Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Headless with Cassandra: The nyt⨍aбrik project ...

Headless with Cassandra: The nyt⨍aбrik project at the New York Times

Systems architect Michael Laing gave this presentation at FOSDEM 2014

Cassandra provides the global persistence layer for the New York Times nyt⨍aбrik project. This presentation focuses on the use of Cassandra as the high performance distributed data store supporting the nyt⨍aбrik.

nyt⨍aбrik (in production January 2014) is reliable, low latency messaging middleware connecting internal clients at the New York Times (breaking news, user generated content, etc) with millions of external clients around the world. The primary technologies employed are: RabbitMQ (AMQP), Cassandra, and websockets/sockjs. Components developed by the New York TImes will be made open source beginning in 2014.

The New York Times Developers

February 02, 2014
Tweet

More Decks by The New York Times Developers

Other Decks in Technology

Transcript

  1. Headless with Cassandra a simple, reliable persistence layer for the

    nyt aбrik global messaging platform Michael Laing 2014-02-02
  2. Messaging everywhere • Simpler • Scalable • Resilient Add just

    enough structure to the Internet cloud nyt aбrik
  3. Technical problems: Our solutions • Backbone messaging: AMQP (RabbitMQ) •

    Client messaging: websockets / sockjs (pylws - to be open sourced) • Global, scalable resources: cloud (AWS)
  4. Problems: architecture • RabbitMQ ◦ Excellent for routing ◦ Excellent

    for queuing ◦ Not a database • Websockets / sockjs ◦ Excellent for message interchange ◦ Really not a database! • But we need a message cache ◦ Unconnected clients ◦ Archiving ◦ Analysis
  5. So what is nyt aбrik? • It’s an architectural platform

    that allows dozens of NYTimes systems and millions of client devices to rapidly exchange billions of messages • It’s a ‘chat’ system for things that belong to us and to our clients and partners - phones, web browsers, refrigerators, advertisements, etc. • (It’s also a system ‘for the rest of us’) • It needs a cache
  6. A simple message structure • A message has: ◦ message_uuid

    (a version 1 UUID) ◦ replica_uuid (a version 1 UUID) ◦ metadata (JSON) ◦ optional body (BLOB - large ones are referenced in metadata) ◦ a time-to-live (ttl - all ttls are < 30 days)
  7. Simple message indexing • A message has one or more

    ‘paths’ carried in its metadata • Each path is comprised of: ◦ collection ◦ hash_key ◦ range_key (implicit = message_uuid) • An example: ◦ collection: ‘feeds.breaking-news’ ◦ hash_key: 12345 ◦ path: ‘feeds.breaking_news.12345’[UUIDs]
  8. Simple query patterns: get latest • Get latest messages in

    a subtree: ◦ Walk a subtree of the path ◦ return the latest message for each complete path found • Used to: ◦ Get the latest versions of news items within a category, e.g. query path ‘feeds.breaking-news.#’ will retrieve the latest version of each breaking news item ◦ Get the latest versions of client information for a client
  9. Simple query patterns: get all • Get all unexpired messages

    for a path up to a limit: ◦ Find the path ◦ Return messages in reverse date order up to the limit • Used to: ◦ Get metrics from a time bucket, e.g. query path ‘metrics. searchcloud.minute.2014-02-01T09:39Z’ will retrieve all the messages in that bucket ◦ Get all the unexpired versions of a specific information set, e.g. a to do list
  10. Other simple query patterns • Get a message by message_uuid:

    • Get all messages by time bucket (journal) • Get a range of paths
  11. I love relational whatever! • I remember pre-SQL ◦ CODASYL

    ◦ Cullinet ◦ Pick ◦ Track/block ◦ ... • I started with relational algebra and calculus • Some nerdy stories… ok I’ll keep it short!
  12. Relational = Beautiful • IMHO: the mathematical grounding provides elegance

    and power • But! Another story, older and perhaps more relevant... ◦ Reality cannot always be addressed by closed form solutions • Some factors push you out of the SQL sweet spot: ◦ Time ◦ Space ◦ Volume
  13. Reality bites • Goals for the nyt aбrik message cache:

    ◦ globally distributed ◦ high volume ◦ low cost ◦ resilient • NoSQL is the answer (read up on the CAP theorem if you don’t know it)
  14. Reality bites again on the other leg • NoSQL doesn’t

    do as much for us - we have to do it ourselves • Answers: ◦ do it in the application ◦ simplify ◦ re-simplify ◦ ok really simplify this time
  15. Criteria for cache • Multi-region (global) • Open source (no

    license fee) • Scalable cost and volume • Manageable - not Java
  16. Possible answers • AWS Dynamo ◦ great scalability and manageability

    ◦ our first implementation ◦ not multi-region... • Riak ◦ scalable, manageable ◦ have to pay for multi-region • Cassandra ◦ scalable, might be manageable (Java…) ◦ new version w improved language, new interface library... do it!
  17. Caveat emptor • All interaction with the cache is strictly

    isolated in nyt aбrik - we can switch cache backends very quickly • We are willing to dive into open source code to contribute fixes and already have with Cassandra (python interface)
  18. Choices choices... Initial requirements are pretty small: hundreds of reads/writes

    per second • Aggressive (2.0.n) or “safe” (1.2.n)? ◦ 1.2 has enough features but uses Java 6: difficult to manage on small machines ◦ 2.0 uses Java 7: MUCH better behaved on small machines • Features? ◦ Minimize the use of newer features: secondary indexes, leveled compaction, etc.
  19. Mistakes • Using the ‘collections’ feature to implement message structure

    ◦ The entire collection is read whenever a message is read ◦ “Should have known better” - restructured tables to remove collections • Black launch then launch 8 Jan and aftermath... ◦ Application oversights create 10-100X expected volumes ◦ Some paths written to millions of times resulting in huge rows ◦ Nodes fail and are rebuilt ◦ Queuing, parallelized workers, autoscaling, etc. compensate.for errors so... ◦ No one notices
  20. Amazon Web Services • Regional cluster ◦ 6 nodes: 2

    per zone ◦ m1-medium: 1 virtual CPU, 3.4GB memory, 400GB disk (these machines are WAY small! we launched anyway) ◦ replication factor = 3 • Each region supporting 10 to 100 other nyt aбrik instances • 2 regions currently: Dublin and Oregon - may add Tokyo, São Paulo, Singapore, Sydney,...
  21. Lessons / Advice Keep it simple - use the defaults

    Keep it simple - evolutionary design
  22. Staying in the Cassandra sweet spots • Starting out? Use

    version 2, use cql3, use the defaults, be wary of features • Really. USE THE DEFAULTS! Have a good reason to deviate. • A good reason: we never use ‘delete’, are careful w overwrites, and manage data size with truncates and ttls. Hence we can: ◦ Garbage collect immediately (gc_grace_period = 0) ◦ Avoid periodic repair of nodes (big load on small machines)
  23. Evolve your design • Cassandra is not happy about some

    schema changes ◦ avoid dropping and recreating ◦ this will get better • Watch usage patterns and progressively simplify ◦ Writes are so cheap that we run versions of tables in parallel ◦ We gradually migrate code to use new versions • Much of our tweaking has to do with avoiding ‘large rows’
  24. nyt aбrik: next? Metrics - generated by internal systems User

    events - generated by client devices Result: substantially higher volumes
  25. Metrics: gotta love them too! • First project going into

    production this week: searchcloud ◦ what are people searching for ◦ not too much volume ◦ no rollup or cache access initially • Underway: Cassandra metrics! ◦ 1400+ metrics ◦ differential protocol buffers ◦ blog posts soon • Future: metrics supporting analytical client apps
  26. Events happen... • Lots of potential user events • Websockets

    provides an efficient 2-way connection for gathering events • Scaling needed for the cache: ◦ up: bigger instance types to regain the Cassandra sweet spot ◦ out: more nodes ◦ nothing else changes :)