Elasticsearch is a powerful tool for storing and analyzing data, but when using it for logs and other time oriented information it can become problematic to keep all of your history. Chaos Search was started to make it easy for you to keep all of your data and make it usable in S3, so that you can have the best of both worlds. In this episode the CTO, Thomas Hazel, and VP of Product, Pete Cheslock, describe how they have built a platform to let you keep all of your history, save money, and reduce your operational overhead. They also explain some of the types of data that you can use with Chaos Search, how to load it into S3, and when you might want to choose it over Amazon Athena for our serverless data analysis.
- Hello and welcome to the Data Engineering Podcast, the show about modern data management
- When you’re ready to build your next pipeline you’ll need somewhere to deploy it, so check out Linode. With private networking, shared block storage, node balancers, and a 40Gbit network, all controlled by a brand new API you’ve got everything you need to run a bullet-proof data platform. Go to dataengineeringpodcast.com/linode to get a $/0 credit and launch a new server in under a minute.
- You work hard to make sure that your data is reliable and accurate, but can you say the same about the deployment of your machine learning models? The Skafos platform from Metis Machine was built to give your data scientists the end-to-end support that they need throughout the machine learning lifecycle. Skafos maximizes interoperability with your existing tools and platforms, and offers real-time insights and the ability to be up and running with cloud-based production scale infrastructure instantaneously. Request a demo at dataengineeringpodcast.com/metis-machine to learn more about how Metis Machine is operationalizing data science.
- Go to dataengineeringpodcast.com to subscribe to the show, sign up for the mailing list, read the show notes, and get in touch.
- Join the community in the new Zulip chat workspace at dataengineeringpodcast.com/chat
- Your host is Tobias Macey and today I’m interviewing Pete Cheslock and Thomas Hazel about Chaos Search and their effort to bring historical depth to your Elasticsearch data
- How did you get involved in the area of data management?
- Can you start by explaining what you have built at Chaos Search and the problems that you are trying to solve with it?
- What types of data are you focused on supporting?
- What are the challenges inherent to scaling an elasticsearch infrastructure to large volumes of log or metric data?
- Is there any need for an Elasticsearch cluster in addition to Chaos Search?
- For someone who is using Chaos Search, what mechanisms/formats would they use for loading their data into S3?
- What are the benefits of implementing the Elasticsearch API on top of your data in S3 as opposed to using systems such as Presto or Drill to interact with the same information via SQL?
- Given that the S3 API has become a de facto standard for many other object storage platforms, what would be involved in running Chaos Search on data stored outside of AWS?
- What mechanisms do you use to allow for such drastic space savings of indexed data in S3 versus in an Elasticsearch cluster?
- What is the system architecture that you have built to allow for querying terabytes of data in S3?
- What are the biggest contributors to query latency and what have you done to mitigate them?
- What are the options for access control when running queries against the data stored in S3?
- What are some of the most interesting or unexpected uses of Chaos Search and access to large amounts of historical log information that you have seen?
- What are your plans for the future of Chaos Search?
- From your perspective, what is the biggest gap in the tooling or technology for data management today?
- Chaos Search
- AWS S3
- Distributed Systems
- Information Theory
- Inverted Index
- AWS KMS
- OpenStack Swift
- Elastic Beats
- Data Lake