Every business needs a pipeline for their critical data, even if it is just pasting into a spreadsheet. As the organization grows and gains more customers, the requirements for that pipeline will change. In this episode Christian Heinzmann, Head of Data Warehousing at Grubhub, discusses the various requirements for data pipelines and how the overall system architecture evolves as more data is being processed. He also covers the changes in how the output of the pipelines are used, how that impacts the expectations for accuracy and availability, and some useful advice on build vs. buy for the components of a data platform.
Your data platform needs to be scalable, fault tolerant, and performant, which means that you need the same from your cloud provider. Linode has been powering production systems for over 17 years, and now they’ve launched a fully managed Kubernetes platform. With the combined power of the Kubernetes engine for flexible and scalable deployments, and features like dedicated CPU instances, GPU instances, and object storage you’ve got everything you need to build a bulletproof data pipeline. If you go to dataengineeringpodcast.com/linode today you’ll even get a $60 credit to use on building your own cluster, or object storage, or reliable backups, or… And while you’re there don’t forget to thank them for being a long-time supporter of the Data Engineering Podcast!
- Hello and welcome to the Data Engineering Podcast, the show about modern data management
- When you’re ready to build your next pipeline, or want to test out the projects you hear about on the show, you’ll need somewhere to deploy it, so check out Linode. With 200Gbit private networking, scalable shared block storage, and a 40Gbit public network, you’ve got everything you need to run a fast, reliable, and bullet-proof data platform. If you need global distribution, they’ve got that covered too with world-wide datacenters including new ones in Toronto and Mumbai. Go to dataengineeringpodcast.com/linode today to get a $20 credit and launch a new server in under a minute.
- Go to dataengineeringpodcast.com to subscribe to the show, sign up for the mailing list, read the show notes, and get in touch.
- Join the community in the new Zulip chat workspace at dataengineeringpodcast.com/chat
- Your host is Tobias Macey and today I’m interviewing Christian Heinzmann about how data pipelines evolve as your business grows
- How did you get involved in the area of data management?
- Can you start by sharing your definition of a data pipeline?
- At what point in the life of a project or organization should you start thinking about building a pipeline?
- In the early stages when the scale of the data and business are still small, what are some of the design characteristics that you should be targeting for your pipeline?
- What metrics/use cases should you be optimizing for at this point?
- What are some of the indicators that you look for to signal that you are reaching the next order of magnitude in terms of scale?
- How do the design requirements for a data pipeline change as you reach this stage?
- What are some of the challenges and complexities that begin to present themselves as you build and run your pipeline at medium scale?
- What are some of the changes that are necessary as you move to a large scale data pipeline?
- At each level of scale it is important to minimize the impact of the ETL process on the source systems. What are some strategies that you have employed to avoid degrading the performance of the application systems?
- In recent years there has been a shift to using data lakes as a staging ground before performing transformations. What are your thoughts on that approach?
- When performing transformations there is a potential for discarding information or losing fidelity. How have you worked to reduce the impact of this effect?
- Transformations of the source data can be brittle when the format or volume changes. How do you design the pipeline to be resilient to these types of changes?
- What are your selection criteria when determining what workflow or ETL engines to use in your pipeline?
- How has your preference of build vs buy changed at different scales of operation and as new/different projects become available?
- What are some of the dead ends or edge cases that you have had to deal with in your current role at Grubhub?
- What are some of the common mistakes or overlooked aspects of building a data pipeline that you have seen?
- What are your plans for improving your current pipeline at Grubhub?
- What are some references that you recommend for anyone who is designing a new data platform?
- From your perspective, what is the biggest gap in the tooling or technology for data management today?
- Scaling ETL blog post
- Data Warehouse
- Amazon EMR
- A Primer on Enterprise Data Curation
- Pub/Sub (Publish-Subscribe Pattern)
- Change Data Capture
- Data Lineage
- AirBnB Engineering Blog