A data lake can be a highly valuable resource, as long as it is well built and well managed. Unfortunately, that can be a complex and time-consuming effort, requiring specialized knowledge and diverting resources from your primary business. In this episode Yoni Iny, CTO of Upsolver, discusses the various components that are necessary for a successful data lake project, how the Upsolver platform is architected, and how modern data lakes can benefit your organization.
Do you want to try out some of the tools and applications that you heard about on the Data Engineering Podcast? Do you have some ETL jobs that need somewhere to run? Check out Linode at promo.linode.com/dataengineeringpodcast or use the code dataengineering2018 and get a $20 credit (that’s 4 months free!) to try out their fast and reliable Linux virtual servers. They’ve got lightning fast networking and SSD servers with plenty of power and storage to run whatever you want to experiment on.
- Hello and welcome to the Data Engineering Podcast, the show about modern data management
- When you’re ready to build your next pipeline you’ll need somewhere to deploy it, so check out Linode. With private networking, shared block storage, node balancers, and a 40Gbit network, all controlled by a brand new API you’ve got everything you need to run a bullet-proof data platform. Go to dataengineeringpodcast.com/linode to get a $20 credit and launch a new server in under a minute.
- Go to dataengineeringpodcast.com to subscribe to the show, sign up for the mailing list, read the show notes, and get in touch.
- Join the community in the new Zulip chat workspace at dataengineeringpodcast.com/chat
- Your host is Tobias Macey and today I’m interviewing Yoni Iny about Upsolver, a data lake platform that lets developers integrate and analyze streaming data with ease
- How did you get involved in the area of data management?
- Can you start by describing what Upsolver is and how it got started?
- What are your goals for the platform?
- There are a lot of opinions on both sides of the data lake argument. When is it the right choice for a data platform?
- What are the shortcomings of a data lake architecture?
- How is Upsolver architected?
- How has that architecture changed over time?
- How do you manage schema validation for incoming data?
- What would you do differently if you were to start over today?
- What are the biggest challenges at each of the major stages of the data lake?
- What is the workflow for a user of Upsolver and how does it compare to a self-managed data lake?
- When is Upsolver the wrong choice for an organization considering implementation of a data platform?
- Is there a particular scale or level of data maturity for an organization at which they would be better served by moving management of their data lake in house?
- What features or improvements do you have planned for the future of Upsolver?
- From your perspective, what is the biggest gap in the tooling or technology for data management today?
- Data Lake
- Israeli Army
- Data Warehouse
- Data Engineering Podcast Episode About Data Curation
- Three Vs
- Spot Instances
- Object Storage
- Data Engineering Podcast Episode About Data Serialization Formats
- Run Length Encoding
- CSV (Comma Separated Values)
- Protocol Buffers