Authors: Ramesh Shanmugam & Aditi Verma
Flink Forward San Francisco is a couple of days away! In case you haven’t booked your tickets yet, here’s a sneak preview of our session Scaling a real-time streaming warehouse with Apache Flink, Parquet and Kubernetes, on April 2, 2019, to give you some more insight into what you can expect at the conference next week.
If you haven’t registered already, make sure to book your last minute tickets while they last! Spots are limited so hurry up to secure your place at Flink Forward and learn more about the exciting world of Apache Flink!
Branch is the industry-leading mobile measurement and deep linking platform. For this, we process more than 20 billion events and store several terabytes of data per day.
In this talk, we cover our learnings and challenges running and scaling an Apache Flink Parquet warehouse on Kubernetes. We share our challenges around memory management and failure recovery. We also talk in detail about our current Apache Flink infrastructure, recovery and auto-scaling mechanisms.
This talk covers a detailed overview of our challenges around writing columnar file formats with Flink. We also talk about the decisions taken and learnings around migrating Flink jobs from Mesos on Kubernetes. Then we talk about auto scaling Flink jobs on Kubernetes, as well as efficiently handling failure scenarios
Learnings from running Apache Flink clusters on Mesos and Kubernetes
Takeaways from writing Parquet files with ApacheFlink
Make sure to secure your spot by registering on the Flink Forward website today. The event includes multiple tracks and it’s a unique opportunity to bring your knowledge and stream processing expertise to the next level! Sessions cover among other Flink use cases, technology deep dives, Apache Flink and stream processing ecosystem talks and deep dives so don’t miss out on the exciting conference schedule!
About the authors:
Aditi Verma
Ramesh Shanmugam