I am a senior big data developer at Nielsen.
My work focuses on building massive data pipelines (~250 Billion events/day) and infrastructure for running machine learning algorithms.
My projects run on AWS using a variety of technologies like Kafka, Spark, Airflow, Kubernetes and more.
I like to continuously experiment with new technologies, tackle challenging problems, and find those better, more elegant and cost effective solutions.
Sessions
- Scale16.06.2021 18:10 – 18:40KesselhausIntermediate
Should you consume Kafka in a stream OR batch? When should you choose each one? What is more efficient, and cost effective? Should you even care?
In this talk we’ll give you the tools and metrics to decide which solution you should apply when, and show you a real life example with cost…