GOTO Chicago 2017

Tuesday May 2
11:40 –
Location: Vevey 1-2

Processing Data of Any Size with Apache Beam


Rewriting code as you scale is a terrible waste of time. You have perfectly working code, but it doesn't scale. You really need code that works at any size, whether that's a megabyte or a terabyte. Beam allows you to learn a single API and process data as it grows. You don't have to rewrite at every step.

In this session, we will talk about Beam and its API. We'll see how Beam execute on Big Data or small data. We'll touch on some of the advanced features that make Beam an interesting choice.

Jesse Anderson
Managing Director at Big Data Institute
Organized by