Cloud Native Data Pipelines
This video is also available in the GOTO Play video app! Download it to enjoy offline access to our conference videos while on the move.
Big Data companies (e.g. LinkedIn, Facebook, Google, and Twitter) have historically built custom data pipelines over bare metal in custom-designed data centers. In order to meet strict requirements on data security, fault-tolerance, cost control, job scalability, network topology, and compute and storage placement, they need to closely manage their core technology. In recent years, many companies with Big Data needs have started migrating to one of the public cloud vendors. How does the public cloud change the game? Specifically, how can companies effectively marry cloud best-practices with big data technology in order to leverage the benefits of both? Agari, a leading email security company, is applying big data best practices to both the security industry and to the cloud in order to secure the world against email-bourne threats. We do this by building both batch and stream processing predictive data pipelines in the AWS cloud. Come to this talk to learn about our architectural best practices and technologies.
-
Processing Data of Any Size with Apache BeamJesse AndersonTuesday May 2 @ 11:40 AM
-
Apache Spark Beyond Shuffling - Why it isn't Magic - but also where there is some really cool MagicHolden KarauTuesday May 2 @ 3:40 PM
-
Apache Flink - The State of the Art in Streaming ComputationJamie GrierTuesday May 2 @ 1:30 PM
-
Fast Data Architectures for Streaming ApplicationsDean WamplerTuesday May 2 @ 10:35 AM
-
Cloud Native Data PipelinesSid AnandTuesday May 2 @ 2:35 PM
-
Stream All Things - Patterns of Modern Data IntegrationGwen ShapiraTuesday May 2 @ 4:45 PM