We will host a series of meetup events to showcase some of the cutting edge developments that are happening in data stream processing in the open sourced tools and industry. We will invite the PMCs of Apache Beam, Airflow, Spark, Flink, Kafka, etc.. and teams who use these tools to share practical experiences, best practices in building data stream processing pipelines. The community of real-time technology leaders and developers will come together to share best practices and use cases as well as explore the vision and future of data streaming.
* 6:00pm~6:30pm: Checkin and Food/drink
* 6:30pm~8:00pm: Tech talks
* 8:00pm~8:30pm: Networking and closing
Tech Talk 1: Apache Beam - fully language portable and scalable batch and streaming data processing.
Abstract: Apache Beam introduces a unified programming model for batch and streaming data processing. Beam provides SDKs for various programming languages, for example, Java, Python and Go and Beam pipelines are executed in a runner, for example, Apache Flink, Apache Spark and Google Cloud Dataflow. Beam provides a portability framework that allows a given runner to execute transforms defined in any given SDK. This lets Beam runners maintain one implementation that supports all current and future SDKs. Additionally, this allows Beam runners to use transforms from different SDK languages in the same pipeline. In this talk, we'll look into Apache Beam’s portability framework and its benefits.