- ｢予約商品の価格保証｣では、お客様が対象商品を予約注文した時点から発送手続きに入る時点、または発売日のいずれか早い時点までの期間中のAmazon.co.jp のサイト上で表示される最低販売価格が、お支払いいただく金額となります。予約商品の価格保証について詳しくはヘルプページをご覧ください。 詳細はこちら (細則もこちらからご覧いただけます)
Stream Processing with Apache Spark: Mastering Structured Streaming and Spark Streaming (英語) ペーパーバック – 2019/6/27
Kindle 端末は必要ありません。無料 Kindle アプリのいずれかをダウンロードすると、スマートフォン、タブレットPCで Kindle 本をお読みいただけます。
To build analytics tools that provide faster insights, knowing how to process data in real time is a must, and moving from batch processing to stream processing is absolutely required. Fortunately, the Spark in-memory framework/platform for processing data has added an extension devoted to fault-tolerant stream processing: Spark Streaming.
If you're familiar with Apache Spark and want to learn how to implement it for streaming jobs, this practical book is a must.
- Understand how Spark Streaming fits in the big picture
- Learn core concepts such as Spark RDDs, Spark Streaming clusters, and the fundamentals of a DStream
- Discover how to create a robust deployment
- Dive into streaming algorithmics
- Learn how to tune, measure, and monitor Spark Streaming
Gerard Maas is the lead engineer at Kensu.io, an early stage startup where he works on context management for big-data environments. Previous to that, he led the design and development of the data processing pipeline of Virdata.com, a startup building a cloud-native IoT platform, where Scala, Apache Spark and Spark Streaming were crucial building blocks. He enjoys contributing to open source projects, small and large.
Through his career in technology companies like Alcatel-Lucent, Bell Labs, Sony and Technicolor, he has been mostly involved in the interaction of services and devices, from early days service adaptation when mobile screens only had few text lines, passing through multi-device interactions to IoT device management.
He has a degree in Computer Engineering from the Simón Bolívar University, Venezuela.
François Garillot worked on Scala's type system in 2006, earned his PhD from the French École Polytechnique in 2011, and worked at Typesafe, after a brief stint in Internet advertising. He's worked on interactive interfaces to the Scala compiler, while nourishing a strong enthusiasm for data analytics in his spare time, until Apache Spark let him fullfill this passion as his main job. He received the first Spark Certification in November 2014, and worked in London and Philadelphia, among other places.
In his spare time, he can be found practicing one of a half-dozen ways of making coffee, climbing up or skiing down a not-necessarily-Alpine mountain, or sailing a not-necessarily coastal course.