Stream processing¶
Stream processing is a computing paradigm focused on the high-speed calculation and analysis of data streams.^[600-developer-big-data-big-data.md]
It is a key capability within modern big data frameworks, specifically highlighted in the Apache Spark ecosystem.^[600-developer-big-data-big-data.md] Unlike batch processing methods, stream processing involves real-time data ingestion and computation, often referred to as flow calculation or stream computing.^[600-developer-big-data-big-data.md]
Implementation¶
In the context of the Spark framework, stream processing is handled by Spark Streaming.^[600-developer-big-data-big-data.md] This component allows for the continuous processing of live data streams and is considered one of the core modules alongside Spark SQL and the core Spark calculation engine.^[600-developer-big-data-big-data.md]
Related Concepts¶
- Apache Spark
- [[Big data]]
- [[Batch processing]]
Sources¶
- 600-developer-big-data-big-data.md