Declarative workflows for building Spark Streaming
Spark Streaming is an extension of the core Spark API that enables stream processing from a variety of sources. Spark is a extensible and programmable framework for massive distributed processing of datasets, called Resilient Distributed Datasets (RDD). Spark Streaming receives input data streams and divides the data into batches, which are then processed by the Spark engine to generate the results.
Spark Streaming data is organized into a sequence of DStreams, represented internally as a sequence of RDDs.
StreamingPro is not only a complete application, but also a extensible and programmable framework for spark streaming (also include spark,storm) that can easily be used to build your streaming application.
StreamingPro also make it possible that all you should do to build streaming program is assembling components(eg. SQL Component) in configuration file.
- Pure Spark Streaming(Or normal Spark) program (Storm in future)
- No need of coding, only declarative workflows
- Rest API for interactive
- SQL-Oriented workflows support
- Data continuously streamed in & processed in near real-time
- dynamically CURD of workflows at runtime via Rest API
- Flexible workflows (input, output, parsers, etc...)
- High performance
- Scalable
链接: http://pan.baidu.com/s/1minPZny 密码: 9jsz
链接: http://pan.baidu.com/s/1eRZjib4 密码: 48fr
- Three steps to run your first application
- 三步跑起你的第一个应用
- 利用StreamingPro实现SQL-交互式查询
- 使用Spark SQL 构建流式处理程序
- 使用Spark SQL构建批处理程序
- 流式计算常见模块用法说明
- Properties
- Build
- Run your first application
- Submit application
- dynamically CURD of workflows at runtime via Rest API
- Recovery
- Useful modules introduction
- Other runtime support
If no picture show,please click me
If github is too slow to view ,please click me