...
- A developer should be able to create pipelines that contain aggregations (GROUP BY -> count/sum/unique)
- A developer should be able to create a pipeline with multiple sources
- A developer should be able to use a Spark ML job as a pipeline stage
- A developer should be able to rerun failed pipeline runs without reconfiguring the pipeline
- A developer should be able to de-duplicate records in a pipeline
- A developer should be able to join multiple branches of a pipeline
- A developer should be able to use an Explore action as a pipeline stage
- A developer should be able to create pipelines that contain Spark Streaming jobs
- A developer should be able to create pipelines that run based on various conditions, including input data availability and Kafka events
...