Java Application, uses Apache Spark, handles batch as well as streaming processing
- Updated
Jan 22, 2021 - Java
Java Application, uses Apache Spark, handles batch as well as streaming processing
A sandbox environment designed to simulate a pseudo-distributed Hadoop cluster with integrated Apache Spark and Kafka components. It allows developers to prototype and experiment with big data workflows, test distributed computing patterns, and explore cluster behavior in a contained virtual setup.
Add a description, image, and links to the dataframes-api topic page so that developers can more easily learn about it.
To associate your repository with the dataframes-api topic, visit your repo's landing page and select "manage topics."