Skip to content

In this project, I will work on a real-world dataset, extracted from Kaggle, on San Francisco crime incidents, and will provide statistical analyses of the data using Apache Spark Structured Streaming. Will create a Kafka server to produce data, and ingest data through Spark Structured Streaming.

Notifications You must be signed in to change notification settings

onkargagare/San-Francisco-crime-statistics-and-spark-streaming

Repository files navigation

1.How did changing values on the SparkSession property parameters affect the throughput and latency of the data?

--> The processedRowsPerSecond increase and decrease affected the throughput and latency of the data, the next batch of data is kept on hold until the current batch is processed.

2.What were the 2-3 most efficient SparkSession property key/value pairs? Through testing multiple variations on values, how can you tell these were the most optimal?

--> processedRowsPerSecond depends on these settings- spark.streaming.kafka.maxRatePerPartition,
spark.default.parallelism,
spark.sql.shuffle.partitions

About

In this project, I will work on a real-world dataset, extracted from Kaggle, on San Francisco crime incidents, and will provide statistical analyses of the data using Apache Spark Structured Streaming. Will create a Kafka server to produce data, and ingest data through Spark Structured Streaming.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages