This document summarizes a system that processes billions of daily events using Apache Spark and Kafka. The system ingests 1.5-2 million events per minute from various sources, generates over 30 million events per minute, and stores over 5 TB of raw data daily. It uses Spark Streaming to process real-time data from Kafka, Spark jobs to handle raw data stored in Parquet format, and updates MySQL 1500 times per second with aggregated results. The system is run on Amazon EC2 infrastructure and stores data on S3 and Glacier.