This document discusses using Spark Streaming to dynamically flag transactions while controlling their throughput based on various parameters. It describes decoupling the task creation process from the task flagging process to allow real-time visibility on task flow rates and the ability to perform analytics. It provides two examples of flagging strategies - random sampling which dynamically adjusts sampling rates, and transaction analysis which flags outliers. It also covers maintaining state information and getting data into and out of Spark Streaming and Kafka.
Related topics: