This document summarizes a presentation about reducing tail latency in Cassandra clusters using a regression-based replica selection algorithm. It discusses how tail latency occurs in distributed systems and how previous approaches used replica selection to reduce it. The proposed approach uses linear regression models to predict query execution times and select the fastest replica. Experimental results on homogeneous and heterogeneous server clusters show the approach reduces tail latency metrics like p999 while maintaining throughput. However, it degrades some lower percentile metrics. Future work could explore more advanced machine learning models.