This document summarizes a research paper that proposes a custom spatiotemporal fusion method for video saliency detection. The method involves taking video frames and computing colour and motion saliency. It then performs temporal fusion and pixel saliency fusion. Colour information then guides a spatiotemporal diffusion process using a permutation matrix. The results show the proposed method achieves overall best performance compared to other state-of-the-art saliency detection methods on a publicly available dataset, based on five global saliency evaluation metrics.