This document provides examples of using SparkR to perform distributed computing tasks like word counting on HDFS files, distributed k-means clustering of large datasets, and saving/loading k-means models to/from HDFS. It shows how to use SparkR functions like mapreduce, to.dfs, from.dfs, and hdfs.write/hdfs.read to parallelize work across a cluster and handle large amounts of data.