This document is a final report for a CS799 course that explores using reinforcement learning to train an agent to play a chasing game. The author defines the game environment and mechanics, then uses Q-learning with an epsilon-greedy exploration strategy to train an agent to maximize its score by collecting vegetables while avoiding walls, minerals, and other players. The agent is trained in multiple phases to first avoid walls, then minerals, and finally other players while collecting vegetables. Results are presented comparing training with different exploration vs exploitation settings.