This document presents a method for estimating the pose of a mobile robot using fusion of IMU and vision data via an extended Kalman filter. It discusses using IMU to provide fast updates but is prone to drift, while vision can reduce drift but has limitations like occlusion. The proposed method uses an Arduino, camera, and robot to collect IMU and image data, applies SURF and RANSAC for feature matching, and fuses the data in an EKF. Experimental results show the fused method improves accuracy over individual sensors, with position errors under 15cm and orientation errors under 1 degree. The paper concludes the method provides accurate indoor localization suitable for mobile robots.