This document describes research into developing a conversational robot that can integrate speech acts and sensorimotor acts when resolving ambiguities. The robot needs to decide whether to ask a clarifying question or perform a sensory action like moving its head to see from a different perspective. The researchers present a planning algorithm that treats speech acts and sensory actions in a common framework by calculating the expected costs and information rewards of different actions. They evaluate the algorithm's performance under various settings and discuss possible extensions.