The document summarizes a research paper titled "Continuous Deep Q-Learning with Model-based Acceleration" presented at ICML 2016. It proposes a method that incorporates advantages of both model-free and model-based reinforcement learning. The method uses deep Q-learning with normalized advantage functions to learn a parameterized Q-function for continuous state-action spaces. It accelerates the learning process by using trajectory optimization from an imagined model to generate exploratory behaviors during data collection.