1) The document examines what neural machine translation models learn about morphology through experiments analyzing the hidden states of NMT models.
2) It finds that character-based word representations better capture morphological information than word-based representations, and that lower encoder layers learn more about a word's structure while higher layers improve translation.
3) The target language does not significantly impact how much the model learns about source language morphology, and decoder states do not capture rich morphological information.