This document discusses integrating deep learning models from Python frameworks like PyTorch and CNTK into Flink pipelines for real-time inference on streaming data. It presents ONNX as a standard format to export models, and explores options like using Java Embedded Python (JEP) to run models within Flink jobs or creating a microservice. Examples demonstrate named entity recognition and sentiment analysis on Twitter data streams using Flair models with JEP. Overall, the easiest approaches currently are using JEP to avoid converting models to ONNX and rewriting code, though native ONNX support in frameworks like Menoh, VESPA and Deeplearning4j may improve in the future.