Also started working on openvino integration, but differently: based on openvino/model_server container images, using the latest models from ultralytics (26). It will allow spreading the load over multiple servers/devices (Jetson Nano should in theory work as well), and will allow deploying to the OS of your liking. Will update here if I get it working properly.
Preliminary results show that inference with yolo26n is about 20% faster than yolo11n (on CPU, without any GPU).