Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

At the moment, no unfortunately. However, to my recent knowledge of open source alternatives, the vLLM team published a separate repository for omni models now:

https://github.com/vllm-project/vllm-omni

I have not yet tested out if this does full speech to speech, but this seems like a promising workspace for omni-modal models.





Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: