Discussion about this post

User's avatar
Neural Foundry's avatar

Meta's Omnilingual ASR covering 1,600 languages is actualy a massive contribution to open source. The fact they got below 10% character error rates for 78% of those languages is impressive, especialy when you consider 500 of them had zero ASR coverage before. The zero shot capability for 5,000 languages via in-context learning feels like a diferent approach from the usual scaling paradigm we see with these other models. Curious how this stacks up against Whisper in practial deployments.

No posts

Ready for more?