Discussion about this post

User's avatar
Neural Foundry's avatar

Meta's Omnilingual ASR covering 1,600 languages is actualy a massive contribution to open source. The fact they got below 10% character error rates for 78% of those languages is impressive, especialy when you consider 500 of them had zero ASR coverage before. The zero shot capability for 5,000 languages via in-context learning feels like a diferent approach from the usual scaling paradigm we see with these other models. Curious how this stacks up against Whisper in practial deployments.

Expand full comment

No posts

Ready for more?