Advertisement Β· 728 Γ— 90

Posts by Laurent Mazare

I have a bunch of catch up to do in ocaml, effects/multicore/oxcaml all seem very exciting to play with!

9 months ago 2 0 1 0

Thanks for the kind words @anil.recoil.org , just to mention that these models are a team effort from a bunch of folks here at Kyutai! The MLX bugs are most likely mine though :) hopefully I'll get some free cycles soon and make an ocaml implementation that will have less of these!

9 months ago 2 0 1 0

πŸš€ Say hello to unmute.sh β€” a modular voice AI system built on our in-house low latency text-to-speech and speech-to-text engines. It works in English πŸ‡¬πŸ‡§ and French πŸ‡«πŸ‡· and you can customize the voice and personality.
πŸŽ™οΈTry it live and tell us what you think!

10 months ago 2 0 0 0

I'll present a dive into Moshi 🟒 and our translation model Hibiki πŸ‡«πŸ‡·β™»οΈπŸ‡¬πŸ‡§ as part of the next @convai-rg.bsky.social reading group πŸ‘¨β€πŸ«πŸ“—.

πŸ“… 13th of March πŸ•°οΈ 11am ET, 4pm in Paris.

I'll discuss Mimi πŸ—œοΈ and multi-stream audio modeling πŸ”Š.
Join on Zoom, replay on YT.

⬛ ⬛ 🟧 🟧 🟨 🟨 🟩 🟩 🟩 ⬛
⬛ 🟧 🟧 🟨 🟨 🟩 🟩 🟩 ⬛ ⬛

1 year ago 6 1 0 1

Ah, we already have inference codebases in python/rust/swift but maybe we need an ocaml one too for folks that interact with the French ocaml community πŸ˜„

1 year ago 5 0 0 0
Video

Afraid of missing out on French pop culture references because you don't speak the language? Fear no more and try our Hibiki speech-to-speech translation modelβ€” no more FOMO! πŸ‡«πŸ‡·βœ¨ #Translation #AI

1 year ago 7 1 1 0
Video

Even Kavinsky 🎧πŸͺ© can't break Hibiki! Just like Moshi, Hibiki is robust to extreme background conditions πŸ’₯πŸ”Š.

1 year ago 8 4 0 1
Advertisement
Video

We just released Hibiki 🟒, a real time speech-to-speech translation πŸ‡«πŸ‡· -> πŸ‡¬πŸ‡§. It preserves the voice of the user, and the smaller variant can run on iPhone as showed by Neil in this video.
Find the code on github github.com/kyutai-labs/... and the weights on HF and give it a spin!

1 year ago 8 2 0 1
Preview
GitHub - nu-dialogue/j-moshi: J-Moshi: A Japanese Full-duplex Spoken Dialogue System J-Moshi: A Japanese Full-duplex Spoken Dialogue System - nu-dialogue/j-moshi

Very impressive to hear this Japanese πŸ‡―πŸ‡΅ version of moshi 🟒. I don't speak the language so I cannot understand what it's trying to tell me but at least it sounds great πŸ˜…
github.com/nu-dialogue/...

1 year ago 3 0 0 0

Getting our latest LLM to run on the edge was pretty fun, it had been a while since I last used swift and it's still a pretty enjoyable language!

1 year ago 1 0 0 0

That's certainly the plan, ideally this will make it possible to run a moshi locally on a mobile phone but there is still a lot to be done before we get there!

1 year ago 1 0 0 0

Super proud of our first publicly released text model, helium-1 preview, a 2B model trained on 6 languages. It should be a great fit for on-device applications. Already available in candle/transformers, can't wait to see what the community builds with it! #OpenSource #AI #FTW!

1 year ago 7 0 1 0
Preview
GitHub - LaurentMazare/ug: Experimental compiler for deep learning models Experimental compiler for deep learning models. Contribute to LaurentMazare/ug development by creating an account on GitHub.

It's certainly open-source but also pretty much a work in progress :) github.com/LaurentMazar... There is no AMD support for now but as long as the programming model is similar to cuda or metal, adding a new backend should be easy (or at least that's the goal).

1 year ago 2 0 0 0

Ah, sounds like a nice part to have! Note that the bindings are now developed by JS here github.com/janestreet/t... Also lately I've worked on a small compiler for deep learning models, inspired by triton/tinygrad. It's written in rust, but ocaml would be a great choice for writing such a compiler.

1 year ago 5 0 2 0

The benchmark code with all the matmul variants is on github: github.com/LaurentMazar...

1 year ago 0 0 0 0
Post image

Last week we've received a new M4pro mac mini so I've benched it with various matmul variants and the results are pretty impressive for a tiny form factor. Even with a naive approach it reaches ~5.2TFlops in f32 (so probably more than 10TFlops in bf16), and that's just using the GPU, no NPU for now.

1 year ago 3 0 1 0