Mention the standalone rust example in the readme.
This commit is contained in:
parent
e8de86bdab
commit
94f1692cc6
10
README.md
10
README.md
|
|
@ -36,6 +36,16 @@ python -m moshi_mlx.run_inference --hf-repo kyutai/stt-2.6b-en-mlx bria.mp3 --te
|
||||||
### Rust implementation
|
### Rust implementation
|
||||||
[[Hugging Face]](https://huggingface.co/kyutai/stt-2.6b-en-candle)
|
[[Hugging Face]](https://huggingface.co/kyutai/stt-2.6b-en-candle)
|
||||||
|
|
||||||
|
A standalone Rust example is provided in the `stt-rs` directory in this repo.
|
||||||
|
This can be used as follows:
|
||||||
|
```bash
|
||||||
|
cd stt-rs
|
||||||
|
cargo run --features cuda -r -- bria.mp3
|
||||||
|
```
|
||||||
|
|
||||||
|
### Rust server
|
||||||
|
[[Hugging Face]](https://huggingface.co/kyutai/stt-2.6b-en-candle)
|
||||||
|
|
||||||
The Rust implementation provides a server that can process multiple streaming
|
The Rust implementation provides a server that can process multiple streaming
|
||||||
queries in parallel. Dependening on the amount of memory on your GPU, you may
|
queries in parallel. Dependening on the amount of memory on your GPU, you may
|
||||||
have to adjust the batch size from the config file. For a L40S GPU, a batch size
|
have to adjust the batch size from the config file. For a L40S GPU, a batch size
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue
Block a user