pandora-s sanchit-gandhi commited on
Commit
b45b4dc
·
1 Parent(s): 6c177b5

Update README.md (#26)

Browse files

- Update README.md (bfa49d6b2a0d67790db60e8493bcfc8e1e10c58b)


Co-authored-by: Sanchit Gandhi <sanchit-gandhi@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -80,7 +80,7 @@ We recommend deploying with the following best practices:
80
  For the best user experience, we recommend to simply instantiate vLLM with the default parameters which will automatically set a maximum model length of 131072 (~ca. 3h).
81
  - We strongly recommend using websockets to set up audio streaming sessions. For more info on how to do so, check [Usage](#usage).
82
  - We recommend using a delay of 480ms as we found it to be the sweet spot of performance and low latency. If, however, you want to adapt the delay, you can change the `"transcription_delay_ms": 480` parameter
83
- in the [tekken.json](https://huggingface.co/mistralai/Voxtral-Mini-4B-Realtime-2602/blob/main/params.json) file to any multiple of 80ms between 80 and 2400.
84
 
85
  ## Benchmark Results
86
 
 
80
  For the best user experience, we recommend to simply instantiate vLLM with the default parameters which will automatically set a maximum model length of 131072 (~ca. 3h).
81
  - We strongly recommend using websockets to set up audio streaming sessions. For more info on how to do so, check [Usage](#usage).
82
  - We recommend using a delay of 480ms as we found it to be the sweet spot of performance and low latency. If, however, you want to adapt the delay, you can change the `"transcription_delay_ms": 480` parameter
83
+ in the [tekken.json](https://huggingface.co/mistralai/Voxtral-Mini-4B-Realtime-2602/blob/main/params.json) file to any multiple of 80ms between 80 and 1200, as well as 2400 as a standalone value.
84
 
85
  ## Benchmark Results
86