Update README.md
Browse files
README.md
CHANGED
|
@@ -107,7 +107,7 @@ pipe({'audio': audio, 'turns': turns, 'sampling_rate': sr}, max_new_tokens=30)
|
|
| 107 |
|
| 108 |
The model uses a pre-trained [Llama3.2-1B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.2-1B) backbone as well as the encoder part of [whisper-large-v3-turbo](https://huggingface.co/openai/whisper-large-v3-turbo).
|
| 109 |
|
| 110 |
-
|
| 111 |
|
| 112 |
We use a knowledge-distillation loss where Ultravox is trying to match the logits of the text-based Llama backbone.
|
| 113 |
|
|
|
|
| 107 |
|
| 108 |
The model uses a pre-trained [Llama3.2-1B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.2-1B) backbone as well as the encoder part of [whisper-large-v3-turbo](https://huggingface.co/openai/whisper-large-v3-turbo).
|
| 109 |
|
| 110 |
+
The multi-modal adapter is trained, the Whisper encoder is fine-tuned, while the Llama model is kept frozen.
|
| 111 |
|
| 112 |
We use a knowledge-distillation loss where Ultravox is trying to match the logits of the text-based Llama backbone.
|
| 113 |
|