You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Interactive chat mode (multi-turn with stateful KV cache)
144
+
uv run lfm2-audio-infer LFM2.5-Audio-1.5B-ONNX --mode interleaved --chat \
145
+
--output output.wav --precision q4
146
+
# Commands in chat mode:
147
+
# /audio <file> [text] - Send audio with optional text
148
+
# <text> - Send text message
149
+
# reset - Clear conversation state
150
+
# quit - Exit
151
+
```
152
+
153
+
> **Note:** Audio inference requires the model directory path (not a single .onnx file) since it loads multiple components. Use `--precision` to select quantization level (fp16, q4, q8).
154
+
113
155
## 5. Testing
114
156
115
157
Tests verify ONNX exports against PyTorch reference models.
@@ -149,6 +191,9 @@ uv run lfm2-bench --model LiquidAI/LFM2.5-1.2B-Instruct \
0 commit comments