Skip to content

Commit ec2d22c

Browse files
committed
Deployed 998776c to 0.3 with MkDocs 1.6.0 and mike 2.1.3
1 parent 998776c commit ec2d22c

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

0.3/overview/index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1005,7 +1005,7 @@ <h3 id="key-features">Key Features</h3>
10051005
<li><strong>Supports a Wide Variety of Hardware:</strong> Run with different brands of GPUs in Apple MacBooks, Windows PCs, and Linux servers.</li>
10061006
<li><strong>Scales with Your GPU Inventory:</strong> Easily add more GPUs or nodes to scale up your operations.</li>
10071007
<li><strong>Distributed Inference</strong>: Supports both single-node multi-GPU and multi-node inference and serving.</li>
1008-
<li><strong>Multiple Inference Backends</strong>: Supports llama-box(llama.cpp) and vLLM as the inference backend.</li>
1008+
<li><strong>Multiple Inference Backends</strong>: Supports llama-box (llama.cpp) and vLLM as the inference backend.</li>
10091009
<li><strong>Lightweight Python Package:</strong> Minimal dependencies and operational overhead.</li>
10101010
<li><strong>OpenAI-compatible APIs:</strong> Serve APIs that are compatible with OpenAI standards.</li>
10111011
<li><strong>User and API key management:</strong> Simplified management of users and API keys.</li>

0.3/search/search_index.json

Lines changed: 1 addition & 1 deletion
Large diffs are not rendered by default.

0.3/user-guide/model-management/index.html

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1429,11 +1429,11 @@ <h4 id="manual">Manual</h4>
14291429
</ul>
14301430
<p>Select a GPU from the list. The model instance will attempt to deploy to this GPU if resources permit.</p>
14311431
<h3 id="backend">Backend</h3>
1432-
<p>The inference backend. Currently, GPUStack supports two backends: llama-box and vLLM. GPUStack automatically selects the backend based on the model's information.</p>
1432+
<p>The inference backend. Currently, GPUStack supports two backends: llama-box and vLLM. GPUStack automatically selects the backend based on the model's configuration.</p>
14331433
<p>For more details, please refer to the <a href="../inference-backends/">Inference Backends</a> section.</p>
14341434
<h3 id="backend-parameters">Backend Parameters</h3>
1435-
<p>Input the parameters for the backend you want to customize while running the model instance. Should be in the format <code>--parameter=value</code>, <code>--bool-parameter</code> or separate <code>--parameter</code> and <code>value</code> in two fields.
1436-
For example, <code>--ctx-size=8192</code> for llama-box.</p>
1435+
<p>Input the parameters for the backend you want to customize when running the model. The parameter should be in the format <code>--parameter=value</code>, <code>--bool-parameter</code> or as separate fields for <code>--parameter</code> and <code>value</code>.
1436+
For example, use <code>--ctx-size=8192</code> for llama-box.</p>
14371437
<p>For full list of supported parameters, please refer to the <a href="../inference-backends/">Inference Backends</a> section.</p>
14381438
<h3 id="allow-cpu-offloading">Allow CPU Offloading</h3>
14391439
<div class="admonition note">

0 commit comments

Comments
 (0)