Skip to content

Commit d85c47f

Browse files
authored
update 2.1.30 llm.html (#2876)
1 parent 9700007 commit d85c47f

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

xpu/2.1.30+xpu/_sources/tutorials/llm.rst.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ Optimized Models
4848

4949
*Note*: The above verified models (including other models in the same model family, like "codellama/CodeLlama-7b-hf" from LLAMA family) are well supported with all optimizations like indirect access KV cache, fused ROPE, and prepacked TPP Linear (fp16). For other LLMs families, we are working in progress to cover those optimizations, which will expand the model list above.
5050

51-
Check `LLM best known practice <https://github.com/intel/intel-extension-for-pytorch/tree/v2.1.30%2Bxpu/examples/gpu/inference/python/llm>`_ for instructions to install/setup environment and example scripts..
51+
Check `LLM best known practice <https://github.com/intel/intel-extension-for-pytorch/tree/release/xpu/2.1.30/examples/gpu/inference/python/llm>`_ for instructions to install/setup environment and example scripts..
5252

5353
Optimization Methodologies
5454
--------------------------

xpu/2.1.30+xpu/tutorials/llm.html

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -163,7 +163,7 @@ <h2>Optimized Models<a class="headerlink" href="#optimized-models" title="Permal
163163
</tbody>
164164
</table>
165165
<p><em>Note</em>: The above verified models (including other models in the same model family, like “codellama/CodeLlama-7b-hf” from LLAMA family) are well supported with all optimizations like indirect access KV cache, fused ROPE, and prepacked TPP Linear (fp16). For other LLMs families, we are working in progress to cover those optimizations, which will expand the model list above.</p>
166-
<p>Check <a class="reference external" href="https://github.com/intel/intel-extension-for-pytorch/tree/v2.1.30%2Bxpu/examples/gpu/inference/python/llm">LLM best known practice</a> for instructions to install/setup environment and example scripts..</p>
166+
<p>Check <a class="reference external" href="https://github.com/intel/intel-extension-for-pytorch/tree/release/xpu/2.1.30/examples/gpu/inference/python/llm">LLM best known practice</a> for instructions to install/setup environment and example scripts..</p>
167167
</section>
168168
<section id="optimization-methodologies">
169169
<h2>Optimization Methodologies<a class="headerlink" href="#optimization-methodologies" title="Permalink to this heading"></a></h2>
@@ -260,4 +260,4 @@ <h2>Weight Only Quantization INT4<a class="headerlink" href="#weight-only-quanti
260260
</script>
261261

262262
</body>
263-
</html>
263+
</html>

0 commit comments

Comments
 (0)