You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
remove req.txt in LLM inference example env_activate (#3848)
* remove req.txt in LLM inference example env_activate * CI fail debug logs * disable exit on error since the error is expected * CI script will take care of `set` commands * remove KMP BARRIER env for testing; remove torchscript in README.md * restore KMP BARRIER env variables --------- Co-authored-by: DiweiSun <105627594+DiweiSun@users.noreply.github.com>
Copy file name to clipboardExpand all lines: examples/cpu/inference/python/python-scripts/README.md
-25Lines changed: 0 additions & 25 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,7 +23,6 @@ We provided inference examples for eager mode as well as graph mode, in which th
23
23
Eager mode is the default execution mode in PyTorch, the codes are executed in a “define-by-run” paradigm, so it is flexible, interactive and easy to debug.
24
24
On the other hand, in graph mode the codes are executed in “define-and-run” paradigm, which means the building of the entire computation graph is required before running the function.
25
25
During the graph compilation process, optimizations like layer fusion and folding are applied, and the compiled graphs are more friendly for backend optimizations, leading to accelerated execution.
26
-
TorchScript and TorchDynamo are the 2 graph compiling tools that PyTorch\* provides.
27
26
28
27
From numerical precision perspective, we provided inference examples for [BFloat16](#bfloat16) and [INT8 quantization](#int8) in addition to the default [Float32](#float32) precision.
29
28
Low-precision approaches including [Automatic Mixed Precision (AMP)](https://pytorch.org/docs/stable/amp.html) and [quantization](https://pytorch.org/docs/stable/quantization.html) are commonly used in PyTorch\* to improve performance.
@@ -44,12 +43,6 @@ Running ResNet50 inference in eager mode:
0 commit comments