We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent be2dda7 commit c6c0899Copy full SHA for c6c0899
backends/cadence/aot/compiler.py
@@ -144,7 +144,6 @@ def fuse_pt2(
144
return converted_graph_module
145
146
147
-# Note: this is the one-liner API to quantize and fuse a model.
148
def quantize_pt2(
149
model: torch.nn.Module,
150
inputs: tuple[object, ...],
@@ -158,6 +157,8 @@ def quantize_pt2(
158
157
not, the inputs will be used for calibration instead, which is useful for
159
unit tests but should not be used for end-to-end use cases.
160
Returns a GraphModule with the quantized model.
+ Note: this function should not be called directly in general. Please use
161
+ quantize_and_export_to_executorch for most needs.
162
"""
163
# Make the model inference mode by calling model.eval()
164
model.eval()
0 commit comments