We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4bdd3df commit 4f8ebfdCopy full SHA for 4f8ebfd
backends/cuda/cuda_backend.py
@@ -129,6 +129,8 @@ def preprocess(
129
user_input_placeholders.append(node.meta["val"])
130
131
options: dict[str, typing.Any] = {
132
+ # Better model precision
133
+ "emulate_precision_casts": True,
134
# Embed CUDA kernel binaries directly into the compiled shared object
135
"aot_inductor.embed_kernel_binary": True,
136
# Do not link against the full PyTorch/libtorch library
0 commit comments