-
Notifications
You must be signed in to change notification settings - Fork 577
Unbreak test models llama CI #6026
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/6026
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 04b1dda with merge base b6e6d06 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This pull request was exported from Phabricator. Differential Revision: D64074891 |
This pull request was exported from Phabricator. Differential Revision: D64074891 |
Summary: Pull Request resolved: #6026 Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Differential Revision: D64074891
91fba09
to
7e54dab
Compare
Summary: Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Reviewed By: Jack-Khuu Differential Revision: D64074891
7e54dab
to
bcc8b14
Compare
This pull request was exported from Phabricator. Differential Revision: D64074891 |
1 similar comment
This pull request was exported from Phabricator. Differential Revision: D64074891 |
Summary: Pull Request resolved: #6026 Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Reviewed By: Jack-Khuu Differential Revision: D64074891
bcc8b14
to
0c2febd
Compare
Summary: Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Reviewed By: Jack-Khuu Differential Revision: D64074891
0c2febd
to
70cc91f
Compare
This pull request was exported from Phabricator. Differential Revision: D64074891 |
Summary: Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Reviewed By: digantdesai, Jack-Khuu, tugsbayasgalan Differential Revision: D64074891
Summary: Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975 Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917. The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out. We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987. The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic. Reviewed By: digantdesai, Jack-Khuu, tugsbayasgalan Differential Revision: D64074891
70cc91f
to
75125be
Compare
75125be
to
04b1dda
Compare
This pull request was exported from Phabricator. Differential Revision: D64074891 |
1 similar comment
This pull request was exported from Phabricator. Differential Revision: D64074891 |
This pull request has been merged in 72b3bb3. |
Summary: Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models: - Quantized ViT - Quantized Mobilebert - Quantized Emformer Predict - Quantized Emformer Transcribe Differential Revision: D64081319
pytorch#6242) Summary: Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models: - Quantized ViT - Quantized Mobilebert - Quantized Emformer Predict - Quantized Emformer Transcribe Reviewed By: digantdesai Differential Revision: D64081319
pytorch#6242) Summary: Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models: - Quantized ViT - Quantized Mobilebert - Quantized Emformer Predict - Quantized Emformer Transcribe Reviewed By: digantdesai Differential Revision: D64081319
pytorch#6242) Summary: Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models: - Quantized ViT - Quantized Mobilebert - Quantized Emformer Predict - Quantized Emformer Transcribe Reviewed By: digantdesai Differential Revision: D64081319
#6242) Summary: Pull Request resolved: #6242 Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change #6026, quantization XNNPack options can be enabled by default for the following models: - Quantized ViT - Quantized Mobilebert - Quantized Emformer Predict - Quantized Emformer Transcribe Reviewed By: digantdesai Differential Revision: D64081319 fbshipit-source-id: 4e8ff77af442dfded043c5a5583466afec6beb4e
Summary:
Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975
Was able to confirm although the problem happens in
ConvertToLinear
but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.The exported graph looks the same but the partitioner was behaving differently and causes the
ConvertToLinear
pass to error out.We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.
The current approach I'm taking avoids decomposing linear by using
to_edge_lower_and_transform
API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.Differential Revision: D64074891