Skip to content

Unbreak test models llama CI #6026

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 1 commit into from
Closed

Unbreak test models llama CI #6026

wants to merge 1 commit into from

Conversation

larryliu0820
Copy link
Contributor

Summary:
Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in ConvertToLinear but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the ConvertToLinear pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using to_edge_lower_and_transform API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Differential Revision: D64074891

Copy link

pytorch-bot bot commented Oct 8, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/6026

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 04b1dda with merge base b6e6d06 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Oct 8, 2024
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

larryliu0820 added a commit that referenced this pull request Oct 8, 2024
Summary:
Pull Request resolved: #6026

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Differential Revision: D64074891
facebook-github-bot pushed a commit that referenced this pull request Oct 8, 2024
Summary:

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Reviewed By: Jack-Khuu

Differential Revision: D64074891
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

1 similar comment
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

larryliu0820 added a commit that referenced this pull request Oct 8, 2024
Summary:
Pull Request resolved: #6026

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Reviewed By: Jack-Khuu

Differential Revision: D64074891
facebook-github-bot pushed a commit that referenced this pull request Oct 8, 2024
Summary:

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Reviewed By: Jack-Khuu

Differential Revision: D64074891
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

Summary:

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Reviewed By: digantdesai, Jack-Khuu, tugsbayasgalan

Differential Revision: D64074891
facebook-github-bot pushed a commit that referenced this pull request Oct 9, 2024
Summary:

Did a bunch of debugging on OSS CI:https://github.com/pytorch/executorch/actions/runs/11241297226/job/31252590975

Was able to confirm although the problem happens in `ConvertToLinear` but the root cause is we are partitioning the graph differently between these two pytorch nightly: dev20240916 and dev20240917.

The exported graph looks the same but the partitioner was behaving differently and causes the `ConvertToLinear` pass to error out.

We can't really revert back to dev20240916 nightly because it breaks other CI jobs, see #5987.

The current approach I'm taking avoids decomposing linear by using `to_edge_lower_and_transform` API. This avoids jumping into the rabbit hole of debugging the partitioning & tagging logic.

Reviewed By: digantdesai, Jack-Khuu, tugsbayasgalan

Differential Revision: D64074891
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

1 similar comment
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D64074891

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 72b3bb3.

qma added a commit to qma/executorch that referenced this pull request Oct 15, 2024
Summary:
Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models:

- Quantized ViT
- Quantized Mobilebert
- Quantized Emformer Predict
- Quantized Emformer Transcribe

Differential Revision: D64081319
qma added a commit to qma/executorch that referenced this pull request Oct 16, 2024
pytorch#6242)

Summary:

Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models:

- Quantized ViT
- Quantized Mobilebert
- Quantized Emformer Predict
- Quantized Emformer Transcribe

Reviewed By: digantdesai

Differential Revision: D64081319
qma added a commit to qma/executorch that referenced this pull request Oct 16, 2024
pytorch#6242)

Summary:

Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models:

- Quantized ViT
- Quantized Mobilebert
- Quantized Emformer Predict
- Quantized Emformer Transcribe

Reviewed By: digantdesai

Differential Revision: D64081319
qma added a commit to qma/executorch that referenced this pull request Oct 16, 2024
pytorch#6242)

Summary:

Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change pytorch#6026, quantization XNNPack options can be enabled by default for the following models:

- Quantized ViT
- Quantized Mobilebert
- Quantized Emformer Predict
- Quantized Emformer Transcribe

Reviewed By: digantdesai

Differential Revision: D64081319
facebook-github-bot pushed a commit that referenced this pull request Oct 17, 2024
#6242)

Summary:
Pull Request resolved: #6242

Since master has migrated aot_compiler to use to_edge_transform_and_lower in a previous change #6026, quantization XNNPack options can be enabled by default for the following models:

- Quantized ViT
- Quantized Mobilebert
- Quantized Emformer Predict
- Quantized Emformer Transcribe

Reviewed By: digantdesai

Differential Revision: D64081319

fbshipit-source-id: 4e8ff77af442dfded043c5a5583466afec6beb4e
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported Merged
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants