Skip to content

Conversation

jerryzh168
Copy link
Contributor

@jerryzh168 jerryzh168 commented Jan 31, 2024

Summary:
This is a follow up to #118605 to remove fold_quantize flag from
convert_pt2e

Test Plan: CI

Differential Revision: D53247301

BC Breaking Note:

flag fold_quantize set to True convert_pt2e and now we'll fold the quantize op in the weight by default, so users will see model size reduction by default after pt2e quantization.
2.2

folded_model = convert_pt2e(model, fold_quantize=True)

non_folded_model = convert_pt2e(model)

2.3

folded_model = convert_pt2e(model)

non_folded_model = convert_pt2e(model, fold_quantize=False)

cc @ezyang @gchanan @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @peterbell10 @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @aakhundov @ColinPeppler

Copy link

pytorch-bot bot commented Jan 31, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/118701

Note: Links to docs will display an error until the docs builds have been completed.

❌ 1 New Failure

As of commit df455cb with merge base 46ef735 (image):

NEW FAILURE - The following job has failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added the release notes: quantization release notes category label Jan 31, 2024
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D53247301

jerryzh168 added a commit to jerryzh168/executorch-1 that referenced this pull request Jan 31, 2024
Summary:
X-link: pytorch/pytorch#118701

This is a follow up to pytorch/pytorch#118605 to remove `fold_quantize` flag from
`convert_pt2e`

Differential Revision: D53247301
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D53247301

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Feb 1, 2024
Summary:
X-link: pytorch/executorch#1766


This is a follow up to pytorch#118605 to remove `fold_quantize` flag from
`convert_pt2e`

Test Plan: CI

Reviewed By: andrewor14

Differential Revision: D53247301
jerryzh168 added a commit to jerryzh168/executorch-1 that referenced this pull request Feb 1, 2024
Summary:

X-link: pytorch/pytorch#118701

This is a follow up to pytorch/pytorch#118605 to remove `fold_quantize` flag from
`convert_pt2e`

Reviewed By: andrewor14

Differential Revision: D53247301
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D53247301

@jerryzh168 jerryzh168 added the module: bc-breaking Related to a BC-breaking change label Feb 2, 2024
@jerryzh168 jerryzh168 added the suppress-bc-linter Suppresses the failures of API backward-compatibility linter (Lint/bc_linter) label Feb 2, 2024
facebook-github-bot pushed a commit to pytorch/executorch that referenced this pull request Feb 7, 2024
Summary:
Pull Request resolved: #1766

X-link: pytorch/pytorch#118701

This is a follow up to pytorch/pytorch#118605 to set `fold_quantize` flag to True in `convert_pt2e`

Reviewed By: andrewor14, digantdesai

Differential Revision: D53247301

fbshipit-source-id: 5b2dbbc76487a8779f30c483b5ff4f054ba1ae8c
@facebook-github-bot
Copy link
Contributor

@pytorchbot merge -f 'Landed internally'

(Initiating merge automatically since Phabricator Diff has merged, using force because this PR might not pass merge_rules.json but landed internally)

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged immediately since you used the force (-f) flag, bypassing any CI checks (ETA: 1-5 minutes). Please use -f as last resort and instead consider -i/--ignore-current to continue the merge ignoring current failures. This will allow currently pending tests to finish and report signal before the merge.

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@jerryzh168 jerryzh168 changed the title [quant][pt2e][bc-breaking] Remove fold_quantize flag [quant][pt2e][bc-breaking] Set fold_quantize to True in convert_pt2e Feb 7, 2024
@facebook-github-bot
Copy link
Contributor

@pytorchbot revert -m="Diff reverted internally" -c="ghfirst"

This Pull Request has been reverted by a revert inside Meta. To re-land this change, please open another pull request, assign the same reviewers, fix the CI failures that caused the revert and make sure that the failing CI runs on the PR by applying the proper ciflow label (e.g., ciflow/trunk).)

@pytorchmergebot
Copy link
Collaborator

@pytorchbot successfully started a revert job. Check the current status here.
Questions? Feedback? Please reach out to the PyTorch DevX Team

pytorchmergebot added a commit that referenced this pull request Feb 7, 2024
@pytorchmergebot
Copy link
Collaborator

@jerryzh168 your PR has been successfully reverted.

pytorch-bot bot pushed a commit that referenced this pull request Feb 8, 2024
Summary:
This is a follow up to #118605 to remove `fold_quantize` flag from
`convert_pt2e`

Test Plan: CI

Differential Revision: D53247301

BC Breaking Note:

flag `fold_quantize` set to True `convert_pt2e` and now we'll fold the quantize op in the weight by default, so users will see model size reduction by default after pt2e quantization.
2.2
```
folded_model = convert_pt2e(model, fold_quantize=True)

non_folded_model = convert_pt2e(model)
```

2.3
```
folded_model = convert_pt2e(model)

non_folded_model = convert_pt2e(model, fold_quantize=False)
```

Pull Request resolved: #118701
Approved by: https://github.com/andrewor14, https://github.com/leslie-fang-intel
pytorch-bot bot pushed a commit that referenced this pull request Feb 8, 2024
@jerryzh168 jerryzh168 closed this Feb 10, 2024
clee2000 pushed a commit that referenced this pull request Feb 14, 2024
Summary:
This is a follow up to #118605 to remove `fold_quantize` flag from
`convert_pt2e`

Test Plan: CI

Differential Revision: D53247301

BC Breaking Note:

flag `fold_quantize` set to True `convert_pt2e` and now we'll fold the quantize op in the weight by default, so users will see model size reduction by default after pt2e quantization.
2.2
```
folded_model = convert_pt2e(model, fold_quantize=True)

non_folded_model = convert_pt2e(model)
```

2.3
```
folded_model = convert_pt2e(model)

non_folded_model = convert_pt2e(model, fold_quantize=False)
```

Pull Request resolved: #118701
Approved by: https://github.com/andrewor14, https://github.com/leslie-fang-intel
clee2000 pushed a commit that referenced this pull request Feb 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ciflow/trunk Trigger trunk jobs on your pull request fb-exported Merged module: bc-breaking Related to a BC-breaking change module: inductor release notes: quantization release notes category Reverted suppress-bc-linter Suppresses the failures of API backward-compatibility linter (Lint/bc_linter)
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants