Skip to content

Conversation

@salilsdesai
Copy link
Contributor

Summary:
This version just does dequantize, fp32 softmax, quantize.
Another version of actual quantized softmax using qnnpack will be added next

Test Plan:
From fbcode:
buck test caffe2/test:quantization -- test_qsoftmax

Benchmarking: See summary of D34996486

Reviewed By: kimishpatel

Differential Revision: D34943147

Summary:
This version just does dequantize, fp32 softmax, quantize.
Another version of actual quantized softmax using qnnpack will be added next

Test Plan:
From fbcode:
```buck test caffe2/test:quantization -- test_qsoftmax```

Benchmarking: See summary of D34996486

Reviewed By: kimishpatel

Differential Revision: D34943147

fbshipit-source-id: dcebecd55e8da989ab4937c2b8db38baaf32b700
@facebook-github-bot
Copy link
Contributor

facebook-github-bot commented Mar 31, 2022

🔗 Helpful links

💊 CI failures summary and remediations

As of commit 1e34275 (more details on the Dr. CI page):


💚 💚 Looks good so far! There are no failures yet. 💚 💚


This comment was automatically generated by Dr. CI (expand for details).

Please report bugs/suggestions to the (internal) Dr. CI Users group.

Click here to manually regenerate this comment.

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D34943147

facebook-github-bot pushed a commit that referenced this pull request Mar 31, 2022
Summary:
Pull Request resolved: #75017

This version just does dequantize, fp32 softmax, quantize.
Another version of actual quantized softmax using qnnpack will be added next

Test Plan:
From fbcode:
```buck test caffe2/test:quantization -- test_qsoftmax```

Benchmarking: See summary of D34996486

Reviewed By: kimishpatel

Differential Revision: D34943147

fbshipit-source-id: 426a0780803597a21460139c67960891d6e9cc81
@github-actions
Copy link
Contributor

Hey @salilsdesai.
You've committed this PR, but it does not have both a 'release notes: ...' and 'topics: ...' label. Please add one of each to the PR. The 'release notes: ...' label should represent the part of PyTorch that this PR changes (fx, autograd, distributed, etc) and the 'topics: ...' label should represent the kind of PR it is (not user facing, new feature, bug fix, perf improvement, etc). The list of valid labels can be found here for the 'release notes: ...' and here for the 'topics: ...'.
For changes that are 'topic: not user facing' there is no need for a release notes label.

vkuzo added a commit that referenced this pull request Apr 1, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the eager mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`. A future PR will
add FX graph mode workflow integration.

Test plan:

```
python test/test_quantization.py TestQuantizeEagerPTQStatic.test_softmax
```

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 1, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 1, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 1, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 1, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 097b951
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 0af2cec
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 9eb7056
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 4, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: e555ffe
Pull Request resolved: #75106
@facebook-github-bot
Copy link
Contributor

This pull request has been reverted by 1bcae0d. To re-land this change, please open another pull request, assignthe same reviewers, fix the CI failures that caused the revert and make sure that the failing CI runs on the PR by applying the proper ciflow label (e.g., ciflow/trunk).

@facebook-github-bot
Copy link
Contributor

This pull request has been reverted by b99b7182aeaf77d07978708b3e1630260e342974. To re-land this change, please open another pull request, assignthe same reviewers, fix the CI failures that caused the revert and make sure that the failing CI runs on the PR by applying the proper ciflow label (e.g., ciflow/trunk).

vkuzo added a commit that referenced this pull request Apr 13, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 9ec762e
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 13, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 13, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 13, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 8dec27f
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 14, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 535666a
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 14, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 19, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Apr 19, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: 7d83ed1
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 20, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

ghstack-source-id: f45b660
Pull Request resolved: #75106
vkuzo added a commit that referenced this pull request Apr 20, 2022
Summary:

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test plan:

```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Differential Revision: [D35324817](https://our.internmc.facebook.com/intern/diff/D35324817)

[ghstack-poisoned]
facebook-github-bot pushed a commit that referenced this pull request Apr 20, 2022
Summary:
Pull Request resolved: #75106

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test Plan:
```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Reviewed By: kimishpatel, andrewor14

Differential Revision: D35324817

Pulled By: vkuzo

fbshipit-source-id: 710ae3bedf8a6ad1dc411cd9808fdd0ce743e757
pytorchmergebot pushed a commit that referenced this pull request Apr 20, 2022
Summary:
Pull Request resolved: #75106

In #75017 a quantized softmax
kernel was added. This PR adds the FX graph mode quantization workflow
integration to swap `nn.Softmax` to `nnq.Softmax`.

Test Plan:
```
python test/test_quantization.py TestQuantizeFxOps.test_fixed_qparams_ops
```

Reviewed By: kimishpatel, andrewor14

Differential Revision: D35324817

Pulled By: vkuzo

fbshipit-source-id: 710ae3bedf8a6ad1dc411cd9808fdd0ce743e757
(cherry picked from commit d67603c)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants