-
Notifications
You must be signed in to change notification settings - Fork 26.3k
[quant] Add benchmarks for quantized embeddingbag module #43296
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Summary: Use common config for float and quantized embedding_bag modules Test Plan: ``` python -m pt.qembeddingbag_test Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 35.738 Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 62.708 python -m pt.embeddingbag_test Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 46.878 Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 103.904 ``` Reviewers: Subscribers: Tasks: Tags: [ghstack-poisoned]
Summary: Use common config for float and quantized embedding_bag modules Test Plan: ``` python -m pt.qembeddingbag_test Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 35.738 Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 62.708 python -m pt.embeddingbag_test Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 46.878 Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 103.904 ``` Reviewers: Subscribers: Tasks: Tags: ghstack-source-id: 2644726 Pull Request resolved: #43296
💊 CI failures summary and remediationsAs of commit 77a941e (more details on the Dr. CI page): ✅ None of the CI failures appear to be your fault 💚
🚧 1 ongoing upstream failure:These were probably caused by upstream breakages that are not fixed yet:
This comment was automatically generated by Dr. CI (expand for details).Follow this link to opt-out of these comments for your Pull Requests.Please report bugs/suggestions on the GitHub issue tracker or post in the (internal) Dr. CI Users group. This comment has been revised 4 times. |
vkuzo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for merging it with floating point!
| embedding_dim=dim, | ||
| mode=mode, | ||
| include_last_offset=include_last_offset).to(device=device) | ||
| numpy.random.seed((1 << 32) - 1) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
just curious, what's the context on this line?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It was used in the embeddingbag_test. So I did the same here for consistency
Summary: Use common config for float and quantized embedding_bag modules Test Plan: ``` python -m pt.qembeddingbag_test Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 35.738 Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 62.708 python -m pt.embeddingbag_test Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 46.878 Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 103.904 ``` Reviewers: Subscribers: Tasks: Tags: [ghstack-poisoned]
Summary: Use common config for float and quantized embedding_bag modules Test Plan: ``` python -m pt.qembeddingbag_test Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 35.738 Benchmarking PyTorch: qEmbeddingBag Mode: Eager Name: qEmbeddingBag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 62.708 python -m pt.embeddingbag_test Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetTrue_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: True, device: cpu Forward Execution Time (us) : 46.878 Benchmarking PyTorch: embeddingbag Mode: Eager Name: embeddingbag_embeddingbags10_dim4_modesum_input_size8_offset0_sparseTrue_include_last_offsetFalse_cpu Input: embeddingbags: 10, dim: 4, mode: sum, input_size: 8, offset: 0, sparse: True, include_last_offset: False, device: cpu Forward Execution Time (us) : 103.904 ``` Reviewers: Subscribers: Tasks: Tags: ghstack-source-id: 079cf2f Pull Request resolved: #43296
|
This pull request has been merged in 7024ce8. |
Stack from ghstack:
Summary:
Use common config for float and quantized embedding_bag modules
Test Plan:
Reviewers:
Subscribers:
Tasks:
Tags:
Differential Revision: D23245531