Skip to content
This repository was archived by the owner on Aug 1, 2023. It is now read-only.

Conversation

@theweiho
Copy link
Contributor

Summary:
pytorch/pytorch#8899 had added CUDA support for torch.unique()

pytorch/pytorch#16145 has some timing stats that could be relevant


Experiment results: https://fb.quip.com/olQOA853j0mb
Words per second (gpu-unique_wps_avg_vs_base): 1.046x
Total train time (gpu-unique_total_train_time_vs_base; excl ar_AR-fr_XX): 0.987x

Even though train time reduction is pretty minimal (probably overshadowed by random variance, scheduling delay, etc), WPS does seem to be ~5% faster - so might as well land this.

Training time for ar_AR-fr_XX increased significantly - but that's b/c it trained for many more updates (gpu-unique_num_updates_avg_vs_base) - and also ended up w/ +1.43 BLEU. I think this is probably just an anomaly?

Differential Revision: D15073468

@theweiho theweiho force-pushed the export-D15073468 branch from b29dff7 to e2c9d87 Compare May 16, 2019 21:15
theweiho added a commit to theweiho/translate that referenced this pull request May 16, 2019
…torch#537)

Summary:
Pull Request resolved: pytorch#537

pytorch/pytorch#8899 had added CUDA support for `torch.unique()`

pytorch/pytorch#16145 has some timing stats that could be relevant

 ---

Experiment results: https://fb.quip.com/olQOA853j0mb
Words per second (`gpu-unique_wps_avg_vs_base`): 1.046x
Total train time (`gpu-unique_total_train_time_vs_base`; excl ar_AR-fr_XX): 0.987x

Even though train time reduction is pretty minimal (probably overshadowed by random variance, scheduling delay, etc), WPS does seem to be ~5% faster - so might as well land this.

Training time for ar_AR-fr_XX increased significantly - but that's b/c it trained for many more updates (`gpu-unique_num_updates_avg_vs_base`) - and also ended up w/ +1.43 BLEU. I think this is probably just an anomaly?

Differential Revision: D15073468

fbshipit-source-id: 713288fc7c77f582840f270dd2e343a3b63f8fe5
…torch#537)

Summary:
Pull Request resolved: pytorch#537

pytorch/pytorch#8899 had added CUDA support for `torch.unique()`

pytorch/pytorch#16145 has some timing stats that could be relevant

 ---

Experiment results: https://fb.quip.com/olQOA853j0mb
Words per second (`gpu-unique_wps_avg_vs_base`): 1.046x
Total train time (`gpu-unique_total_train_time_vs_base`; excl ar_AR-fr_XX): 0.987x

Even though train time reduction is pretty minimal (probably overshadowed by random variance, scheduling delay, etc), WPS does seem to be ~5% faster - so might as well land this.

Training time for ar_AR-fr_XX increased significantly - but that's b/c it trained for many more updates (`gpu-unique_num_updates_avg_vs_base`) - and also ended up w/ +1.43 BLEU. I think this is probably just an anomaly?

Differential Revision: D15073468

fbshipit-source-id: 29c7eaaddd63d629866c7314920fe27b22690603
@facebook-github-bot
Copy link

This pull request has been merged in 2abcc08.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants