Skip to content

Conversation

@szagoruyko
Copy link
Contributor

also:

  • rename pad to padding for consistency
  • ConvTranspose2d used to take kw and kh which it didn't need

@soumith soumith merged commit 45d6212 into pytorch:master Dec 25, 2016
jjsjann123 pushed a commit to jjsjann123/pytorch that referenced this pull request Sep 23, 2020
Removing support for Kernel IR nodes from IrGraphGenerator
jjsjann123 pushed a commit to jjsjann123/pytorch that referenced this pull request Sep 24, 2020
Removing support for Kernel IR nodes from IrGraphGenerator
KyleCZH pushed a commit to KyleCZH/pytorch that referenced this pull request Sep 20, 2021
eellison pushed a commit to eellison/pytorch that referenced this pull request Jun 29, 2022
akashveramd pushed a commit to akashveramd/pytorch that referenced this pull request Apr 9, 2025
* initial stub for gemm_gemm_xdl_cshuffle

* set up example code

* compiles

* prevent integer overflow

* harmonize interface between ref_gemm and ref_batched_gemm

* batched_gemm_gemm

* fix example

* host tensor gen: diagonal pattern in lowest two-dimensions only

* make c descriptors containing only integral constants

* clean up

* add BlockwiseGemmXdlops_v2 while exploring an unified approach

* implement proper interface

* tidy up example

* fix compilation warnings

* coarsely controlled 2nd gemm padding

* remove rocm-cmake's hard requirement for certain revision

* clang-format

* resolve merge conflict

* fix compilation error on gfx10

* adds acc0 elementwise op to interface

* add gemm_gemm instances and tests

* avoid LDS data hazard

* fix build

Co-authored-by: Chao Liu <chao.liu2@amd.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants