Skip to content

Commit

Permalink
embedding decomp
Browse files Browse the repository at this point in the history
  • Loading branch information
samdow committed Jun 16, 2022
1 parent 825f439 commit 3b68942
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
3 changes: 1 addition & 2 deletions functorch/_src/eager_transforms.py
Original file line number Diff line number Diff line change
Expand Up @@ -1339,7 +1339,6 @@ def _register_python_decomposition_vmap(decomp):
_register_jit_decomposition(torch.ops.aten._softmax_backward_data.default)
_register_jit_decomposition(torch.ops.aten.log_sigmoid_forward.default)
_register_jit_decomposition(torch.ops.aten.native_layer_norm_backward.default)
_register_jit_decomposition(torch.ops.aten.native_batch_norm_backward.default)
_register_jit_decomposition(torch.ops.aten.cudnn_batch_norm_backward.default)
_register_jit_decomposition(torch.ops.aten.embedding_dense_backward.default)
_register_python_decomposition_vmap(torch.ops.aten.mse_loss_backward.default)
_register_python_decomposition_vmap(torch.ops.aten.addr.default)
1 change: 1 addition & 0 deletions functorch/csrc/DynamicLayer.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -504,6 +504,7 @@ TORCH_LIBRARY_IMPL(aten, FT_DYNAMIC_LAYER_FRONT_MODE_KEY, m) {
JVP_DECOMP(native_layer_norm_backward);
JVP_DECOMP(native_batch_norm_backward);
JVP_DECOMP(cudnn_batch_norm_backward);
JVP_DECOMP(embedding_dense_backward);
}


Expand Down

0 comments on commit 3b68942

Please sign in to comment.