Skip to content

Instantly share code, notes, and snippets.

import torch
self_tensors = [torch.rand(2, 3, requires_grad=True),
torch.rand(4, 5, requires_grad=True)]
self_tensors = [torch.rand(2, 3),
torch.rand(4, 5)]
flat_tensors = []
for tensor in self_tensors:
['__abs__',
'__add__',
'__and__',
'__array__',
'__array_priority__',
'__array_wrap__',
'__bool__',
'__class__',
'__contains__',
'__cuda_array_interface__',
aten::_cudnn_rnn_backward(Tensor input, Tensor[] weight, int weight_stride0, Tensor weight_buf, Tensor hx, Tensor? cx, Tensor output, Tensor? grad_output, Tensor? grad_hy, Tensor? grad_cy, int mode, int hidden_size, int num_layers, bool batch_first, float dropout, bool train, bool bidirectional, int[] batch_sizes, Tensor? dropout_state, Tensor reserve, bool[4] output_mask) -> (Tensor, Tensor, Tensor, Tensor[]) aten::_cudnn_rnn_backward(Tensor input, Tensor[] weight, int weight_stride0, Tensor weight_buf, Tensor hx, Tensor? cx, Tensor output, Tensor? grad_output, Tensor? grad_hy, Tensor? grad_cy, int mode, int hidden_size, int num_layers, bool batch_first, float dropout, bool train, bool bidirectional, int[] batch_sizes, BoolTensor? dropout_state, Tensor reserve, bool[4] output_mask) -> (Tensor, Tensor, Tensor, Tensor[])
aten::_embedding_bag_dense_backward(Tensor grad, Tensor indices, Tensor offsets, Tensor offset2bag, Tensor bag_size, Tensor maximum_indices, int num_weights, bool scale_grad_by_freq, int mode)
diff --git a/build/aten/src/ATen/Declarations.yaml b/build/aten/src/ATen/Declarations.yaml
index 31207af18..c36fd7cd5 100644
--- a/build/aten/src/ATen/Declarations.yaml
+++ b/build/aten/src/ATen/Declarations.yaml
@@ -23590,7 +23590,7 @@
with_gil: false
deprecated: false
- name: _copy_same_type_
- matches_jit_signature: false
+ matches_jit_signature: true
diff --git a/build/aten/src/ATen/Declarations.yaml b/build/aten/src/ATen/Declarations.yaml
index 7ca5b4659..f6b91b7cc 100644
--- a/build/aten/src/ATen/Declarations.yaml
+++ b/build/aten/src/ATen/Declarations.yaml
@@ -25896,8 +25896,8 @@
with_gil: false
deprecated: false
- name: embedding_backward
- matches_jit_signature: false
- schema_string: aten::embedding_backward(Tensor grad, IndexTensor indices, int num_weights,
diff --git a/build/aten/src/ATen/Declarations.yaml b/build/aten/src/ATen/Declarations.yaml
index 7ca5b4659..249b9a9b5 100644
--- a/build/aten/src/ATen/Declarations.yaml
+++ b/build/aten/src/ATen/Declarations.yaml
@@ -25896,8 +25896,8 @@
with_gil: false
deprecated: false
- name: embedding_backward
- matches_jit_signature: false
- schema_string: aten::embedding_backward(Tensor grad, IndexTensor indices, int num_weights,
diff --git a/torch/csrc/jit/generated/register_aten_ops_0.cpp b/torch/csrc/jit/generated/register_aten_ops_0.cpp
index 5a9c7d601..b8b4d86cc 100644
--- a/torch/csrc/jit/generated/register_aten_ops_0.cpp
+++ b/torch/csrc/jit/generated/register_aten_ops_0.cpp
@@ -601,7 +601,7 @@ RegisterOperators reg({
}
),
Operator(
- "aten::_sparse_coo_tensor_with_dims_and_tensors(int sparse_dim, int dense_dim, int[] size, Tensor indices, Tensor values, *, ScalarType dtype=float, Layout layout=strided, Device device=\"cpu\") -> Tensor",
+ "aten::_sparse_coo_tensor_with_dims_and_tensors(int sparse_dim, int dense_dim, int[] size, Tensor indices, Tensor values, *, ScalarType dtype, Layout layout, Device device) -> Tensor",
diff --git a/build/aten/src/ATen/Declarations.yaml b/build/aten/src/ATen/Declarations.yaml
index 7ca5b4659..e3281b884 100644
--- a/build/aten/src/ATen/Declarations.yaml
+++ b/build/aten/src/ATen/Declarations.yaml
@@ -18798,9 +18798,9 @@
with_gil: false
deprecated: false
- name: _cudnn_init_dropout_state
- matches_jit_signature: false
+ matches_jit_signature: true
@cpuhrsch
cpuhrsch / -
Created February 27, 2019 23:11
diff --git a/build/aten/src/ATen/Declarations.yaml b/build/aten/src/ATen/Declarations.yaml
index 7ca5b4659..852e2cade 100644
--- a/build/aten/src/ATen/Declarations.yaml
+++ b/build/aten/src/ATen/Declarations.yaml
@@ -18800,7 +18800,7 @@
- name: _cudnn_init_dropout_state
matches_jit_signature: false
schema_string: aten::_cudnn_init_dropout_state(float dropout, bool train, int dropout_seed,
- TensorOptions options) -> Tensor
+ ScalarType dtype, Layout layout, Device device) -> Tensor
@cpuhrsch
cpuhrsch / -
Created February 26, 2019 22:47
This file has been truncated, but you can view the full file.
diff --git a/build/aten/src/ATen/CPUByteType.cpp b/build/aten/src/ATen/CPUByteType.cpp
index d50b87d1e..50c08d9f5 100644
--- a/build/aten/src/ATen/CPUByteType.cpp
+++ b/build/aten/src/ATen/CPUByteType.cpp
@@ -2603,9 +2603,9 @@ Tensor CPUByteType::histc(const Tensor & self, int64_t bins, Scalar min, Scalar
const OptionalDeviceGuard device_guard(device_of(self));
return at::native::_histc_cpu(/* actuals */ self, bins, min, max);
}
-Tensor & CPUByteType::adaptive_avg_pool2d_out(Tensor & output, const Tensor & self, IntArrayRef output_size) const {
+Tensor & CPUByteType::adaptive_avg_pool2d_out(Tensor & out, const Tensor & self, IntArrayRef output_size) const {