#pragma once // @generated by torchgen/gen.py from Function.h #include #include #include #include #include #include #include #include #include #include #include #include #include namespace at { // aten::infinitely_differentiable_gelu_backward(Tensor grad, Tensor self) -> Tensor inline at::Tensor infinitely_differentiable_gelu_backward(const at::Tensor & grad, const at::Tensor & self) { return at::_ops::infinitely_differentiable_gelu_backward::call(grad, self); } }