|
|
#pragma once |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#include <c10/core/MemoryFormat.h> |
|
|
#include <c10/core/Scalar.h> |
|
|
#include <ATen/core/Reduction.h> |
|
|
|
|
|
#if defined(AT_PER_OPERATOR_HEADERS) && defined(TORCH_ASSERT_ONLY_METHOD_OPERATORS) |
|
|
#error This change adds a dependency on all pytorch operators, meaning the \ |
|
|
file will need to be re-compiled every time an operator is changed or added. \ |
|
|
Consider including a specific operator from \ |
|
|
<ATen/ops/{my_operator}_compositeimplicitautogradnestedtensor_dispatch.h>. \ |
|
|
See NOTE [TORCH_ASSERT_ONLY_METHOD_OPERATORS]. |
|
|
#endif |
|
|
|
|
|
#include <ATen/ops/reshape_compositeimplicitautogradnestedtensor_dispatch.h> |
|
|
#include <ATen/ops/reshape_as_compositeimplicitautogradnestedtensor_dispatch.h> |
|
|
#include <ATen/ops/unbind_compositeimplicitautogradnestedtensor_dispatch.h> |
|
|
|
|
|
|
|
|
|
|
|
|