#pragma once // ${generated_comment} #include #include #include #include #include #include // for size_t #include // for function #include // for unique_ptr #include #include namespace at { struct Quantizer; }; namespace torch { namespace autograd { using Variable = at::Tensor; using at::Context; using at::Device; using at::Dimname; using at::DimnameList; using at::Generator; using at::IntArrayRef; using at::MemoryFormat; using at::QScheme; using at::Scalar; using at::ScalarType; using at::Storage; using at::Tensor; using at::TensorList; using at::TensorOptions; using at::Quantizer; // This is temporary typedef to enable Quantizer in aten native function API // we'll remove them when we are actually exposing Quantizer class // to frontend using ConstQuantizerPtr = const c10::intrusive_ptr&; using c10::optional; namespace VariableType { TORCH_API std::vector allCUDATypes(); TORCH_API std::vector allXPUTypes(); TORCH_API std::vector allCPUTypes(); at::Tensor & unpack(Tensor & t, const char * name, int pos); const at::Tensor & unpack(const Tensor & t, const char * name, int pos); at::Tensor unpack_opt(const Tensor & t, const char * name, int pos); std::vector unpack(at::ITensorListRef tl, const char *name, int pos); }; }} // namespace torch::autograd