Skip to content

Files

Latest commit

60d62a4 · Oct 16, 2023

History

History
22540 lines (22352 loc) · 1.82 MB

torch_npu_apis.md

File metadata and controls

22540 lines (22352 loc) · 1.82 MB

API of PyTorch Ascend Adapter

PyTorch Ascend Adapter named torch_npu is committed to maintaining compatibility with PyTorch's native API so that users can better use torch_npu. The current API comparison between torch_npu and PyTorch is as follows:

NOTE:

  1. Marked as compatible, please note if there are any limitations.
  2. Not indicating whether it is compatible means that the compatibility of this API has not been tested and will be improved in the future.

torch

PyTorch API

Compatibility

Limitations

torch.default_generator

     

torch.SymInt

Y

  

torch.SymFloat

Y

  

torch.SymBool

Y

  

torch.Tag

Y

  

torch.Tag.name

     

torch.is_tensor

Y

  

torch.is_storage

Y

  

torch.is_complex

Y

  

torch.is_conj

     

torch.is_floating_point

Y

  

torch.is_nonzero

Y

  

torch.set_default_dtype

Y

  

torch.get_default_dtype

Y

  

torch.set_default_device

     

torch.set_default_tensor_type

Y

  

torch.numel

Y

  

torch.set_printoptions

Y

  

torch.set_flush_denormal

Y

  

torch.tensor

Y

  

torch.sparse_coo_tensor

     

torch.sparse_csr_tensor

     

torch.sparse_csc_tensor

     

torch.sparse_bsr_tensor

     

torch.sparse_bsc_tensor

     

torch.asarray

Y

  

torch.as_tensor

Y

  

torch.as_strided

Y

  

torch.from_numpy

Y

  

torch.from_dlpack

     

torch.frombuffer

Y

  

torch.zeros

Y

Only support fp16,fp32,int64,bool

torch.zeros_like

Y

Only support fp16,fp32,int64,bool

torch.ones

Y

Only support fp16,fp32,int64

torch.ones_like

Y

Only support fp16,fp32,int64

torch.arange

Y

Only support fp16,fp32,int64,bool

torch.range

Y

Only support fp16,fp32,int64,bool

torch.linspace

     

torch.logspace

     

torch.eye

Y

Only support fp16,fp32,int64,bool

torch.empty

Y

  

torch.empty_like

Y

Only support fp16,fp32,int64,bool

torch.empty_strided

Y

Only support fp16,fp32,int64,bool

torch.full

Y

Only support fp16,fp32,int64,bool

torch.full_like

     

torch.quantize_per_tensor

     

torch.quantize_per_channel

     

torch.dequantize

     

torch.complex

     

torch.polar

     

torch.heaviside

     

torch.adjoint

     

torch.argwhere

Y

Only support fp16,fp32,int64,bool

torch.cat

Y

Only support fp16,fp32,int64,bool

torch.concat

Y

Only support fp16,fp32,int64,bool

torch.concatenate

Y

Only support fp16,fp32,int64,bool

torch.conj

Y

Only support fp16,fp32,int64,bool

torch.chunk

Y

Only support fp16,fp32,int64,bool

torch.dsplit

Y

Only support fp16,fp32,int64,bool

torch.column_stack

     

torch.dstack

Y

Only support fp16,fp32,int64,bool

torch.gather

Y

Only support fp16,fp32,int64,bool

torch.hsplit

Y

Only support fp16,fp32,int64,bool

torch.hstack

Y

Only support fp16,fp32,int64,bool

torch.index_add

Y

Only support fp16,fp32,int64,bool

torch.index_copy

Y

Only support fp16,fp32,int64,bool

torch.index_reduce

Y

Only support fp16,fp32,int64,bool

torch.index_select

     

torch.masked_select

Y

Only support fp16,fp32,int64,bool

torch.movedim

     

torch.moveaxis

Y

  

torch.narrow

Y

Only support fp16,fp32,int64,bool

torch.narrow_copy

Y

Only support fp16,fp32,int64,bool

torch.nonzero

Y

Only support fp16,fp32,int64,bool

torch.permute

Y

Only support fp16,fp32,int64,bool

torch.reshape

     

torch.row_stack

Y

Only support fp16,fp32,int64,bool

torch.select

     

torch.scatter

Y

Only support fp16,fp32,int64,bool

torch.diagonal_scatter

Y

Only support fp16,fp32,int64,bool

torch.select_scatter

Y

Only support fp16,fp32,int64,bool

torch.slice_scatter

     

torch.scatter_add

     

torch.scatter_reduce

     

torch.split

Y

Only support fp16,fp32,int64,bool

torch.squeeze

Y

Only support fp16,fp32,int64

torch.stack

Y

Only support fp16,fp32,int64,bool

torch.swapaxes

     

torch.swapdims

     

torch.t

Y

  

torch.take

Y

Only support fp16,fp32,int64,bool

torch.take_along_dim

Y

Only support fp16,fp32,int64,bool

torch.tensor_split

Y

Only support fp16,fp32,int64,bool

torch.tile

Y

Only support fp16,fp32,int64,bool

torch.transpose

Y

Only support fp16,fp32,int64,bool

torch.unbind

     

torch.unsqueeze

Y

Only support fp16,fp32,int64,bool

torch.vsplit

Y

Only support fp16,fp32,int64,bool

torch.vstack

Y

Only support fp16,fp32,int64,bool

torch.where

Y

Only support fp16,fp32,int64,bool;Don`t support 8-dimensional shape

torch.Generator

Y

  

torch.Generator.device

     

torch.Generator.get_state

Y

  

torch.Generator.initial_seed

Y

  

torch.Generator.manual_seed

Y

  

torch.Generator.seed

Y

  

torch.Generator.set_state

Y

  

torch.seed

Y

  

torch.manual_seed

Y

  

torch.initial_seed

Y

  

torch.get_rng_state

Y

  

torch.set_rng_state

Y

  

torch.bernoulli

Y

Only support fp16,fp32,int64,bool

torch.multinomial

Y

Only support fp16,fp32,int64,bool

torch.normal

Y

Only support fp16,fp32,int64,bool

torch.poisson

     

torch.rand

     

torch.rand_like

Y

Only support fp16,fp32,int64,bool

torch.randint

     

torch.randint_like

Y

Only support fp16,fp32,int64,bool

torch.randn

     

torch.randn_like

Y

Only support fp16,fp32,int64,bool

torch.randperm

Y

Only support fp16,fp32,int64,bool

torch.quasirandom.SobolEngine

     

torch.quasirandom.SobolEngine.draw

     

torch.quasirandom.SobolEngine.draw_base2

     

torch.quasirandom.SobolEngine.fast_forward

     

torch.quasirandom.SobolEngine.reset

     

torch.save

Y

  

torch.load

Y

  

torch.get_num_threads

Y

  

torch.set_num_threads

Y

  

torch.get_num_interop_threads

Y

  

torch.set_num_interop_threads

Y

  

torch.no_grad

Y

  

torch.enable_grad

Y

  

torch.set_grad_enabled

Y

  

torch.is_grad_enabled

Y

  

torch.inference_mode

Y

  

torch.is_inference_mode_enabled

Y

  

torch.abs

Y

Only support fp16,fp32,int64,bool

torch.absolute

Y

Only support fp16,fp32,int64,bool

torch.acos

Y

Only support fp16,fp32,int64,bool

torch.arccos

Y

Only support fp16,fp32,int64,bool

torch.acosh

Y

Only support fp16,fp32

torch.arccosh

Y

Only support fp16,fp32

torch.add

Y

Only support fp16,fp32,int64,bool

torch.addcdiv

Y

Only support fp16,fp32,int64,bool

torch.addcmul

Y

Only support fp16,fp32,int64,bool

torch.angle

Y

Only support fp16,fp32,int64,bool

torch.asin

Y

Only support fp16,fp32,int64,bool

torch.arcsin

     

torch.asinh

Y

Only support fp16,fp32,int64,bool

torch.arcsinh

Y

Only support fp16,fp32,int64,bool

torch.atan

Y

Only support fp16,fp32,int64,bool

torch.arctan

Y

Only support fp16,fp32,int64,bool

torch.atanh

Y

Only support fp16,fp32,int64,bool

torch.arctanh

Y

Only support fp16,fp32,int64,bool

torch.atan2

     

torch.arctan2

Y

Only support fp16,fp32,int64,bool

torch.bitwise_not

Y

Only support fp16,fp32,int64,bool

torch.bitwise_and

Y

Only support fp16,fp32,int64,bool

torch.bitwise_or

Y

Only support fp16,fp32,int64,bool

torch.bitwise_xor

Y

Only support fp16,fp32,int64,bool

torch.bitwise_left_shift

     

torch.bitwise_right_shift

     

torch.ceil

Y

Only support fp16,fp32,int64,bool

torch.clamp

Y

Only support fp16,fp32,int64,bool

torch.clip

Y

Only support fp16,fp32,int64,bool

torch.conj_physical

     

torch.copysign

     

torch.cos

Y

Only support fp16,fp32,int64,bool

torch.cosh

Y

Only support fp16,fp32

torch.deg2rad

Y

Only support fp16,fp32,int64,bool

torch.div

Y

Only support fp16,fp32,int64,bool

torch.divide

Y

Only support fp16,fp32,int64,bool

torch.digamma

     

torch.erf

Y

Only support fp16,fp32,int64,bool

torch.erfc

Y

Only support fp16,fp32,int64,bool

torch.erfinv

Y

Only support fp16,fp32,int64,bool

torch.exp

Y

Only support fp16,fp32,int64,bool

torch.exp2

Y

Only support fp16,fp32,int64,bool

torch.expm1

Y

Only support fp16,fp32,int64,bool

torch.fake_quantize_per_channel_affine

     

torch.fake_quantize_per_tensor_affine

     

torch.fix

Y

Only support fp16,fp32,int64,bool

torch.float_power

Y

Only support fp16,fp32,int64,bool

torch.floor

Y

Only support fp16,fp32,int64,bool

torch.floor_divide

Y

Only support fp16,fp32,int64,bool

torch.fmod

Y

Only support fp16,fp32,int64,bool

torch.frac

     

torch.frexp

     

torch.gradient

     

torch.imag

     

torch.ldexp

Y

Only support fp16,fp32,int64,bool

torch.lerp

Y

Only support fp16,fp32,int64,bool

torch.lgamma

     

torch.log

Y

Only support fp16,fp32,int64,bool

torch.log10

     

torch.log1p

Y

Only support fp16,fp32,int64,bool

torch.log2

Y

Only support fp16,fp32,int64,bool

torch.logaddexp

Y

Only support fp16,fp32,int64,bool

torch.logaddexp2

Y

Only support fp16,fp32,int64,bool

torch.logical_and

Y

Only support fp16,fp32,int64,bool

torch.logical_not

Y

Only support fp16,fp32,int64,bool

torch.logical_or

Y

Only support fp16,fp32,int64,bool

torch.logical_xor

Y

Only support fp16,fp32,int64,bool

torch.logit

     

torch.hypot

     

torch.i0

     

torch.igamma

     

torch.igammac

     

torch.mul

Y

Only support fp16,fp32,int64,bool

torch.multiply

Y

Only support fp16,fp32,int64,bool

torch.mvlgamma

     

torch.nan_to_num

     

torch.neg

Y

Only support fp16,fp32,int64,bool

torch.negative

Y

Only support fp16,fp32,int64,bool

torch.nextafter

     

torch.polygamma

     

torch.positive

Y

Only support fp16,fp32,int64,bool

torch.pow

     

torch.quantized_batch_norm

     

torch.quantized_max_pool1d

     

torch.quantized_max_pool2d

     

torch.rad2deg

Y

Only support fp16,fp32,int64,bool

torch.real

Y

Only support fp16,fp32,int64,bool

torch.reciprocal

Y

Only support fp16,fp32,int64,bool

torch.remainder

Y

Only support fp16,fp32,int64,bool

torch.round

Y

Only support fp16,fp32,int64,bool

torch.rsqrt

Y

Only support fp16,fp32,int64,bool

torch.sigmoid

Y

Only support fp16,fp32,int64,bool

torch.sign

Y

Only support fp16,fp32,int64,bool

torch.sgn

Y

Only support fp16,fp32,int64,bool

torch.signbit

     

torch.sin

Y

Only support fp16,fp32,int64,bool

torch.sinc

     

torch.sinh

Y

Only support fp16,fp32

torch.softmax

Y

Only support fp16,fp32,int64,bool

torch.sqrt

Y

Only support fp16,fp32

torch.square

Y

Only support fp16,fp32,int64

torch.sub

Y

Only support fp16,fp32,int64

torch.subtract

     

torch.tan

Y

Only support fp16,fp32,int64

torch.tanh

Y

Only support fp16,fp32,int64

torch.true_divide

Y

Only support fp16,fp32,int64,bool

torch.trunc

Y

Only support fp16,fp32,int64,bool

torch.xlogy

Y

Only support fp16,fp32

torch.argmax

Y

Only support fp16,fp32,int64

torch.argmin

Y

Only support fp16,fp32,int64,bool

torch.amax

Y

Only support fp16,fp32,int64,bool

torch.amin

Y

Only support fp16,fp32,int64,bool

torch.aminmax

Y

Only support fp16,fp32,int64,bool

torch.all

Y

Only support fp16,fp32,int64,bool

torch.any

Y

Only support fp16,fp32,int64,bool

torch.max

Y

Only support fp16,fp32,int64,bool

torch.min

Y

Only support fp16,fp32,int64,bool

torch.dist

     

torch.logsumexp

Y

Only support fp16,fp32,int64,bool

torch.mean

Y

Only support fp16,fp32,int64,bool

torch.nanmean

     

torch.median

Y

Only support fp16,fp32,int64

torch.nanmedian

     

torch.mode

     

torch.norm

     

torch.nansum

     

torch.prod

Y

Only support fp16,fp32,int64,bool

torch.quantile

     

torch.nanquantile

     

torch.std

Y

Only support fp16,fp32

torch.std_mean

     

torch.sum

Y

Only support fp16,fp32,int64,bool;Don`t support dtype parameter

torch.unique

Y

Only support fp16,fp32,int64,bool

torch.unique_consecutive

     

torch.var

Y

Only support fp16,fp32

torch.var_mean

     

torch.count_nonzero

Y

Only support fp16,fp32,int64,bool

torch.allclose

Y

Only support fp16,fp32,int64,bool

torch.argsort

Y

Only support fp16,fp32,int64,bool

torch.eq

Y

Only support fp16,fp32,int64,bool

torch.equal

Y

Only support fp16,fp32,int64,bool

torch.ge

Y

Only support fp16,fp32,int64,bool

torch.greater_equal

Y

Only support fp16,fp32,int64,bool

torch.gt

Y

Only support fp16,fp32,int64,bool

torch.greater

Y

Only support fp16,fp32,int64,bool

torch.isclose

Y

Only support fp16,fp32,int64,bool

torch.isfinite

Y

Only support fp16,fp32,int64,bool

torch.isin

     

torch.isinf

Y

Only support fp16,fp32,int64,bool

torch.isposinf

Y

Only support fp16,fp32,int64,bool

torch.isneginf

Y

Only support fp16,fp32,int64,bool

torch.isnan

Y

Only support fp16,fp32,int64,bool

torch.isreal

     

torch.kthvalue

Y

Only support fp16,fp32

torch.le

Y

Only support fp16,fp32,int64,bool

torch.less_equal

Y

Only support fp16,fp32,int64,bool

torch.lt

Y

Only support fp16,fp32,int64,bool

torch.less

Y

Only support fp16,fp32,int64,bool

torch.maximum

Y

Only support fp16,fp32,int64,bool

torch.minimum

Y

Only support fp16,fp32,int64,bool

torch.fmax

     

torch.fmin

     

torch.ne

Y

Only support fp16,fp32,int64,bool

torch.not_equal

Y

Only support fp16,fp32,int64,bool

torch.sort

Y

Only support fp16,fp32,int64,bool

torch.topk

Y

Only support fp16,fp32,int64,bool;Don`t support sorted=False

torch.msort

Y

Only support fp16,fp32,int64,bool

torch.stft

     

torch.istft

     

torch.bartlett_window

     

torch.blackman_window

     

torch.hamming_window

     

torch.hann_window

     

torch.kaiser_window

     

torch.atleast_1d

     

torch.atleast_2d

     

torch.atleast_3d

     

torch.bincount

Y

Only support fp16,fp32,int64,bool

torch.block_diag

Y

Only support fp16,fp32,int64,bool

torch.broadcast_tensors

Y

Only support fp16,fp32,int64,bool

torch.broadcast_to

     

torch.broadcast_shapes

Y

Only support fp16,fp32,int64,bool

torch.bucketize

     

torch.cartesian_prod

     

torch.cdist

     

torch.clone

Y

Only support fp16,fp32,int64,bool

torch.combinations

Y

Only support fp16,fp32,int64,bool

torch.corrcoef

     

torch.cov

Y

Only support fp16,fp32,int64,bool

torch.cross

     

torch.cummax

Y

Only support fp16,fp32,int64,bool

torch.cummin

     

torch.cumprod

     

torch.cumsum

Y

Only support fp16,fp32,int64,bool

torch.diag

Y

Only support fp16,fp32,int64,bool

torch.diag_embed

     

torch.diagflat

     

torch.diagonal

Y

Only support fp16,fp32,int64,bool

torch.diff

Y

Only support fp16,fp32,int64,bool

torch.einsum

Y

Only support fp16,fp32,int64,bool

torch.flatten

Y

Only support fp16,fp32,int64,bool

torch.flip

Y

Only support fp16,fp32,int64,bool

torch.fliplr

Y

Only support fp16,fp32,int64,bool

torch.flipud

Y

Only support fp16,fp32,int64,bool

torch.kron

     

torch.rot90

     

torch.gcd

     

torch.histc

Y

Only support fp16,fp32,int64,bool

torch.histogram

     

torch.histogramdd

     

torch.meshgrid

Y

Only support fp16,fp32,int64,bool

torch.lcm

     

torch.logcumsumexp

     

torch.ravel

Y

Only support fp16,fp32,int64,bool

torch.renorm

     

torch.repeat_interleave

Y

Only support fp16,fp32,int64,bool

torch.roll

Y

Only support fp16,fp32,int64,bool

torch.searchsorted

Y

Only support fp16,fp32,int64

torch.tensordot

     

torch.trace

     

torch.tril

Y

Only support fp16,fp32,int64,bool

torch.tril_indices

Y

Only support fp16,fp32,int64,bool

torch.triu

Y

Only support fp16,fp32,int64,bool

torch.triu_indices

Y

Only support fp16,fp32,int64,bool

torch.unflatten

Y

  

torch.vander

     

torch.view_as_real

Y

Only support fp16,fp32,int64,bool

torch.view_as_complex

     

torch.resolve_conj

Y

Only support fp16,fp32,int64,bool

torch.resolve_neg

Y

Only support fp16,fp32,int64,bool

torch.addbmm

Y

Only support fp16,fp32,int64,bool

torch.addmm

Y

Only support fp16,fp32,int64,bool

torch.addmv

Y

Only support fp16,fp32,int64,bool

torch.addr

Y

Only support fp16,fp32,int64,bool

torch.baddbmm

Y

Only support fp16,fp32,int64,bool

torch.bmm

Y

Only support fp16,fp32,int64,bool

torch.chain_matmul

     

torch.cholesky

     

torch.cholesky_inverse

     

torch.cholesky_solve

     

torch.dot

Y

Only support fp16,fp32,int64,bool

torch.geqrf

     

torch.ger

     

torch.inner

     

torch.inverse

     

torch.det

     

torch.logdet

     

torch.slogdet

Y

Only support fp16,fp32,int64,bool

torch.lu

     

torch.lu_solve

     

torch.lu_unpack

     

torch.matmul

Y

Only support fp16,fp32,int64,bool

torch.matrix_power

     

torch.matrix_exp

     

torch.mm

Y

Only support fp16,fp32,int64,bool

torch.mv

     

torch.orgqr

     

torch.ormqr

     

torch.outer

     

torch.pinverse

     

torch.qr

Y

Only support fp16,fp32,int64,bool

torch.svd

     

torch.svd_lowrank

     

torch.pca_lowrank

     

torch.lobpcg

     

torch.trapz

     

torch.trapezoid

Y

Only support fp16,fp32,int64,bool

torch.cumulative_trapezoid

Y

Only support fp16,fp32,int64,bool

torch.triangular_solve

     

torch.vdot

     

torch._foreach_abs

     

torch._foreach_abs_

     

torch._foreach_acos

     

torch._foreach_acos_

     

torch._foreach_asin

     

torch._foreach_asin_

     

torch._foreach_atan

     

torch._foreach_atan_

     

torch._foreach_ceil

     

torch._foreach_ceil_

     

torch._foreach_cos

     

torch._foreach_cos_

     

torch._foreach_cosh

     

torch._foreach_cosh_

     

torch._foreach_erf

     

torch._foreach_erf_

     

torch._foreach_erfc

     

torch._foreach_erfc_

     

torch._foreach_exp

     

torch._foreach_exp_

     

torch._foreach_expm1

     

torch._foreach_expm1_

     

torch._foreach_floor

     

torch._foreach_floor_

     

torch._foreach_log

     

torch._foreach_log_

     

torch._foreach_log10

     

torch._foreach_log10_

     

torch._foreach_log1p

     

torch._foreach_log1p_

     

torch._foreach_log2

     

torch._foreach_log2_

     

torch._foreach_neg

     

torch._foreach_neg_

     

torch._foreach_tan

     

torch._foreach_tan_

     

torch._foreach_sin

     

torch._foreach_sin_

     

torch._foreach_sinh

     

torch._foreach_sinh_

     

torch._foreach_round

     

torch._foreach_round_

     

torch._foreach_sqrt

     

torch._foreach_sqrt_

     

torch._foreach_lgamma

     

torch._foreach_lgamma_

     

torch._foreach_frac

     

torch._foreach_frac_

     

torch._foreach_reciprocal

     

torch._foreach_reciprocal_

     

torch._foreach_sigmoid

     

torch._foreach_sigmoid_

     

torch._foreach_trunc

     

torch._foreach_trunc_

     

torch._foreach_zero_

     

torch.compiled_with_cxx11_abi

Y

  

torch.result_type

Y

  

torch.can_cast

Y

  

torch.promote_types

Y

  

torch.use_deterministic_algorithms

Y

  

torch.are_deterministic_algorithms_enabled

Y

  

torch.is_deterministic_algorithms_warn_only_enabled

     

torch.set_deterministic_debug_mode

Y

  

torch.get_deterministic_debug_mode

Y

  

torch.set_float32_matmul_precision

     

torch.get_float32_matmul_precision

     

torch.set_warn_always

Y

  

torch.is_warn_always_enabled

Y

  

torch.vmap

Y

  

torch._assert

Y

  

torch.sym_float

Y

  

torch.sym_int

Y

  

torch.sym_max

Y

  

torch.sym_min

Y

  

torch.sym_not

Y

  

torch.compile

Y

  

torch.nn

PyTorch API

Compatibility

Limitations

torch.nn.parameter.Parameter

Y

  

torch.nn.parameter.UninitializedParameter

Y

  

torch.nn.parameter.UninitializedParameter.cls_to_become

     

torch.nn.parameter.UninitializedBuffer

Y

  

torch.nn.Module

Y

  

torch.nn.Module.add_module

Y

  

torch.nn.Module.apply

Y

  

torch.nn.Module.bfloat16

Y

  

torch.nn.Module.buffers

Y

  

torch.nn.Module.children

Y

  

torch.nn.Module.compile

     

torch.nn.Module.cpu

Y

  

torch.nn.Module.cuda

Y

  

torch.nn.Module.double

     

torch.nn.Module.eval

Y

  

torch.nn.Module.extra_repr

Y

  

torch.nn.Module.float

Y

  

torch.nn.Module.forward

Y

  

torch.nn.Module.get_buffer

Y

  

torch.nn.Module.get_extra_state

Y

  

torch.nn.Module.get_parameter

Y

  

torch.nn.Module.get_submodule

Y

  

torch.nn.Module.half

Y

  

torch.nn.Module.ipu

     

torch.nn.Module.load_state_dict

Y

  

torch.nn.Module.modules

Y

  

torch.nn.Module.named_buffers

Y

  

torch.nn.Module.named_children

Y

  

torch.nn.Module.named_modules

Y

  

torch.nn.Module.named_parameters

Y

  

torch.nn.Module.parameters

Y

  

torch.nn.Module.register_backward_hook

Y

  

torch.nn.Module.register_buffer

Y

  

torch.nn.Module.register_forward_hook

Y

  

torch.nn.Module.register_forward_pre_hook

Y

  

torch.nn.Module.register_full_backward_hook

Y

  

torch.nn.Module.register_full_backward_pre_hook

Y

  

torch.nn.Module.register_load_state_dict_post_hook

Y

  

torch.nn.Module.register_module

Y

  

torch.nn.Module.register_parameter

Y

  

torch.nn.Module.register_state_dict_pre_hook

Y

  

torch.nn.Module.requires_grad_

Y

  

torch.nn.Module.set_extra_state

Y

  

torch.nn.Module.share_memory

     

torch.nn.Module.state_dict

Y

  

torch.nn.Module.to

Y

  

torch.nn.Module.to_empty

Y

  

torch.nn.Module.train

Y

  

torch.nn.Module.type

Y

  

torch.nn.Module.xpu

     

torch.nn.Module.zero_grad

Y

  

torch.nn.Sequential

Y

  

torch.nn.Sequential.append

Y

  

torch.nn.ModuleList

Y

  

torch.nn.ModuleList.append

Y

  

torch.nn.ModuleList.extend

Y

  

torch.nn.ModuleList.insert

Y

  

torch.nn.ModuleDict

Y

  

torch.nn.ModuleDict.clear

Y

  

torch.nn.ModuleDict.items

Y

  

torch.nn.ModuleDict.keys

Y

  

torch.nn.ModuleDict.pop

Y

  

torch.nn.ModuleDict.update

Y

  

torch.nn.ModuleDict.values

Y

  

torch.nn.ParameterList

Y

  

torch.nn.ParameterList.append

Y

  

torch.nn.ParameterList.extend

Y

  

torch.nn.ParameterDict

Y

  

torch.nn.ParameterDict.clear

Y

  

torch.nn.ParameterDict.copy

Y

  

torch.nn.ParameterDict.fromkeys

Y

  

torch.nn.ParameterDict.get

Y

  

torch.nn.ParameterDict.items

Y

  

torch.nn.ParameterDict.keys

Y

  

torch.nn.ParameterDict.pop

Y

  

torch.nn.ParameterDict.popitem

Y

  

torch.nn.ParameterDict.setdefault

Y

  

torch.nn.ParameterDict.update

Y

  

torch.nn.ParameterDict.values

Y

  

torch.nn.modules.module.register_module_forward_pre_hook

Y

  

torch.nn.modules.module.register_module_forward_hook

Y

  

torch.nn.modules.module.register_module_backward_hook

Y

  

torch.nn.modules.module.register_module_full_backward_pre_hook

     

torch.nn.modules.module.register_module_full_backward_hook

Y

  

torch.nn.modules.module.register_module_buffer_registration_hook

     

torch.nn.modules.module.register_module_module_registration_hook

     

torch.nn.modules.module.register_module_parameter_registration_hook

     

torch.nn.Conv1d

     

torch.nn.Conv2d

     

torch.nn.Conv3d

     

torch.nn.ConvTranspose1d

     

torch.nn.ConvTranspose2d

     

torch.nn.ConvTranspose3d

     

torch.nn.LazyConv1d

     

torch.nn.LazyConv1d.cls_to_become

     

torch.nn.LazyConv2d

     

torch.nn.LazyConv2d.cls_to_become

     

torch.nn.LazyConv3d

     

torch.nn.LazyConv3d.cls_to_become

     

torch.nn.LazyConvTranspose1d

     

torch.nn.LazyConvTranspose1d.cls_to_become

     

torch.nn.LazyConvTranspose2d

     

torch.nn.LazyConvTranspose2d.cls_to_become

     

torch.nn.LazyConvTranspose3d

     

torch.nn.LazyConvTranspose3d.cls_to_become

     

torch.nn.Unfold

     

torch.nn.Fold

     

torch.nn.MaxPool1d

     

torch.nn.MaxPool2d

     

torch.nn.MaxPool3d

     

torch.nn.MaxUnpool1d

Y

Only support fp16,fp32,int64,bool

torch.nn.MaxUnpool2d

Y

Only support fp16,fp32,int64,bool

torch.nn.MaxUnpool3d

     

torch.nn.AvgPool1d

     

torch.nn.AvgPool2d

     

torch.nn.AvgPool3d

     

torch.nn.FractionalMaxPool2d

     

torch.nn.FractionalMaxPool3d

     

torch.nn.LPPool1d

     

torch.nn.LPPool2d

     

torch.nn.AdaptiveMaxPool1d

     

torch.nn.AdaptiveMaxPool2d

     

torch.nn.AdaptiveMaxPool3d

Y

Only support fp16,fp32,int64,bool

torch.nn.AdaptiveAvgPool1d

Y

Only support fp16,fp32,int64,bool

torch.nn.AdaptiveAvgPool2d

Y

Only support fp16,fp32,int64,bool

torch.nn.AdaptiveAvgPool3d

     

torch.nn.ReflectionPad1d

Y

Only support fp16,fp32,int64,bool

torch.nn.ReflectionPad2d

Y

Only support fp16,fp32,int64,bool

torch.nn.ReflectionPad3d

     

torch.nn.ReplicationPad1d

Y

Only support fp16,fp32,int64,bool

torch.nn.ReplicationPad2d

Y

Only support fp16,fp32,int64,bool

torch.nn.ReplicationPad3d

     

torch.nn.ZeroPad1d

     

torch.nn.ZeroPad2d

Y

Only support fp16,fp32,int64,bool

torch.nn.ZeroPad3d

     

torch.nn.ConstantPad1d

Y

Only support fp16,fp32,int64,bool

torch.nn.ConstantPad2d

Y

Only support fp16,fp32,int64,bool

torch.nn.ConstantPad3d

     

torch.nn.ELU

     

torch.nn.Hardshrink

     

torch.nn.Hardsigmoid

Y

Only support fp16,fp32

torch.nn.Hardtanh

Y

Only support fp16,fp32,int64,bool

torch.nn.Hardswish

Y

Only support fp16,fp32

torch.nn.LeakyReLU

Y

Only support fp16,fp32

torch.nn.LogSigmoid

     

torch.nn.MultiheadAttention

     

torch.nn.MultiheadAttention.forward

     

torch.nn.MultiheadAttention.merge_masks

     

torch.nn.PReLU

Y

Only support fp16,fp32,int64,bool

torch.nn.ReLU

Y

Only support fp16,fp32,int64

torch.nn.ReLU6

Y

Only support fp16,fp32,int64,bool

torch.nn.RReLU

     

torch.nn.SELU

Y

Only support fp16,fp32,int64,bool

torch.nn.CELU

     

torch.nn.GELU

Y

Only support fp16,fp32,int64,bool

torch.nn.Sigmoid

Y

Only support fp16,fp32,int64,bool

torch.nn.SiLU

Y

Only support fp16,fp32,int64,bool

torch.nn.Mish

     

torch.nn.Softplus

Y

Only support fp16,fp32

torch.nn.Softshrink

Y

Only support fp16,fp32,int64,bool

torch.nn.Softsign

Y

Only support fp16,fp32,int64,bool

torch.nn.Tanh

Y

Only support fp16,fp32,int64

torch.nn.Tanhshrink

Y

Only support fp16,fp32,int64,bool

torch.nn.Threshold

Y

Only support fp16,fp32,int64

torch.nn.GLU

Y

Only support fp16,fp32,int64,bool

torch.nn.Softmin

Y

Only support fp16,fp32,int64,bool

torch.nn.Softmax

Y

Only support fp16,fp32,int64,bool

torch.nn.Softmax2d

Y

Only support fp16,fp32,int64,bool

torch.nn.LogSoftmax

Y

Only support fp16,fp32,int64,bool

torch.nn.AdaptiveLogSoftmaxWithLoss

     

torch.nn.AdaptiveLogSoftmaxWithLoss.log_prob

     

torch.nn.AdaptiveLogSoftmaxWithLoss.predict

     

torch.nn.BatchNorm1d

Y

Only support fp16,fp32,int64,bool

torch.nn.BatchNorm2d

Y

Only support fp16,fp32,int64,bool

torch.nn.BatchNorm3d

Y

Only support fp16,fp32,int64,bool

torch.nn.LazyBatchNorm1d

     

torch.nn.LazyBatchNorm1d.cls_to_become

     

torch.nn.LazyBatchNorm2d

     

torch.nn.LazyBatchNorm2d.cls_to_become

     

torch.nn.LazyBatchNorm3d

     

torch.nn.LazyBatchNorm3d.cls_to_become

     

torch.nn.GroupNorm

     

torch.nn.SyncBatchNorm

Y

Only support fp16,fp32,int64,bool

torch.nn.SyncBatchNorm.convert_sync_batchnorm

Y

  

torch.nn.InstanceNorm1d

     

torch.nn.InstanceNorm2d

     

torch.nn.InstanceNorm3d

     

torch.nn.LazyInstanceNorm1d

     

torch.nn.LazyInstanceNorm1d.cls_to_become

     

torch.nn.LazyInstanceNorm2d

     

torch.nn.LazyInstanceNorm2d.cls_to_become

     

torch.nn.LazyInstanceNorm3d

     

torch.nn.LazyInstanceNorm3d.cls_to_become

     

torch.nn.LayerNorm

Y

Only support fp16,fp32,int64,bool

torch.nn.LocalResponseNorm

     

torch.nn.RNNBase

     

torch.nn.RNNBase.flatten_parameters

     

torch.nn.RNN

     

torch.nn.LSTM

     

torch.nn.GRU

     

torch.nn.RNNCell

     

torch.nn.LSTMCell

     

torch.nn.GRUCell

     

torch.nn.Transformer

     

torch.nn.Transformer.forward

     

torch.nn.Transformer.generate_square_subsequent_mask

     

torch.nn.TransformerEncoder

     

torch.nn.TransformerEncoder.forward

     

torch.nn.TransformerDecoder

     

torch.nn.TransformerDecoder.forward

     

torch.nn.TransformerEncoderLayer

     

torch.nn.TransformerEncoderLayer.forward

     

torch.nn.TransformerDecoderLayer

     

torch.nn.TransformerDecoderLayer.forward

     

torch.nn.Identity

Y

  

torch.nn.Linear

Y

Only support fp16,fp32,int64,bool

torch.nn.Bilinear

Y

Only support fp16,fp32,int64,bool

torch.nn.LazyLinear

Y

Only support fp16,fp32,int64,bool

torch.nn.LazyLinear.cls_to_become

     

torch.nn.Dropout

     

torch.nn.Dropout1d

     

torch.nn.Dropout2d

Y

Only support fp16,fp32,int64,bool

torch.nn.Dropout3d

     

torch.nn.AlphaDropout

Y

Only support fp16,fp32,int64,bool

torch.nn.FeatureAlphaDropout

Y

Only support fp16,fp32,int64,bool

torch.nn.Embedding

Y

Only support fp16,fp32,int64,bool

torch.nn.Embedding.from_pretrained

Y

  

torch.nn.EmbeddingBag

     

torch.nn.EmbeddingBag.forward

Y

  

torch.nn.EmbeddingBag.from_pretrained

Y

  

torch.nn.CosineSimilarity

     

torch.nn.PairwiseDistance

     

torch.nn.L1Loss

Y

Only support fp16,fp32,int64,bool

torch.nn.MSELoss

Y

Only support fp16,fp32,int64,bool

torch.nn.CrossEntropyLoss

     

torch.nn.CTCLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.NLLLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.PoissonNLLLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.GaussianNLLLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.KLDivLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.BCELoss

     

torch.nn.BCEWithLogitsLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.MarginRankingLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.HingeEmbeddingLoss

Y

Only support fp16,fp32,int64,bool

torch.nn.MultiLabelMarginLoss

     

torch.nn.HuberLoss

     

torch.nn.SmoothL1Loss

Y

Only support fp16,fp32,int64,bool

torch.nn.SoftMarginLoss

     

torch.nn.MultiLabelSoftMarginLoss

Y

Only support fp16,fp32

torch.nn.CosineEmbeddingLoss

     

torch.nn.MultiMarginLoss

     

torch.nn.TripletMarginLoss

     

torch.nn.TripletMarginWithDistanceLoss

     

torch.nn.PixelShuffle

     

torch.nn.PixelUnshuffle

Y

Only support fp16,fp32,int64,bool

torch.nn.Upsample

Y

Only support fp16,fp32,int64,bool

torch.nn.UpsamplingNearest2d

Y

Only support fp16,fp32,int64,bool

torch.nn.UpsamplingBilinear2d

     

torch.nn.ChannelShuffle

     

torch.nn.DataParallel

     

torch.nn.parallel.DistributedDataParallel

Y

  

torch.nn.parallel.DistributedDataParallel.join

Y

  

torch.nn.parallel.DistributedDataParallel.join_hook

Y

  

torch.nn.parallel.DistributedDataParallel.no_sync

Y

  

torch.nn.parallel.DistributedDataParallel.register_comm_hook

Y

  

torch.nn.utils.clip_grad_norm_

     

torch.nn.utils.clip_grad_value_

Y

Only support fp16,fp32,int64,bool

torch.nn.utils.parameters_to_vector

     

torch.nn.utils.vector_to_parameters

     

torch.nn.utils.prune.BasePruningMethod

Y

  

torch.nn.utils.prune.BasePruningMethod.apply

Y

  

torch.nn.utils.prune.BasePruningMethod.apply_mask

Y

  

torch.nn.utils.prune.BasePruningMethod.compute_mask

     

torch.nn.utils.prune.BasePruningMethod.prune

Y

  

torch.nn.utils.prune.BasePruningMethod.remove

Y

  

torch.nn.utils.prune.PruningContainer

Y

  

torch.nn.utils.prune.PruningContainer.add_pruning_method

Y

  

torch.nn.utils.prune.PruningContainer.apply

Y

  

torch.nn.utils.prune.PruningContainer.apply_mask

Y

  

torch.nn.utils.prune.PruningContainer.compute_mask

Y

  

torch.nn.utils.prune.PruningContainer.prune

Y

  

torch.nn.utils.prune.PruningContainer.remove

Y

  

torch.nn.utils.prune.Identity

Y

  

torch.nn.utils.prune.Identity.apply

Y

  

torch.nn.utils.prune.Identity.apply_mask

Y

  

torch.nn.utils.prune.Identity.prune

Y

  

torch.nn.utils.prune.Identity.remove

Y

  

torch.nn.utils.prune.RandomUnstructured

Y

  

torch.nn.utils.prune.RandomUnstructured.apply

Y

  

torch.nn.utils.prune.RandomUnstructured.apply_mask

Y

  

torch.nn.utils.prune.RandomUnstructured.prune

Y

  

torch.nn.utils.prune.RandomUnstructured.remove

     

torch.nn.utils.prune.L1Unstructured

Y

  

torch.nn.utils.prune.L1Unstructured.apply

Y

  

torch.nn.utils.prune.L1Unstructured.apply_mask

Y

  

torch.nn.utils.prune.L1Unstructured.prune

Y

  

torch.nn.utils.prune.L1Unstructured.remove

Y

  

torch.nn.utils.prune.RandomStructured

Y

  

torch.nn.utils.prune.RandomStructured.apply

Y

  

torch.nn.utils.prune.RandomStructured.apply_mask

Y

  

torch.nn.utils.prune.RandomStructured.compute_mask

Y

  

torch.nn.utils.prune.RandomStructured.prune

Y

  

torch.nn.utils.prune.RandomStructured.remove

Y

  

torch.nn.utils.prune.LnStructured

Y

  

torch.nn.utils.prune.LnStructured.apply

Y

  

torch.nn.utils.prune.LnStructured.apply_mask

Y

  

torch.nn.utils.prune.LnStructured.compute_mask

Y

  

torch.nn.utils.prune.LnStructured.prune

Y

  

torch.nn.utils.prune.LnStructured.remove

Y

  

torch.nn.utils.prune.CustomFromMask

Y

  

torch.nn.utils.prune.CustomFromMask.apply

Y

  

torch.nn.utils.prune.CustomFromMask.apply_mask

Y

  

torch.nn.utils.prune.CustomFromMask.prune

Y

  

torch.nn.utils.prune.CustomFromMask.remove

Y

  

torch.nn.utils.prune.identity

Y

  

torch.nn.utils.prune.random_unstructured

Y

  

torch.nn.utils.prune.l1_unstructured

Y

  

torch.nn.utils.prune.random_structured

Y

  

torch.nn.utils.prune.ln_structured

Y

  

torch.nn.utils.prune.global_unstructured

Y

  

torch.nn.utils.prune.custom_from_mask

Y

  

torch.nn.utils.prune.remove

Y

  

torch.nn.utils.prune.is_pruned

Y

  

torch.nn.utils.weight_norm

     

torch.nn.utils.remove_weight_norm

     

torch.nn.utils.spectral_norm

     

torch.nn.utils.remove_spectral_norm

     

torch.nn.utils.skip_init

     

torch.nn.utils.parametrizations.orthogonal

     

torch.nn.utils.parametrizations.spectral_norm

     

torch.nn.utils.parametrize.register_parametrization

     

torch.nn.utils.parametrize.remove_parametrizations

     

torch.nn.utils.parametrize.cached

     

torch.nn.utils.parametrize.is_parametrized

     

torch.nn.utils.parametrize.ParametrizationList

Y

  

torch.nn.utils.parametrize.ParametrizationList.right_inverse

Y

  

torch.nn.utils.stateless.functional_call

     

torch.nn.utils.rnn.PackedSequence

Y

  

torch.nn.utils.rnn.PackedSequence.batch_sizes

     

torch.nn.utils.rnn.PackedSequence.count

Y

  

torch.nn.utils.rnn.PackedSequence.data

     

torch.nn.utils.rnn.PackedSequence.index

Y

  

torch.nn.utils.rnn.PackedSequence.is_cuda

     

torch.nn.utils.rnn.PackedSequence.is_pinned

Y

  

torch.nn.utils.rnn.PackedSequence.sorted_indices

     

torch.nn.utils.rnn.PackedSequence.to

Y

  

torch.nn.utils.rnn.PackedSequence.unsorted_indices

     

torch.nn.utils.rnn.pack_padded_sequence

     

torch.nn.utils.rnn.pad_packed_sequence

     

torch.nn.utils.rnn.pad_sequence

Y

Only support fp16,fp32,int64,bool

torch.nn.utils.rnn.pack_sequence

     

torch.nn.utils.rnn.unpack_sequence

     

torch.nn.utils.rnn.unpad_sequence

     

torch.nn.Flatten

Y

Only support fp16,fp32,int64,bool

torch.nn.Unflatten

Y

Only support fp16,fp32,int64,bool

torch.nn.modules.lazy.LazyModuleMixin

Y

  

torch.nn.modules.lazy.LazyModuleMixin.has_uninitialized_params

Y

  

torch.nn.modules.lazy.LazyModuleMixin.initialize_parameters

Y

  

torch.nn.functional

PyTorch API

Compatibility

Limitations

torch.nn.functional.conv1d

     

torch.nn.functional.conv2d

     

torch.nn.functional.conv3d

     

torch.nn.functional.conv_transpose1d

     

torch.nn.functional.conv_transpose2d

     

torch.nn.functional.conv_transpose3d

     

torch.nn.functional.unfold

     

torch.nn.functional.fold

     

torch.nn.functional.avg_pool1d

     

torch.nn.functional.avg_pool2d

     

torch.nn.functional.avg_pool3d

     

torch.nn.functional.max_pool1d

     

torch.nn.functional.max_pool2d

     

torch.nn.functional.max_pool3d

     

torch.nn.functional.max_unpool1d

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.max_unpool2d

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.max_unpool3d

     

torch.nn.functional.lp_pool1d

     

torch.nn.functional.lp_pool2d

     

torch.nn.functional.adaptive_max_pool1d

     

torch.nn.functional.adaptive_max_pool2d

     

torch.nn.functional.adaptive_max_pool3d

     

torch.nn.functional.adaptive_avg_pool1d

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.adaptive_avg_pool2d

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.adaptive_avg_pool3d

     

torch.nn.functional.fractional_max_pool2d

     

torch.nn.functional.fractional_max_pool3d

     

torch.nn.functional.scaled_dot_product_attention

     

torch.nn.functional.threshold

Y

Only support fp16,fp32,int64

torch.nn.functional.threshold_

Y

Only support fp16,fp32,int64

torch.nn.functional.relu

Y

Only support fp16,fp32,int64

torch.nn.functional.relu_

Y

Only support fp16,fp32,int64

torch.nn.functional.hardtanh

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.hardtanh_

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.hardswish

Y

Only support fp16,fp32

torch.nn.functional.relu6

     

torch.nn.functional.elu

     

torch.nn.functional.elu_

     

torch.nn.functional.selu

     

torch.nn.functional.celu

     

torch.nn.functional.leaky_relu

Y

Only support fp16,fp32

torch.nn.functional.leaky_relu_

Y

Only support fp16,fp32

torch.nn.functional.prelu

     

torch.nn.functional.rrelu

     

torch.nn.functional.rrelu_

     

torch.nn.functional.glu

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.gelu

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.logsigmoid

     

torch.nn.functional.hardshrink

     

torch.nn.functional.tanhshrink

     

torch.nn.functional.softsign

     

torch.nn.functional.softplus

Y

Only support fp16,fp32

torch.nn.functional.softmin

     

torch.nn.functional.softmax

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.softshrink

     

torch.nn.functional.gumbel_softmax

     

torch.nn.functional.log_softmax

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.tanh

Y

Only support fp16,fp32,int64

torch.nn.functional.sigmoid

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.hardsigmoid

Y

Only support fp16,fp32

torch.nn.functional.silu

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.mish

     

torch.nn.functional.batch_norm

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.group_norm

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.instance_norm

     

torch.nn.functional.layer_norm

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.local_response_norm

     

torch.nn.functional.normalize

     

torch.nn.functional.linear

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.bilinear

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.dropout

     

torch.nn.functional.alpha_dropout

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.feature_alpha_dropout

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.dropout1d

     

torch.nn.functional.dropout2d

     

torch.nn.functional.dropout3d

     

torch.nn.functional.embedding

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.embedding_bag

     

torch.nn.functional.one_hot

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.pairwise_distance

     

torch.nn.functional.cosine_similarity

     

torch.nn.functional.pdist

     

torch.nn.functional.binary_cross_entropy

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.binary_cross_entropy_with_logits

     

torch.nn.functional.poisson_nll_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.cosine_embedding_loss

     

torch.nn.functional.cross_entropy

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.ctc_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.gaussian_nll_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.hinge_embedding_loss

     

torch.nn.functional.kl_div

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.l1_loss

     

torch.nn.functional.mse_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.margin_ranking_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.multilabel_margin_loss

     

torch.nn.functional.multilabel_soft_margin_loss

     

torch.nn.functional.multi_margin_loss

     

torch.nn.functional.nll_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.huber_loss

     

torch.nn.functional.smooth_l1_loss

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.soft_margin_loss

     

torch.nn.functional.triplet_margin_loss

     

torch.nn.functional.triplet_margin_with_distance_loss

     

torch.nn.functional.pixel_shuffle

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.pixel_unshuffle

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.pad

     

torch.nn.functional.interpolate

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.upsample

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.upsample_nearest

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.upsample_bilinear

Y

Only support fp16,fp32,int64,bool

torch.nn.functional.grid_sample

     

torch.nn.functional.affine_grid

     

torch.nn.parallel.data_parallel

     

torch.Tensor

PyTorch API

Compatibility

Limitations

torch.Tensor

Y

  

Tensor.T

Y

Only support fp16,fp32,int64,bool

Tensor.H

     

Tensor.mT

     

Tensor.mH

     

Tensor.new_tensor

     

Tensor.new_full

Y

  

Tensor.new_empty

Y

  

Tensor.new_ones

Y

  

Tensor.new_zeros

Y

  

Tensor.is_cuda

     

Tensor.is_quantized

     

Tensor.is_meta

     

Tensor.device

     

Tensor.grad

Y

  

Tensor.ndim

Y

  

Tensor.real

     

Tensor.imag

     

Tensor.nbytes

     

Tensor.itemsize

     

Tensor.abs

Y

Only support fp16,fp32,int64,bool

Tensor.abs_

Y

Only support fp16,fp32,int64,bool

Tensor.absolute

Y

Only support fp16,fp32,int64,bool

Tensor.absolute_

Y

Only support fp16,fp32,int64,bool

Tensor.acos

Y

Only support fp16,fp32,int64,bool

Tensor.acos_

Y

Only support fp16,fp32,int64,bool

Tensor.arccos

Y

Only support fp16,fp32,int64,bool

Tensor.arccos_

Y

Only support fp16,fp32,int64,bool

Tensor.add

Y

Only support fp16,fp32,int64,bool

Tensor.add_

Y

Only support fp16,fp32,int64,bool

Tensor.addbmm

Y

Only support fp16,fp32,int64,bool

Tensor.addbmm_

Y

Only support fp16,fp32,int64,bool

Tensor.addcdiv

Y

Only support fp16,fp32,int64,bool

Tensor.addcdiv_

Y

Only support fp16,fp32,int64,bool

Tensor.addcmul

Y

Only support fp16,fp32,int64,bool

Tensor.addcmul_

Y

Only support fp16,fp32,int64,bool

Tensor.addmm

Y

Only support fp16,fp32,int64,bool

Tensor.addmm_

Y

Only support fp16,fp32,int64,bool

Tensor.sspaddmm

     

Tensor.addmv

Y

Only support fp16,fp32,int64,bool

Tensor.addmv_

Y

Only support fp16,fp32,int64,bool

Tensor.addr

Y

Only support fp16,fp32,int64,bool

Tensor.addr_

Y

Only support fp16,fp32,int64,bool

Tensor.adjoint

     

Tensor.allclose

Y

Only support fp16,fp32,int64,bool

Tensor.amax

Y

Only support fp16,fp32,int64,bool

Tensor.amin

Y

Only support fp16,fp32,int64,bool

Tensor.aminmax

Y

Only support fp16,fp32,int64,bool

Tensor.angle

     

Tensor.apply_

Y

Only support fp16,fp32,int64,bool

Tensor.argmax

Y

Only support fp16,fp32,int64

Tensor.argmin

Y

Only support fp16,fp32,int64,bool

Tensor.argsort

Y

Only support fp16,fp32,int64,bool

Tensor.argwhere

     

Tensor.asin

Y

Only support fp16,fp32,int64,bool

Tensor.asin_

Y

Only support fp16,fp32,int64,bool

Tensor.arcsin

     

Tensor.arcsin_

Y

Only support fp16,fp32,int64,bool

Tensor.as_strided

     

Tensor.atan

Y

Only support fp16,fp32,int64,bool

Tensor.atan_

Y

Only support fp16,fp32,int64,bool

Tensor.arctan

Y

Only support fp16,fp32,int64,bool

Tensor.arctan_

Y

Only support fp16,fp32,int64,bool

Tensor.atan2

     

Tensor.atan2_

     

Tensor.arctan2

     

Tensor.arctan2_

     

Tensor.all

Y

Only support fp16,fp32,int64,bool

Tensor.any

Y

Only support fp16,fp32,int64,bool

Tensor.backward

Y

  

Tensor.baddbmm

Y

Only support fp16,fp32,int64,bool

Tensor.baddbmm_

Y

Only support fp16,fp32,int64,bool

Tensor.bernoulli

Y

Only support fp16,fp32,int64,bool

Tensor.bernoulli_

Y

Only support fp16,fp32,int64,bool

Tensor.bfloat16

Y

Only support fp16,fp32,int64,bool

Tensor.bincount

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_not

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_not_

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_and

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_and_

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_or

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_or_

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_xor

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_xor_

Y

Only support fp16,fp32,int64,bool

Tensor.bitwise_left_shift

     

Tensor.bitwise_left_shift_

     

Tensor.bitwise_right_shift

     

Tensor.bitwise_right_shift_

     

Tensor.bmm

Y

Only support fp16,fp32,int64,bool

Tensor.bool

Y

Only support fp16,fp32,int64,bool

Tensor.byte

Y

Only support fp16,fp32,int64,bool

Tensor.broadcast_to

     

Tensor.cauchy_

     

Tensor.ceil

Y

Only support fp16,fp32,int64,bool

Tensor.ceil_

Y

Only support fp16,fp32,int64,bool

Tensor.char

Y

Only support fp16,fp32,int64,bool

Tensor.cholesky

     

Tensor.cholesky_inverse

     

Tensor.cholesky_solve

     

Tensor.chunk

Y

Only support fp16,fp32,int64,bool

Tensor.clamp

Y

Only support fp16,fp32,int64,bool

Tensor.clamp_

Y

Only support fp16,fp32,int64,bool

Tensor.clip

Y

Only support fp16,fp32,int64,bool

Tensor.clip_

Y

Only support fp16,fp32,int64,bool

Tensor.clone

Y

Only support fp16,fp32,int64,bool

Tensor.contiguous

Y

Only support fp16,fp32,int64,bool

Tensor.copy_

Y

Only support fp16,fp32,int64,bool

Tensor.conj

Y

Only support fp16,fp32,int64,bool

Tensor.conj_physical

     

Tensor.conj_physical_

     

Tensor.resolve_conj

Y

Only support fp16,fp32,int64,bool

Tensor.resolve_neg

Y

Only support fp16,fp32,int64,bool

Tensor.copysign

     

Tensor.copysign_

     

Tensor.cos

Y

Only support fp16,fp32,int64,bool

Tensor.cos_

Y

Only support fp16,fp32,int64,bool

Tensor.cosh

Y

Only support fp16,fp32

Tensor.cosh_

Y

Only support fp16,fp32

Tensor.corrcoef

     

Tensor.count_nonzero

Y

Only support fp16,fp32,int64,bool

Tensor.cov

Y

Only support fp16,fp32,int64,bool

Tensor.acosh

Y

Only support fp16,fp32

Tensor.acosh_

Y

Only support fp16,fp32

Tensor.arccosh

Y

Only support fp16,fp32

Tensor.arccosh_

Y

Only support fp16,fp32

Tensor.cpu

Y

Only support fp16,fp32,int64,bool

Tensor.cross

     

Tensor.cuda

     

Tensor.logcumsumexp

     

Tensor.cummax

Y

Only support fp16,fp32,int64,bool

Tensor.cummin

     

Tensor.cumprod

     

Tensor.cumprod_

     

Tensor.cumsum

Y

Only support fp16,fp32,int64,bool

Tensor.cumsum_

Y

Only support fp16,fp32,int64,bool

Tensor.chalf

     

Tensor.cfloat

     

Tensor.cdouble

     

Tensor.data_ptr

Y

  

Tensor.deg2rad

Y

Only support fp16,fp32,int64,bool

Tensor.dequantize

Y

Only support fp16,fp32,int64,bool

Tensor.det

     

Tensor.dense_dim

     

Tensor.detach

Y

  

Tensor.detach_

Y

  

Tensor.diag

Y

Only support fp16,fp32,int64,bool

Tensor.diag_embed

Y

Only support fp16,fp32,int64,bool

Tensor.diagflat

     

Tensor.diagonal

Y

Only support fp16,fp32,int64,bool

Tensor.diagonal_scatter

Y

Only support fp16,fp32,int64,bool

Tensor.fill_diagonal_

Y

Only support fp16,fp32,int64,bool

Tensor.fmax

     

Tensor.fmin

     

Tensor.diff

Y

Only support fp16,fp32,int64,bool

Tensor.digamma

     

Tensor.digamma_

     

Tensor.dim

Y

  

Tensor.dim_order

     

Tensor.dist

     

Tensor.div

Y

Only support fp16,fp32,int64,bool

Tensor.div_

Y

Only support fp16,fp32,int64,bool

Tensor.divide

Y

Only support fp16,fp32,int64,bool

Tensor.divide_

Y

Only support fp16,fp32,int64,bool

Tensor.dot

Y

Only support fp16,fp32,int64,bool

Tensor.double

Y

Only support fp16,fp32,int64,bool

Tensor.dsplit

Y

  

Tensor.element_size

Y

  

Tensor.eq

Y

Only support fp16,fp32,int64,bool

Tensor.eq_

Y

Only support fp16,fp32,int64,bool

Tensor.equal

Y

Only support fp16,fp32,int64,bool

Tensor.erf

Y

Only support fp16,fp32,int64,bool

Tensor.erf_

Y

Only support fp16,fp32,int64,bool

Tensor.erfc

Y

Only support fp16,fp32,int64,bool

Tensor.erfc_

Y

Only support fp16,fp32,int64,bool

Tensor.erfinv

Y

Only support fp16,fp32,int64,bool

Tensor.erfinv_

Y

Only support fp16,fp32,int64,bool

Tensor.exp

Y

Only support fp16,fp32,int64,bool

Tensor.exp_

Y

Only support fp16,fp32,int64,bool

Tensor.expm1

Y

Only support fp16,fp32,int64,bool

Tensor.expm1_

Y

Only support fp16,fp32,int64,bool

Tensor.expand

Y

Only support fp16,fp32,int64,bool

Tensor.expand_as

Y

Only support fp16,fp32,int64,bool

Tensor.exponential_

     

Tensor.fix

Y

Only support fp16,fp32,int64,bool

Tensor.fix_

Y

Only support fp16,fp32,int64,bool

Tensor.fill_

Y

Only support fp16,fp32,int64,bool

Tensor.flatten

Y

  

Tensor.flip

Y

Only support fp16,fp32,int64,bool

Tensor.fliplr

Y

Only support fp16,fp32,int64,bool

Tensor.flipud

Y

Only support fp16,fp32,int64,bool

Tensor.float

Y

Only support fp16,fp32,int64,bool

Tensor.float_power

Y

Only support fp16,fp32,int64,bool

Tensor.float_power_

Y

Only support fp16,fp32,int64,bool

Tensor.floor

Y

Only support fp16,fp32,int64,bool

Tensor.floor_

Y

Only support fp16,fp32,int64,bool

Tensor.floor_divide

Y

Only support fp16,fp32,int64,bool

Tensor.floor_divide_

Y

Only support fp16,fp32,int64,bool

Tensor.fmod

Y

Only support fp16,fp32,int64,bool

Tensor.fmod_

Y

Only support fp16,fp32,int64,bool

Tensor.frac

Y

Only support fp16,fp32,int64,bool

Tensor.frac_

Y

Only support fp16,fp32,int64,bool

Tensor.frexp

     

Tensor.gather

Y

Only support fp16,fp32,int64,bool

Tensor.gcd

     

Tensor.gcd_

     

Tensor.ge

Y

Only support fp16,fp32,int64,bool

Tensor.ge_

Y

Only support fp16,fp32,int64,bool

Tensor.greater_equal

Y

Only support fp16,fp32,int64,bool

Tensor.greater_equal_

Y

Only support fp16,fp32,int64,bool

Tensor.geometric_

     

Tensor.geqrf

     

Tensor.ger

     

Tensor.get_device

     

Tensor.gt

Y

Only support fp16,fp32,int64,bool

Tensor.gt_

Y

Only support fp16,fp32,int64,bool

Tensor.greater

Y

Only support fp16,fp32,int64,bool

Tensor.greater_

Y

Only support fp16,fp32,int64,bool

Tensor.half

Y

Only support fp16,fp32,int64,bool

Tensor.hardshrink

     

Tensor.heaviside

     

Tensor.histc

Y

Only support fp16,fp32,int64,bool

Tensor.histogram

     

Tensor.hsplit

Y

Only support fp16,fp32,int64,bool

Tensor.hypot

     

Tensor.hypot_

     

Tensor.i0

     

Tensor.i0_

     

Tensor.igamma

     

Tensor.igamma_

     

Tensor.igammac

     

Tensor.igammac_

     

Tensor.index_add_

Y

Only support fp16,fp32,int64,bool

Tensor.index_add

Y

Only support fp16,fp32,int64,bool

Tensor.index_copy_

     

Tensor.index_copy

     

Tensor.index_fill_

Y

Only support fp16,fp32,int64,bool

Tensor.index_fill

Y

Only support fp16,fp32,int64,bool

Tensor.index_put_

Y

Only support fp16,fp32,int64,bool

Tensor.index_put

Y

Only support fp16,fp32,int64,bool

Tensor.index_reduce_

     

Tensor.index_reduce

     

Tensor.index_select

     

Tensor.indices

     

Tensor.inner

Y

Only support fp16,fp32,int64,bool

Tensor.int

Y

Only support fp16,fp32,int64,bool

Tensor.int_repr

     

Tensor.inverse

     

Tensor.isclose

Y

Only support fp16,fp32,int64,bool

Tensor.isfinite

Y

Only support fp16,fp32,int64,bool

Tensor.isinf

Y

Only support fp16,fp32,int64,bool

Tensor.isposinf

Y

Only support fp16,fp32,int64,bool

Tensor.isneginf

Y

Only support fp16,fp32,int64,bool

Tensor.isnan

Y

  

Tensor.is_contiguous

Y

  

Tensor.is_complex

Y

Only support fp16,fp32,int64,bool

Tensor.is_conj

Y

Only support fp16,fp32,int64,bool

Tensor.is_floating_point

Y

  

Tensor.is_inference

Y

  

Tensor.is_leaf

     

Tensor.is_pinned

Y

  

Tensor.is_set_to

Y

  

Tensor.is_shared

     

Tensor.is_signed

Y

  

Tensor.is_sparse

Y

  

Tensor.istft

     

Tensor.isreal

Y

Only support fp16,fp32,int64,bool

Tensor.item

Y

  

Tensor.kthvalue

Y

Only support fp16,fp32

Tensor.lcm

     

Tensor.lcm_

     

Tensor.ldexp

Y

Only support fp16,fp32,int64,bool

Tensor.ldexp_

Y

Only support fp16,fp32,int64,bool

Tensor.le

Y

Only support fp16,fp32,int64,bool

Tensor.le_

Y

Only support fp16,fp32,int64,bool

Tensor.less_equal

Y

Only support fp16,fp32,int64,bool

Tensor.less_equal_

Y

Only support fp16,fp32,int64,bool

Tensor.lerp

Y

Only support fp16,fp32,int64,bool

Tensor.lerp_

Y

Only support fp16,fp32,int64,bool

Tensor.lgamma

     

Tensor.lgamma_

     

Tensor.log

Y

Only support fp16,fp32,int64,bool

Tensor.log_

Y

Only support fp16,fp32,int64,bool

Tensor.logdet

     

Tensor.log10

     

Tensor.log10_

     

Tensor.log1p

Y

Only support fp16,fp32,int64,bool

Tensor.log1p_

Y

Only support fp16,fp32,int64,bool

Tensor.log2

Y

Only support fp16,fp32,int64,bool

Tensor.log2_

Y

Only support fp16,fp32,int64,bool

Tensor.log_normal_

     

Tensor.logaddexp

Y

Only support fp16,fp32,int64,bool

Tensor.logaddexp2

Y

Only support fp16,fp32,int64,bool

Tensor.logsumexp

Y

Only support fp16,fp32,int64,bool

Tensor.logical_and

Y

Only support fp16,fp32,int64,bool

Tensor.logical_and_

Y

Only support fp16,fp32,int64,bool

Tensor.logical_not

Y

Only support fp16,fp32,int64,bool

Tensor.logical_not_

Y

Only support fp16,fp32,int64,bool

Tensor.logical_or

Y

Only support fp16,fp32,int64,bool

Tensor.logical_or_

Y

Only support fp16,fp32,int64,bool

Tensor.logical_xor

Y

Only support fp16,fp32,int64,bool

Tensor.logical_xor_

Y

Only support fp16,fp32,int64,bool

Tensor.logit

     

Tensor.logit_

     

Tensor.long

Y

Only support fp16,fp32,int64,bool

Tensor.lt

Y

Only support fp16,fp32,int64,bool

Tensor.lt_

Y

Only support fp16,fp32,int64,bool

Tensor.less

Y

Only support fp16,fp32,int64,bool

Tensor.less_

Y

Only support fp16,fp32,int64,bool

Tensor.lu

     

Tensor.lu_solve

     

Tensor.as_subclass

     

Tensor.map_

Y

  

Tensor.masked_scatter_

Y

Only support fp16,fp32,int64,bool

Tensor.masked_scatter

Y

Only support fp16,fp32,int64,bool

Tensor.masked_fill_

Y

Only support fp16,fp32,int64,bool

Tensor.masked_fill

Y

Only support fp16,fp32,int64,bool

Tensor.masked_select

Y

Only support fp16,fp32,int64,bool

Tensor.matmul

Y

Only support fp16,fp32,int64,bool

Tensor.matrix_power

Y

Only support fp16,fp32,int64,bool

Tensor.matrix_exp

     

Tensor.max

Y

Only support fp16,fp32,int64,bool

Tensor.maximum

Y

Only support fp16,fp32,int64,bool

Tensor.mean

Y

Only support fp16,fp32,int64,bool

Tensor.nanmean

Y

Only support fp16,fp32,int64,bool

Tensor.median

Y

Only support fp16,fp32,int64

Tensor.nanmedian

     

Tensor.min

Y

Only support fp16,fp32,int64,bool

Tensor.minimum

Y

Only support fp16,fp32,int64,bool

Tensor.mm

Y

Only support fp16,fp32,int64,bool

Tensor.smm

     

Tensor.mode

     

Tensor.movedim

Y

Only support fp16,fp32,int64,bool

Tensor.moveaxis

Y

Only support fp16,fp32,int64,bool

Tensor.msort

Y

Only support fp16,fp32,int64,bool

Tensor.mul

Y

Only support fp16,fp32,int64,bool

Tensor.mul_

Y

Only support fp16,fp32,int64,bool

Tensor.multiply

Y

Only support fp16,fp32,int64,bool

Tensor.multiply_

Y

Only support fp16,fp32,int64,bool

Tensor.multinomial

     

Tensor.mv

     

Tensor.mvlgamma

     

Tensor.mvlgamma_

     

Tensor.nansum

     

Tensor.narrow

Y

Only support fp16,fp32,int64,bool

Tensor.narrow_copy

Y

Only support fp16,fp32,int64,bool

Tensor.ndimension

Y

  

Tensor.nan_to_num

     

Tensor.nan_to_num_

     

Tensor.ne

Y

Only support fp16,fp32,int64,bool

Tensor.ne_

Y

Only support fp16,fp32,int64,bool

Tensor.not_equal

Y

Only support fp16,fp32,int64,bool

Tensor.not_equal_

Y

Only support fp16,fp32,int64,bool

Tensor.neg

Y

Only support fp16,fp32,int64,bool

Tensor.neg_

Y

Only support fp16,fp32,int64,bool

Tensor.negative

Y

Only support fp16,fp32,int64,bool

Tensor.negative_

Y

Only support fp16,fp32,int64,bool

Tensor.nelement

Y

  

Tensor.nextafter

     

Tensor.nextafter_

     

Tensor.nonzero

Y

Only support fp16,fp32,int64,bool

Tensor.norm

     

Tensor.normal_

Y

Only support fp16,fp32,int64,bool

Tensor.numel

Y

  

Tensor.numpy

Y

  

Tensor.orgqr

     

Tensor.ormqr

     

Tensor.outer

Y

Only support fp16,fp32,int64,bool

Tensor.permute

Y

Only support fp16,fp32,int64,bool

Tensor.pin_memory

     

Tensor.pinverse

     

Tensor.polygamma

     

Tensor.polygamma_

     

Tensor.positive

Y

Only support fp16,fp32,int64,bool

Tensor.pow

     

Tensor.pow_

Y

Only support fp16,fp32,int64,bool

Tensor.prod

Y

Only support fp16,fp32,int64,bool

Tensor.put_

     

Tensor.qr

Y

Only support fp16,fp32,int64,bool

Tensor.qscheme

     

Tensor.quantile

     

Tensor.nanquantile

     

Tensor.q_scale

     

Tensor.q_zero_point

     

Tensor.q_per_channel_scales

     

Tensor.q_per_channel_zero_points

     

Tensor.q_per_channel_axis

     

Tensor.rad2deg

Y

Only support fp16,fp32,int64,bool

Tensor.random_

Y

Only support fp16,fp32,int64,bool

Tensor.ravel

Y

Only support fp16,fp32,int64,bool

Tensor.reciprocal

Y

Only support fp16,fp32,int64,bool

Tensor.reciprocal_

Y

Only support fp16,fp32,int64,bool

Tensor.record_stream

     

Tensor.register_hook

Y

  

Tensor.register_post_accumulate_grad_hook

     

Tensor.remainder

Y

Only support fp16,fp32,int64,bool

Tensor.remainder_

Y

Only support fp16,fp32,int64,bool

Tensor.renorm

     

Tensor.renorm_

     

Tensor.repeat

Y

Only support fp16,fp32,int64,bool

Tensor.repeat_interleave

Y

Only support fp16,fp32,int64,bool

Tensor.requires_grad

     

Tensor.requires_grad_

Y

  

Tensor.reshape

Y

Only support fp16,fp32,int64,bool

Tensor.reshape_as

Y

Only support fp16,fp32,int64,bool

Tensor.resize_

     

Tensor.resize_as_

     

Tensor.retain_grad

Y

  

Tensor.retains_grad

Y

  

Tensor.roll

Y

Only support fp16,fp32,int64,bool

Tensor.rot90

Y

Only support fp16,fp32,int64,bool

Tensor.round

Y

Only support fp16,fp32,int64,bool

Tensor.round_

Y

Only support fp16,fp32,int64,bool

Tensor.rsqrt

Y

Only support fp16,fp32,int64,bool

Tensor.rsqrt_

Y

Only support fp16,fp32,int64,bool

Tensor.scatter

Y

Only support fp16,fp32,int64,bool

Tensor.scatter_

Y

Only support fp16,fp32,int64,bool

Tensor.scatter_add_

Y

Only support fp16,fp32,int64,bool

Tensor.scatter_add

Y

Only support fp16,fp32,int64,bool

Tensor.scatter_reduce_

     

Tensor.scatter_reduce

Y

Only support fp16,fp32,int64,bool

Tensor.select

Y

Only support fp16,fp32,int64,bool

Tensor.select_scatter

Y

Only support fp16,fp32,int64,bool

Tensor.set_

Y

Only support fp16,fp32,int64,bool

Tensor.share_memory_

     

Tensor.short

Y

Only support fp16,fp32,int64,bool

Tensor.sigmoid

     

Tensor.sigmoid_

Y

Only support fp16,fp32,int64,bool

Tensor.sign

Y

Only support fp16,fp32,int64,bool

Tensor.sign_

Y

Only support fp16,fp32,int64,bool

Tensor.signbit

     

Tensor.sgn

     

Tensor.sgn_

Y

Only support fp16,fp32,int64,bool

Tensor.sin

Y

Only support fp16,fp32,int64,bool

Tensor.sin_

Y

Only support fp16,fp32,int64,bool

Tensor.sinc

     

Tensor.sinc_

     

Tensor.sinh

Y

Only support fp16,fp32

Tensor.sinh_

Y

Only support fp16,fp32

Tensor.asinh

Y

Only support fp16,fp32,int64,bool

Tensor.asinh_

Y

Only support fp16,fp32,int64,bool

Tensor.arcsinh

Y

Only support fp16,fp32,int64,bool

Tensor.arcsinh_

Y

Only support fp16,fp32,int64,bool

Tensor.shape

     

Tensor.size

Y

  

Tensor.slogdet

Y

Only support fp16,fp32,int64,bool

Tensor.slice_scatter

Y

Only support fp16,fp32,int64,bool

Tensor.softmax

     

Tensor.sort

Y

Only support fp16,fp32,int64,bool

Tensor.split

Y

Only support fp16,fp32,int64,bool

Tensor.sparse_mask

     

Tensor.sparse_dim

     

Tensor.sqrt

Y

Only support fp16,fp32

Tensor.sqrt_

Y

Only support fp16,fp32

Tensor.square

Y

Only support fp16,fp32,int64

Tensor.square_

Y

Only support fp16,fp32,int64,bool

Tensor.squeeze

Y

Only support fp16,fp32,int64,bool

Tensor.squeeze_

Y

  

Tensor.std

Y

Only support fp16,fp32

Tensor.stft

     

Tensor.storage

Y

  

Tensor.untyped_storage

Y

  

Tensor.storage_offset

Y

  

Tensor.storage_type

Y

  

Tensor.stride

Y

  

Tensor.sub

Y

Only support fp16,fp32,int64

Tensor.sub_

Y

Only support fp16,fp32,int64

Tensor.subtract

     

Tensor.subtract_

Y

Only support fp16,fp32,int64,bool

Tensor.sum

Y

Only support fp16,fp32,int64,bool

Tensor.sum_to_size

Y

Only support fp16,fp32,int64,bool

Tensor.svd

     

Tensor.swapaxes

Y

Only support fp16,fp32,int64,bool

Tensor.swapdims

Y

Only support fp16,fp32,int64,bool

Tensor.t

Y

Only support fp16,fp32,int64,bool

Tensor.t_

Y

Only support fp16,fp32,int64,bool

Tensor.tensor_split

Y

Only support fp16,fp32,int64,bool

Tensor.tile

Y

Only support fp16,fp32,int64,bool

Tensor.to

Y

Only support fp16,fp32,int64,bool

to

     

torch.to

     

Tensor.to_mkldnn

     

Tensor.take

Y

Only support fp16,fp32,int64,bool

Tensor.take_along_dim

Y

Only support fp16,fp32,int64,bool

Tensor.tan

Y

Only support fp16,fp32,int64

Tensor.tan_

Y

Only support fp16,fp32,int64,bool

Tensor.tanh

Y

Only support fp16,fp32,int64

Tensor.tanh_

Y

Only support fp16,fp32,int64,bool

Tensor.atanh

Y

Only support fp16,fp32,int64,bool

Tensor.atanh_

Y

Only support fp16,fp32,int64,bool

Tensor.arctanh

     

Tensor.arctanh_

     

Tensor.tolist

Y

  

Tensor.topk

Y

Only support fp16,fp32,int64,bool;Don`t support sorted=False

Tensor.to_dense

     

Tensor.to_sparse

     

to_sparse

     

Tensor.to_sparse_csr

     

Tensor.to_sparse_csc

     

Tensor.to_sparse_bsr

     

Tensor.to_sparse_bsc

     

Tensor.trace

     

Tensor.transpose

Y

Only support fp16,fp32,int64,bool

Tensor.transpose_

     

Tensor.triangular_solve

     

Tensor.tril

Y

Only support fp16,fp32,int64,bool

Tensor.tril_

Y

Only support fp16,fp32,int64,bool

Tensor.triu

Y

Only support fp16,fp32,int64,bool

Tensor.triu_

Y

Only support fp16,fp32,int64,bool

Tensor.true_divide

Y

Only support fp16,fp32,int64,bool

Tensor.true_divide_

Y

Only support fp16,fp32,int64,bool

Tensor.trunc

Y

Only support fp16,fp32,int64,bool

Tensor.trunc_

Y

Only support fp16,fp32,int64,bool

Tensor.type

Y

Only support fp16,fp32,int64,bool

Tensor.type_as

Y

Only support fp16,fp32,int64,bool

Tensor.unbind

Y

Only support fp16,fp32,int64,bool

Tensor.unflatten

Y

Only support fp16,fp32,int64,bool

Tensor.unfold

Y

Only support fp16,fp32,int64,bool

Tensor.uniform_

Y

Only support fp16,fp32,int64,bool

Tensor.unique

Y

Only support fp16,fp32,int64,bool

Tensor.unique_consecutive

     

Tensor.unsqueeze

Y

Only support fp16,fp32,int64,bool

Tensor.unsqueeze_

Y

  

Tensor.values

     

Tensor.var

Y

Only support fp16,fp32

Tensor.vdot

     

Tensor.view

Y

Only support fp16,fp32,int64,bool

view

     

Tensor.view_as

Y

Only support fp16,fp32,int64,bool

Tensor.vsplit

Y

Only support fp16,fp32,int64,bool

Tensor.where

Y

Only support fp16,fp32,int64,bool;Don`t support 8-dimensional shape

Tensor.xlogy

Y

Only support fp16,fp32

Tensor.xlogy_

Y

Only support fp16,fp32

Tensor.zero_

Y

Only support fp16,fp32,int64

Tensor Attributes

PyTorch API

Compatibility

Limitations

torch.dtype

Y

  

torch.device

Y

  

torch.layout

Y

  

torch.memory_format

     

torch.autograd

PyTorch API

Compatibility

Limitations

torch.autograd.Function

Y

  

torch.autograd.profiler.profile

Y

  

torch.autograd.profiler.emit_nvtx

     

torch.autograd.profiler.emit_itt

     

torch.autograd.detect_anomaly

Y

  

torch.autograd.set_detect_anomaly

Y

  

torch.autograd.graph.saved_tensors_hooks

Y

  

torch.autograd.graph.save_on_cpu

Y

  

torch.autograd.graph.disable_saved_tensors_hooks

Y

  

torch.autograd.graph.register_multi_grad_hook

Y

  

torch.autograd.graph.allow_mutation_on_saved_tensors

Y

  

torch.autograd.backward

     

torch.autograd.grad

Y

  

torch.autograd.forward_ad.dual_level

Y

  

torch.autograd.forward_ad.make_dual

Y

  

torch.autograd.forward_ad.unpack_dual

Y

  

torch.autograd.functional.jacobian

Y

  

torch.autograd.functional.hessian

Y

  

torch.autograd.functional.vjp

Y

  

torch.autograd.functional.jvp

Y

  

torch.autograd.functional.vhp

Y

  

torch.autograd.functional.hvp

Y

  

Function.forward

     

Function.backward

     

Function.jvp

     

Function.vmap

     

FunctionCtx.mark_dirty

Y

  

FunctionCtx.mark_non_differentiable

Y

  

FunctionCtx.save_for_backward

Y

  

FunctionCtx.set_materialize_grads

Y

  

torch.autograd.gradcheck

Y

  

torch.autograd.gradgradcheck

Y

  

profile.export_chrome_trace

Y

  

profile.key_averages

Y

  

profile.self_cpu_time_total

     

profile.total_average

Y

  

torch.autograd.profiler.load_nvprof

     

Node.name

     

Node.metadata

     

Node.next_functions

     

Node.register_hook

     

Node.register_prehook

     

torch.library

PyTorch API

Compatibility

Limitations

torch.library.Library

Y

  

torch.library.Library.define

Y

  

torch.library.Library.impl

Y

  

torch.library.fallthrough_kernel

     

torch.cuda

The cuda-related interfaces prefixed with torch.cuda correspond to the npu-related interfaces prefixed with torch.npu. When using them, you only need to replace torch.cuda.xxx with torch.npu.xxx.

Take torch.cuda.set_device as an example.

...

torch.cuda.set_device("cuda:0")
# torch.npu.set_device("cuda:0")

...

PyTorch API

Ascend NPU API

Compatibility

Limitations

torch.cuda.can_device_access_peer

torch.npu.can_device_access_peer

Y

  

torch.cuda.current_blas_handle

torch.npu.current_blas_handle

Y

  

torch.cuda.current_device

torch.npu.current_device

Y

  

torch.cuda.device_count

torch.npu.device_count

Y

  

torch.cuda.device_of

torch.npu.device_of

Y

  

torch.cuda.get_device_name

torch.npu.get_device_name

Y

  

torch.cuda.get_device_properties

torch.npu.get_device_properties

Y

  

torch.cuda.get_sync_debug_mode

torch.npu.get_sync_debug_mode

Y

  

torch.cuda.init

torch.npu.init

Y

  

torch.cuda.is_available

torch.npu.is_available

Y

  

torch.cuda.is_initialized

torch.npu.is_initialized

Y

  

torch.cuda.set_device

torch.npu.set_device

Y

  

torch.cuda.set_stream

torch.npu.set_stream

Y

  

torch.cuda.set_sync_debug_mode

torch.npu.set_sync_debug_mode

Y

  

torch.cuda.stream

torch.npu.stream

Y

  

torch.cuda.utilization

torch.npu.utilization

Y

  

torch.cuda.get_rng_state

torch.npu.get_rng_state

Y

  

torch.cuda.set_rng_state

torch.npu.set_rng_state

Y

  

torch.cuda.set_rng_state_all

torch.npu.set_rng_state_all

Y

  

torch.cuda.manual_seed

torch.npu.manual_seed

Y

  

torch.cuda.manual_seed_all

torch.npu.manual_seed_all

Y

  

torch.cuda.seed

torch.npu.seed

Y

  

torch.cuda.seed_all

torch.npu.seed_all

Y

  

torch.cuda.initial_seed

torch.npu.initial_seed

Y

  

torch.cuda.Stream

torch.npu.Stream

Y

  

torch.cuda.Stream.wait_stream

torch.npu.Stream.wait_stream

Y

  

torch.cuda.Event

torch.npu.Event

Y

  

torch.cuda.Event.elapsed_time

torch.npu.Event.elapsed_time

Y

  

torch.cuda.Event.query

torch.npu.Event.query

Y

  

torch.cuda.Event.wait

torch.npu.Event.wait

Y

  

torch.cuda.empty_cache

torch.npu.empty_cache

Y

  

torch.cuda.memory_stats

torch.npu.memory_stats

Y

  

torch.cuda.memory_summary

torch.npu.memory_summary

Y

  

torch.cuda.memory_allocated

torch.npu.memory_allocated

Y

  

torch.cuda.max_memory_allocated

torch.npu.max_memory_allocated

Y

  

torch.cuda.reset_max_memory_allocated

torch.npu.reset_max_memory_allocated

Y

  

torch.cuda.memory_reserved

torch.npu.memory_reserved

Y

  

torch.cuda.max_memory_reserved

torch.npu.max_memory_reserved

Y

  

torch.cuda.set_per_process_memory_fraction

torch.npu.set_per_process_memory_fraction

Y

  

torch.cuda.memory_cached

torch.npu.memory_cached

Y

  

torch.cuda.max_memory_cached

torch.npu.max_memory_cached

Y

  

torch.cuda.reset_max_memory_cached

torch.npu.reset_max_memory_cached

Y

  

torch.cuda.reset_peak_memory_stats

torch.npu.reset_peak_memory_stats

Y

  

torch.cuda.caching_allocator_alloc

torch.npu.caching_allocator_alloc

Y

  

torch.cuda.caching_allocator_delete

torch.npu.caching_allocator_delete

Y

  

torch.cuda.StreamContext

        

torch.cuda.current_stream

        

torch.cuda.default_stream

        

torch.cuda.device

        

torch.cuda.get_arch_list

        

torch.cuda.get_device_capability

        

torch.cuda.get_gencode_flags

        

torch.cuda.ipc_collect

        

torch.cuda.memory_usage

        

torch.cuda.synchronize

        

torch.cuda.temperature

        

torch.cuda.power_draw

        

torch.cuda.clock_rate

        

torch.cuda.get_rng_state_all

        

torch.cuda.comm.broadcast

        

torch.cuda.comm.broadcast_coalesced

        

torch.cuda.comm.reduce_add

        

torch.cuda.comm.scatter

        

torch.cuda.comm.gather

        

torch.cuda.Stream.query

        

torch.cuda.Stream.record_event

        

torch.cuda.Stream.synchronize

        

torch.cuda.Stream.wait_event

        

torch.cuda.ExternalStream

        

torch.cuda.ExternalStream.query

        

torch.cuda.ExternalStream.record_event

        

torch.cuda.ExternalStream.synchronize

        

torch.cuda.ExternalStream.wait_event

        

torch.cuda.ExternalStream.wait_stream

        

torch.cuda.Event.from_ipc_handle

        

torch.cuda.Event.ipc_handle

        

torch.cuda.Event.record

        

torch.cuda.Event.synchronize

        

torch.cuda.is_current_stream_capturing

        

torch.cuda.graph_pool_handle

        

torch.cuda.CUDAGraph

        

torch.cuda.CUDAGraph.capture_begin

        

torch.cuda.CUDAGraph.capture_end

        

torch.cuda.CUDAGraph.debug_dump

        

torch.cuda.CUDAGraph.enable_debug_mode

        

torch.cuda.CUDAGraph.pool

        

torch.cuda.CUDAGraph.replay

        

torch.cuda.CUDAGraph.reset

        

torch.cuda.graph

        

torch.cuda.make_graphed_callables

        

torch.cuda.list_gpu_processes

        

torch.cuda.mem_get_info

        

torch.cuda.memory_snapshot

        

torch.cuda.get_allocator_backend

        

torch.cuda.CUDAPluggableAllocator

        

torch.cuda.change_current_allocator

        

torch.cuda.nvtx.mark

        

torch.cuda.nvtx.range_push

        

torch.cuda.nvtx.range_pop

        

torch.cuda.jiterator._create_jit_fn

        

torch.cuda.jiterator._create_multi_output_jit_fn

        

torch.cuda._sanitizer.enable_cuda_sanitizer

        

torch.cuda.amp

Just like torch.cuda above, you only need to replace torch.cuda.amp.xxx with torch.npu.amp.xxx.

PyTorch API

Ascend NPU API

Compatibility

Limitations

torch.autocast

-

Y

  

torch.cuda.amp.autocast

torch.npu.amp.autocast

Y

  

torch.cuda.amp.custom_fwd

        

torch.cuda.amp.custom_bwd

torch.npu.amp.custom_bwd

Y

  

torch.cpu.amp.autocast

torch.cpu.amp.autocast

Y

  

torch.cuda.amp.GradScaler

torch.npu.amp.GradScaler

Y

  

torch.cuda.amp.GradScaler.get_backoff_factor

torch.npu.amp.GradScaler.get_backoff_factor

Y

  

torch.cuda.amp.GradScaler.get_growth_factor

torch.npu.amp.GradScaler.get_growth_factor

Y

  

torch.cuda.amp.GradScaler.get_growth_interval

torch.npu.amp.GradScaler.get_growth_interval

Y

  

torch.cuda.amp.GradScaler.get_scale

torch.npu.amp.GradScaler.get_scale

Y

  

torch.cuda.amp.GradScaler.is_enabled

torch.npu.amp.GradScaler.is_enabled

Y

  

torch.cuda.amp.GradScaler.load_state_dict

torch.npu.amp.GradScaler.load_state_dict

Y

  

torch.cuda.amp.GradScaler.scale

torch.npu.amp.GradScaler.scale

Y

  

torch.cuda.amp.GradScaler.set_backoff_factor

torch.npu.amp.GradScaler.set_backoff_factor

Y

  

torch.cuda.amp.GradScaler.set_growth_factor

torch.npu.amp.GradScaler.set_growth_factor

Y

  

torch.cuda.amp.GradScaler.set_growth_interval

torch.npu.amp.GradScaler.set_growth_interval

Y

  

torch.cuda.amp.GradScaler.state_dict

torch.npu.amp.GradScaler.state_dict

Y

  

torch.cuda.amp.GradScaler.step

torch.npu.amp.GradScaler.step

Y

  

torch.cuda.amp.GradScaler.unscale_

torch.npu.amp.GradScaler.unscale_

Y

  

torch.cuda.amp.GradScaler.update

torch.npu.amp.GradScaler.update

Y

  

torch.backends

PyTorch API

Compatibility

Limitations

torch.backends.cpu.get_cpu_capability

     

torch.backends.cuda.is_built

Y

  

torch.backends.cuda.matmul.allow_tf32

Y

  

torch.backends.cuda.matmul.allow_fp16_reduced_precision_reduction

     

torch.backends.cuda.matmul.allow_bf16_reduced_precision_reduction

     

torch.backends.cuda.cufft_plan_cache

     

torch.backends.cuda.cufft_plan_cache.size

     

torch.backends.cuda.cufft_plan_cache.max_size

     

torch.backends.cuda.cufft_plan_cache.clear

     

torch.backends.cuda.preferred_linalg_library

Y

  

torch.backends.cuda.SDPBackend

Y

  

torch.backends.cuda.flash_sdp_enabled

Y

  

torch.backends.cuda.enable_mem_efficient_sdp

Y

  

torch.backends.cuda.mem_efficient_sdp_enabled

Y

  

torch.backends.cuda.enable_flash_sdp

Y

  

torch.backends.cuda.math_sdp_enabled

Y

  

torch.backends.cuda.enable_math_sdp

Y

  

torch.backends.cuda.sdp_kernel

Y

  

torch.backends.cudnn.version

     

torch.backends.cudnn.is_available

Y

  

torch.backends.cudnn.enabled

Y

  

torch.backends.cudnn.allow_tf32

Y

  

torch.backends.cudnn.deterministic

Y

  

torch.backends.cudnn.benchmark

Y

  

torch.backends.cudnn.benchmark_limit

     

torch.backends.mps.is_available

Y

  

torch.backends.mps.is_built

Y

  

torch.backends.mkl.is_available

Y

  

torch.backends.mkl.verbose

Y

  

torch.backends.mkldnn.is_available

Y

  

torch.backends.mkldnn.verbose

Y

  

torch.backends.openmp.is_available

Y

  

torch.backends.opt_einsum.is_available

Y

  

torch.backends.opt_einsum.get_opt_einsum

Y

  

torch.backends.opt_einsum.enabled

     

torch.backends.opt_einsum.strategy

     

torch.distributed

PyTorch API

Compatibility

Limitations

torch.distributed.is_available

Y

  

torch.distributed.init_process_group

Y

  

torch.distributed.is_initialized

Y

  

torch.distributed.is_mpi_available

Y

  

torch.distributed.is_nccl_available

Y

  

torch.distributed.is_gloo_available

Y

  

torch.distributed.is_torchelastic_launched

Y

  

torch.distributed.Backend

Y

  

torch.distributed.Backend.register_backend

     

torch.distributed.get_backend

Y

  

torch.distributed.get_rank

Y

  

torch.distributed.get_world_size

Y

  

torch.distributed.Store

Y

  

torch.distributed.TCPStore

Y

  

torch.distributed.HashStore

Y

  

torch.distributed.FileStore

Y

  

torch.distributed.PrefixStore

Y

  

torch.distributed.Store.set

Y

  

torch.distributed.Store.get

Y

  

torch.distributed.Store.add

Y

  

torch.distributed.Store.compare_set

Y

  

torch.distributed.Store.wait

Y

  

torch.distributed.Store.num_keys

Y

  

torch.distributed.Store.delete_key

Y

  

torch.distributed.Store.set_timeout

Y

  

torch.distributed.new_group

Y

  

torch.distributed.get_group_rank

Y

  

torch.distributed.get_global_rank

Y

  

torch.distributed.get_process_group_ranks

Y

  

torch.distributed.send

Y

  

torch.distributed.recv

Y

  

torch.distributed.isend

     

torch.distributed.irecv

     

torch.distributed.batch_isend_irecv

     

torch.distributed.P2POp

Y

  

torch.distributed.broadcast

Y

  

torch.distributed.broadcast_object_list

Y

  

torch.distributed.all_reduce

Y

  

torch.distributed.reduce

     

torch.distributed.all_gather

Y

  

torch.distributed.all_gather_into_tensor

Y

  

torch.distributed.all_gather_object

Y

  

torch.distributed.gather

     

torch.distributed.gather_object

     

torch.distributed.scatter

     

torch.distributed.scatter_object_list

     

torch.distributed.reduce_scatter

Y

  

torch.distributed.reduce_scatter_tensor

Y

  

torch.distributed.all_to_all_single

Y

  

torch.distributed.all_to_all

Y

  

torch.distributed.barrier

Y

  

torch.distributed.monitored_barrier

     

torch.distributed.ReduceOp

Y

  

torch.distributed.reduce_op

Y

  

torch.distributed.broadcast_multigpu

     

torch.distributed.all_reduce_multigpu

     

torch.distributed.reduce_multigpu

     

torch.distributed.all_gather_multigpu

     

torch.distributed.reduce_scatter_multigpu

     

torch.distributed.DistBackendError

Y

  

torch.distributed.algorithms.join

PyTorch API

Compatibility

Limitations

torch.distributed.algorithms.Join

Y

  

torch.distributed.algorithms.Join.notify_join_context

     

torch.distributed.algorithms.Joinable

Y

  

torch.distributed.algorithms.Joinable.join_device

     

torch.distributed.algorithms.Joinable.join_hook

     

torch.distributed.algorithms.Joinable.join_process_group

     

torch.distributed.algorithms.JoinHook

Y

  

torch.distributed.algorithms.JoinHook.main_hook

Y

  

torch.distributed.algorithms.JoinHook.post_hook

Y

  

torch.distributed.elastic

PyTorch API

Compatibility

Limitations

torch.distributed.elastic.agent.server.ElasticAgent

     

torch.distributed.elastic.agent.server.ElasticAgent.get_worker_group

     

torch.distributed.elastic.agent.server.ElasticAgent.run

     

torch.distributed.elastic.agent.server.WorkerSpec

     

torch.distributed.elastic.agent.server.WorkerSpec.get_entrypoint_name

     

torch.distributed.elastic.agent.server.WorkerState

     

torch.distributed.elastic.agent.server.WorkerState.is_running

     

torch.distributed.elastic.agent.server.Worker

     

torch.distributed.elastic.agent.server.WorkerGroup

     

torch.distributed.elastic.agent.server.local_elastic_agent.LocalElasticAgent

     

torch.distributed.elastic.agent.server.SimpleElasticAgent

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._assign_worker_ranks

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._exit_barrier

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._initialize_workers

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._monitor_workers

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._rendezvous

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._restart_workers

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._shutdown

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._start_workers

     

torch.distributed.elastic.agent.server.SimpleElasticAgent._stop_workers

     

torch.distributed.elastic.agent.server.api.RunResult

     

torch.distributed.elastic.multiprocessing.start_processes

     

torch.distributed.elastic.multiprocessing.api.PContext

     

torch.distributed.elastic.multiprocessing.api.MultiprocessContext

     

torch.distributed.elastic.multiprocessing.api.SubprocessContext

     

torch.distributed.elastic.multiprocessing.api.RunProcsResult

     

torch.distributed.elastic.multiprocessing.errors.record

     

torch.distributed.elastic.multiprocessing.errors.ChildFailedError

     

torch.distributed.elastic.multiprocessing.errors.ErrorHandler

     

torch.distributed.elastic.multiprocessing.errors.ProcessFailure

     

torch.distributed.elastic.rendezvous.RendezvousParameters

     

torch.distributed.elastic.rendezvous.RendezvousParameters.get

     

torch.distributed.elastic.rendezvous.RendezvousParameters.get_as_bool

     

torch.distributed.elastic.rendezvous.RendezvousParameters.get_as_int

     

torch.distributed.elastic.rendezvous.RendezvousHandlerRegistry

     

torch.distributed.elastic.rendezvous.RendezvousHandlerRegistry.create_handler

     

torch.distributed.elastic.rendezvous.RendezvousHandlerRegistry.register

     

torch.distributed.elastic.rendezvous.RendezvousHandler

     

torch.distributed.elastic.rendezvous.RendezvousHandler.get_backend

     

torch.distributed.elastic.rendezvous.RendezvousHandler.get_run_id

     

torch.distributed.elastic.rendezvous.RendezvousHandler.is_closed

     

torch.distributed.elastic.rendezvous.RendezvousHandler.next_rendezvous

     

torch.distributed.elastic.rendezvous.RendezvousHandler.num_nodes_waiting

     

torch.distributed.elastic.rendezvous.RendezvousHandler.set_closed

     

torch.distributed.elastic.rendezvous.RendezvousHandler.shutdown

     

torch.distributed.elastic.rendezvous.RendezvousError

     

torch.distributed.elastic.rendezvous.RendezvousClosedError

     

torch.distributed.elastic.rendezvous.RendezvousTimeoutError

     

torch.distributed.elastic.rendezvous.RendezvousConnectionError

     

torch.distributed.elastic.rendezvous.RendezvousStateError

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.create_handler

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.DynamicRendezvousHandler

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.DynamicRendezvousHandler.from_backend

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousBackend

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousBackend.get_state

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousBackend.name

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousBackend.set_state

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousTimeout

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousTimeout.close

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousTimeout.heartbeat

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousTimeout.join

     

torch.distributed.elastic.rendezvous.dynamic_rendezvous.RendezvousTimeout.last_call

     

torch.distributed.elastic.rendezvous.c10d_rendezvous_backend.create_backend

     

torch.distributed.elastic.rendezvous.c10d_rendezvous_backend.C10dRendezvousBackend

     

torch.distributed.elastic.rendezvous.c10d_rendezvous_backend.C10dRendezvousBackend.get_state

     

torch.distributed.elastic.rendezvous.c10d_rendezvous_backend.C10dRendezvousBackend.name

     

torch.distributed.elastic.rendezvous.c10d_rendezvous_backend.C10dRendezvousBackend.set_state

     

torch.distributed.elastic.rendezvous.etcd_rendezvous_backend.create_backend

     

torch.distributed.elastic.rendezvous.etcd_rendezvous_backend.EtcdRendezvousBackend

     

torch.distributed.elastic.rendezvous.etcd_rendezvous_backend.EtcdRendezvousBackend.get_state

     

torch.distributed.elastic.rendezvous.etcd_rendezvous_backend.EtcdRendezvousBackend.name

     

torch.distributed.elastic.rendezvous.etcd_rendezvous_backend.EtcdRendezvousBackend.set_state

     

torch.distributed.elastic.rendezvous.etcd_rendezvous.EtcdRendezvousHandler

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore.add

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore.check

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore.get

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore.set

     

torch.distributed.elastic.rendezvous.etcd_store.EtcdStore.wait

     

torch.distributed.elastic.rendezvous.etcd_server.EtcdServer

     

torch.distributed.elastic.timer.configure

     

torch.distributed.elastic.timer.expires

     

torch.distributed.elastic.timer.LocalTimerServer

     

torch.distributed.elastic.timer.LocalTimerClient

     

torch.distributed.elastic.timer.FileTimerServer

     

torch.distributed.elastic.timer.FileTimerClient

     

torch.distributed.elastic.timer.TimerRequest

     

torch.distributed.elastic.timer.TimerServer

     

torch.distributed.elastic.timer.TimerServer.clear_timers

     

torch.distributed.elastic.timer.TimerServer.get_expired_timers

     

torch.distributed.elastic.timer.TimerServer.register_timers

     

torch.distributed.elastic.timer.TimerClient

     

torch.distributed.elastic.timer.TimerClient.acquire

     

torch.distributed.elastic.timer.TimerClient.release

     

torch.distributed.elastic.metrics.api.MetricHandler

     

torch.distributed.elastic.metrics.api.ConsoleMetricHandler

     

torch.distributed.elastic.metrics.api.NullMetricHandler

     

torch.distributed.elastic.metrics.configure

     

torch.distributed.elastic.metrics.prof

     

torch.distributed.elastic.metrics.put_metric

     

torch.distributed.elastic.events.record

     

torch.distributed.elastic.events.get_logging_handler

     

torch.distributed.elastic.events.api.Event

     

torch.distributed.elastic.events.api.EventSource

     

torch.distributed.elastic.events.api.EventMetadataValue

     

torch.distributed.fsdp

PyTorch API

Compatibility

Limitations

torch.distributed.fsdp.FullyShardedDataParallel

     

torch.distributed.fsdp.FullyShardedDataParallel.apply

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.clip_grad_norm_

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.flatten_sharded_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.forward

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.fsdp_modules

     

torch.distributed.fsdp.FullyShardedDataParallel.full_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.get_state_dict_type

     

torch.distributed.fsdp.FullyShardedDataParallel.module

     

torch.distributed.fsdp.FullyShardedDataParallel.named_buffers

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.named_parameters

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.no_sync

     

torch.distributed.fsdp.FullyShardedDataParallel.optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.optim_state_dict_to_load

     

torch.distributed.fsdp.FullyShardedDataParallel.register_comm_hook

Y

  

torch.distributed.fsdp.FullyShardedDataParallel.rekey_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.scatter_full_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.set_state_dict_type

     

torch.distributed.fsdp.FullyShardedDataParallel.shard_full_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.sharded_optim_state_dict

     

torch.distributed.fsdp.FullyShardedDataParallel.state_dict_type

     

torch.distributed.fsdp.FullyShardedDataParallel.summon_full_params

     

torch.distributed.fsdp.BackwardPrefetch

Y

  

torch.distributed.fsdp.ShardingStrategy

Y

  

torch.distributed.fsdp.MixedPrecision

     

torch.distributed.fsdp.CPUOffload

Y

  

torch.distributed.fsdp.StateDictConfig

     

torch.distributed.fsdp.FullStateDictConfig

     

torch.distributed.fsdp.ShardedStateDictConfig

     

torch.distributed.fsdp.LocalStateDictConfig

     

torch.distributed.fsdp.OptimStateDictConfig

     

torch.distributed.fsdp.FullOptimStateDictConfig

     

torch.distributed.fsdp.ShardedOptimStateDictConfig

     

torch.distributed.fsdp.LocalOptimStateDictConfig

     

torch.distributed.fsdp.StateDictSettings

     

torch.distributed.optim

PyTorch API

Compatibility

Limitations

torch.distributed.optim.DistributedOptimizer

     

torch.distributed.optim.DistributedOptimizer.step

     

torch.distributed.optim.PostLocalSGDOptimizer

Y

  

torch.distributed.optim.PostLocalSGDOptimizer.load_state_dict

Y

  

torch.distributed.optim.PostLocalSGDOptimizer.state_dict

Y

  

torch.distributed.optim.PostLocalSGDOptimizer.step

Y

  

torch.distributed.optim.ZeroRedundancyOptimizer

Y

  

torch.distributed.optim.ZeroRedundancyOptimizer.add_param_group

Y

  

torch.distributed.optim.ZeroRedundancyOptimizer.consolidate_state_dict

     

torch.distributed.optim.ZeroRedundancyOptimizer.join_hook

     

torch.distributed.optim.ZeroRedundancyOptimizer.load_state_dict

     

torch.distributed.optim.ZeroRedundancyOptimizer.state_dict

Y

  

torch.distributed.optim.ZeroRedundancyOptimizer.step

     

torch.distributed.tensor.parallel

PyTorch API

Compatibility

Limitations

torch.distributed.tensor.parallel.parallelize_module

     

torch.distributed.tensor.parallel.style.RowwiseParallel

     

torch.distributed.tensor.parallel.style.ColwiseParallel

     

torch.distributed.tensor.parallel.style.PairwiseParallel

     

torch.distributed.tensor.parallel.style.SequenceParallel

     

torch.distributed.tensor.parallel.style.make_input_replicate_1d

     

torch.distributed.tensor.parallel.style.make_input_reshard_replicate

     

torch.distributed.tensor.parallel.style.make_input_shard_1d

     

torch.distributed.tensor.parallel.style.make_input_shard_1d_last_dim

     

torch.distributed.tensor.parallel.style.make_output_replicate_1d

     

torch.distributed.tensor.parallel.style.make_output_reshard_tensor

     

torch.distributed.tensor.parallel.style.make_output_shard_1d

     

torch.distributed.tensor.parallel.style.make_output_tensor

     

torch.distributed.tensor.parallel.fsdp.enable_2d_with_fsdp

     

torch.distributed.tensor.parallel.ddp.pre_dp_module_transform

     

torch.distributed.checkpoint

PyTorch API

Compatibility

Limitations

torch.distributed.checkpoint.load_state_dict

     

torch.distributed.checkpoint.save_state_dict

     

torch.distributed.checkpoint.StorageReader

     

torch.distributed.checkpoint.StorageReader.prepare_global_plan

     

torch.distributed.checkpoint.StorageReader.prepare_local_plan

     

torch.distributed.checkpoint.StorageReader.read_data

     

torch.distributed.checkpoint.StorageReader.read_metadata

     

torch.distributed.checkpoint.StorageReader.set_up_storage_reader

     

torch.distributed.checkpoint.StorageWriter

     

torch.distributed.checkpoint.StorageWriter.finish

     

torch.distributed.checkpoint.StorageWriter.prepare_global_plan

     

torch.distributed.checkpoint.StorageWriter.prepare_local_plan

     

torch.distributed.checkpoint.StorageWriter.set_up_storage_writer

     

torch.distributed.checkpoint.StorageWriter.write_data

     

torch.distributed.checkpoint.LoadPlanner

     

torch.distributed.checkpoint.LoadPlanner.commit_tensor

     

torch.distributed.checkpoint.LoadPlanner.create_global_plan

     

torch.distributed.checkpoint.LoadPlanner.create_local_plan

     

torch.distributed.checkpoint.LoadPlanner.finish_plan

     

torch.distributed.checkpoint.LoadPlanner.load_bytes

     

torch.distributed.checkpoint.LoadPlanner.resolve_tensor

     

torch.distributed.checkpoint.LoadPlanner.set_up_planner

     

torch.distributed.checkpoint.LoadPlan

     

torch.distributed.checkpoint.ReadItem

     

torch.distributed.checkpoint.SavePlanner

     

torch.distributed.checkpoint.SavePlanner.create_global_plan

     

torch.distributed.checkpoint.SavePlanner.create_local_plan

     

torch.distributed.checkpoint.SavePlanner.finish_plan

     

torch.distributed.checkpoint.SavePlanner.resolve_data

     

torch.distributed.checkpoint.SavePlanner.set_up_planner

     

torch.distributed.checkpoint.SavePlan

     

torch.distributed.checkpoint.WriteItem

     

torch.distributed.checkpoint.FileSystemReader

     

torch.distributed.checkpoint.FileSystemWriter

     

torch.distributed.checkpoint.DefaultSavePlanner

     

torch.distributed.checkpoint.DefaultSavePlanner.lookup_object

     

torch.distributed.checkpoint.DefaultSavePlanner.transform_object

     

torch.distributed.checkpoint.DefaultLoadPlanner

     

torch.distributed.checkpoint.DefaultLoadPlanner.lookup_tensor

     

torch.distributed.checkpoint.DefaultLoadPlanner.transform_tensor

     

torch.distributions

PyTorch API

Compatibility

Limitations

torch.distributions.distribution.Distribution

     

torch.distributions.distribution.Distribution.arg_constraints

     

torch.distributions.distribution.Distribution.batch_shape

     

torch.distributions.distribution.Distribution.cdf

     

torch.distributions.distribution.Distribution.entropy

     

torch.distributions.distribution.Distribution.enumerate_support

     

torch.distributions.distribution.Distribution.event_shape

     

torch.distributions.distribution.Distribution.expand

     

torch.distributions.distribution.Distribution.icdf

     

torch.distributions.distribution.Distribution.log_prob

     

torch.distributions.distribution.Distribution.mean

     

torch.distributions.distribution.Distribution.mode

     

torch.distributions.distribution.Distribution.perplexity

     

torch.distributions.distribution.Distribution.rsample

     

torch.distributions.distribution.Distribution.sample

     

torch.distributions.distribution.Distribution.sample_n

     

torch.distributions.distribution.Distribution.set_default_validate_args

     

torch.distributions.distribution.Distribution.stddev

     

torch.distributions.distribution.Distribution.support

     

torch.distributions.distribution.Distribution.variance

     

torch.distributions.exp_family.ExponentialFamily

     

torch.distributions.exp_family.ExponentialFamily.entropy

     

torch.distributions.bernoulli.Bernoulli

     

torch.distributions.bernoulli.Bernoulli.arg_constraints

     

torch.distributions.bernoulli.Bernoulli.entropy

     

torch.distributions.bernoulli.Bernoulli.enumerate_support

     

torch.distributions.bernoulli.Bernoulli.expand

     

torch.distributions.bernoulli.Bernoulli.has_enumerate_support

     

torch.distributions.bernoulli.Bernoulli.log_prob

     

torch.distributions.bernoulli.Bernoulli.logits

     

torch.distributions.bernoulli.Bernoulli.mean

     

torch.distributions.bernoulli.Bernoulli.mode

     

torch.distributions.bernoulli.Bernoulli.param_shape

     

torch.distributions.bernoulli.Bernoulli.probs

     

torch.distributions.bernoulli.Bernoulli.sample

     

torch.distributions.bernoulli.Bernoulli.support

     

torch.distributions.bernoulli.Bernoulli.variance

     

torch.distributions.beta.Beta

     

torch.distributions.beta.Beta.arg_constraints

     

torch.distributions.beta.Beta.concentration0

     

torch.distributions.beta.Beta.concentration1

     

torch.distributions.beta.Beta.entropy

     

torch.distributions.beta.Beta.expand

     

torch.distributions.beta.Beta.has_rsample

     

torch.distributions.beta.Beta.log_prob

     

torch.distributions.beta.Beta.mean

     

torch.distributions.beta.Beta.mode

     

torch.distributions.beta.Beta.rsample

     

torch.distributions.beta.Beta.support

     

torch.distributions.beta.Beta.variance

     

torch.distributions.binomial.Binomial

     

torch.distributions.binomial.Binomial.arg_constraints

     

torch.distributions.binomial.Binomial.entropy

     

torch.distributions.binomial.Binomial.enumerate_support

     

torch.distributions.binomial.Binomial.expand

     

torch.distributions.binomial.Binomial.has_enumerate_support

     

torch.distributions.binomial.Binomial.log_prob

     

torch.distributions.binomial.Binomial.logits

     

torch.distributions.binomial.Binomial.mean

     

torch.distributions.binomial.Binomial.mode

     

torch.distributions.binomial.Binomial.param_shape

     

torch.distributions.binomial.Binomial.probs

     

torch.distributions.binomial.Binomial.sample

     

torch.distributions.binomial.Binomial.support

     

torch.distributions.binomial.Binomial.variance

     

torch.distributions.categorical.Categorical

     

torch.distributions.categorical.Categorical.arg_constraints

     

torch.distributions.categorical.Categorical.entropy

     

torch.distributions.categorical.Categorical.enumerate_support

     

torch.distributions.categorical.Categorical.expand

     

torch.distributions.categorical.Categorical.has_enumerate_support

     

torch.distributions.categorical.Categorical.log_prob

     

torch.distributions.categorical.Categorical.logits

     

torch.distributions.categorical.Categorical.mean

     

torch.distributions.categorical.Categorical.mode

     

torch.distributions.categorical.Categorical.param_shape

     

torch.distributions.categorical.Categorical.probs

     

torch.distributions.categorical.Categorical.sample

     

torch.distributions.categorical.Categorical.support

     

torch.distributions.categorical.Categorical.variance

     

torch.distributions.cauchy.Cauchy

     

torch.distributions.cauchy.Cauchy.arg_constraints

     

torch.distributions.cauchy.Cauchy.cdf

     

torch.distributions.cauchy.Cauchy.entropy

     

torch.distributions.cauchy.Cauchy.expand

     

torch.distributions.cauchy.Cauchy.has_rsample

     

torch.distributions.cauchy.Cauchy.icdf

     

torch.distributions.cauchy.Cauchy.log_prob

     

torch.distributions.cauchy.Cauchy.mean

     

torch.distributions.cauchy.Cauchy.mode

     

torch.distributions.cauchy.Cauchy.rsample

     

torch.distributions.cauchy.Cauchy.support

     

torch.distributions.cauchy.Cauchy.variance

     

torch.distributions.chi2.Chi2

     

torch.distributions.chi2.Chi2.arg_constraints

     

torch.distributions.chi2.Chi2.df

     

torch.distributions.chi2.Chi2.expand

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.arg_constraints

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.cdf

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.entropy

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.expand

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.has_rsample

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.icdf

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.log_prob

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.logits

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.mean

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.param_shape

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.probs

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.rsample

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.sample

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.stddev

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.support

     

torch.distributions.continuous_bernoulli.ContinuousBernoulli.variance

     

torch.distributions.dirichlet.Dirichlet

     

torch.distributions.dirichlet.Dirichlet.arg_constraints

     

torch.distributions.dirichlet.Dirichlet.entropy

     

torch.distributions.dirichlet.Dirichlet.expand

     

torch.distributions.dirichlet.Dirichlet.has_rsample

     

torch.distributions.dirichlet.Dirichlet.log_prob

     

torch.distributions.dirichlet.Dirichlet.mean

     

torch.distributions.dirichlet.Dirichlet.mode

     

torch.distributions.dirichlet.Dirichlet.rsample

     

torch.distributions.dirichlet.Dirichlet.support

     

torch.distributions.dirichlet.Dirichlet.variance

     

torch.distributions.exponential.Exponential

     

torch.distributions.exponential.Exponential.arg_constraints

     

torch.distributions.exponential.Exponential.cdf

     

torch.distributions.exponential.Exponential.entropy

     

torch.distributions.exponential.Exponential.expand

     

torch.distributions.exponential.Exponential.has_rsample

     

torch.distributions.exponential.Exponential.icdf

     

torch.distributions.exponential.Exponential.log_prob

     

torch.distributions.exponential.Exponential.mean

     

torch.distributions.exponential.Exponential.mode

     

torch.distributions.exponential.Exponential.rsample

     

torch.distributions.exponential.Exponential.stddev

     

torch.distributions.exponential.Exponential.support

     

torch.distributions.exponential.Exponential.variance

     

torch.distributions.fishersnedecor.FisherSnedecor

     

torch.distributions.fishersnedecor.FisherSnedecor.arg_constraints

     

torch.distributions.fishersnedecor.FisherSnedecor.expand

     

torch.distributions.fishersnedecor.FisherSnedecor.has_rsample

     

torch.distributions.fishersnedecor.FisherSnedecor.log_prob

     

torch.distributions.fishersnedecor.FisherSnedecor.mean

     

torch.distributions.fishersnedecor.FisherSnedecor.mode

     

torch.distributions.fishersnedecor.FisherSnedecor.rsample

     

torch.distributions.fishersnedecor.FisherSnedecor.support

     

torch.distributions.fishersnedecor.FisherSnedecor.variance

     

torch.distributions.gamma.Gamma

     

torch.distributions.gamma.Gamma.arg_constraints

     

torch.distributions.gamma.Gamma.cdf

     

torch.distributions.gamma.Gamma.entropy

     

torch.distributions.gamma.Gamma.expand

     

torch.distributions.gamma.Gamma.has_rsample

     

torch.distributions.gamma.Gamma.log_prob

     

torch.distributions.gamma.Gamma.mean

     

torch.distributions.gamma.Gamma.mode

     

torch.distributions.gamma.Gamma.rsample

     

torch.distributions.gamma.Gamma.support

     

torch.distributions.gamma.Gamma.variance

     

torch.distributions.geometric.Geometric

     

torch.distributions.geometric.Geometric.arg_constraints

     

torch.distributions.geometric.Geometric.entropy

     

torch.distributions.geometric.Geometric.expand

     

torch.distributions.geometric.Geometric.log_prob

     

torch.distributions.geometric.Geometric.logits

     

torch.distributions.geometric.Geometric.mean

     

torch.distributions.geometric.Geometric.mode

     

torch.distributions.geometric.Geometric.probs

     

torch.distributions.geometric.Geometric.sample

     

torch.distributions.geometric.Geometric.support

     

torch.distributions.geometric.Geometric.variance

     

torch.distributions.gumbel.Gumbel

     

torch.distributions.gumbel.Gumbel.arg_constraints

     

torch.distributions.gumbel.Gumbel.entropy

     

torch.distributions.gumbel.Gumbel.expand

     

torch.distributions.gumbel.Gumbel.log_prob

     

torch.distributions.gumbel.Gumbel.mean

     

torch.distributions.gumbel.Gumbel.mode

     

torch.distributions.gumbel.Gumbel.stddev

     

torch.distributions.gumbel.Gumbel.support

     

torch.distributions.gumbel.Gumbel.variance

     

torch.distributions.half_cauchy.HalfCauchy

     

torch.distributions.half_cauchy.HalfCauchy.arg_constraints

     

torch.distributions.half_cauchy.HalfCauchy.cdf

     

torch.distributions.half_cauchy.HalfCauchy.entropy

     

torch.distributions.half_cauchy.HalfCauchy.expand

     

torch.distributions.half_cauchy.HalfCauchy.has_rsample

     

torch.distributions.half_cauchy.HalfCauchy.icdf

     

torch.distributions.half_cauchy.HalfCauchy.log_prob

     

torch.distributions.half_cauchy.HalfCauchy.mean

     

torch.distributions.half_cauchy.HalfCauchy.mode

     

torch.distributions.half_cauchy.HalfCauchy.scale

     

torch.distributions.half_cauchy.HalfCauchy.support

     

torch.distributions.half_cauchy.HalfCauchy.variance

     

torch.distributions.half_normal.HalfNormal

     

torch.distributions.half_normal.HalfNormal.arg_constraints

     

torch.distributions.half_normal.HalfNormal.cdf

     

torch.distributions.half_normal.HalfNormal.entropy

     

torch.distributions.half_normal.HalfNormal.expand

     

torch.distributions.half_normal.HalfNormal.has_rsample

     

torch.distributions.half_normal.HalfNormal.icdf

     

torch.distributions.half_normal.HalfNormal.log_prob

     

torch.distributions.half_normal.HalfNormal.mean

     

torch.distributions.half_normal.HalfNormal.mode

     

torch.distributions.half_normal.HalfNormal.scale

     

torch.distributions.half_normal.HalfNormal.support

     

torch.distributions.half_normal.HalfNormal.variance

     

torch.distributions.independent.Independent

     

torch.distributions.independent.Independent.arg_constraints

     

torch.distributions.independent.Independent.entropy

     

torch.distributions.independent.Independent.enumerate_support

     

torch.distributions.independent.Independent.expand

     

torch.distributions.independent.Independent.has_enumerate_support

     

torch.distributions.independent.Independent.has_rsample

     

torch.distributions.independent.Independent.log_prob

     

torch.distributions.independent.Independent.mean

     

torch.distributions.independent.Independent.mode

     

torch.distributions.independent.Independent.rsample

     

torch.distributions.independent.Independent.sample

     

torch.distributions.independent.Independent.support

     

torch.distributions.independent.Independent.variance

     

torch.distributions.kumaraswamy.Kumaraswamy

     

torch.distributions.kumaraswamy.Kumaraswamy.arg_constraints

     

torch.distributions.kumaraswamy.Kumaraswamy.entropy

     

torch.distributions.kumaraswamy.Kumaraswamy.expand

     

torch.distributions.kumaraswamy.Kumaraswamy.has_rsample

     

torch.distributions.kumaraswamy.Kumaraswamy.mean

     

torch.distributions.kumaraswamy.Kumaraswamy.mode

     

torch.distributions.kumaraswamy.Kumaraswamy.support

     

torch.distributions.kumaraswamy.Kumaraswamy.variance

     

torch.distributions.lkj_cholesky.LKJCholesky

     

torch.distributions.lkj_cholesky.LKJCholesky.arg_constraints

     

torch.distributions.lkj_cholesky.LKJCholesky.expand

     

torch.distributions.lkj_cholesky.LKJCholesky.log_prob

     

torch.distributions.lkj_cholesky.LKJCholesky.sample

     

torch.distributions.lkj_cholesky.LKJCholesky.support

     

torch.distributions.laplace.Laplace

     

torch.distributions.laplace.Laplace.arg_constraints

     

torch.distributions.laplace.Laplace.cdf

     

torch.distributions.laplace.Laplace.entropy

     

torch.distributions.laplace.Laplace.expand

     

torch.distributions.laplace.Laplace.has_rsample

     

torch.distributions.laplace.Laplace.icdf

     

torch.distributions.laplace.Laplace.log_prob

     

torch.distributions.laplace.Laplace.mean

     

torch.distributions.laplace.Laplace.mode

     

torch.distributions.laplace.Laplace.rsample

     

torch.distributions.laplace.Laplace.stddev

     

torch.distributions.laplace.Laplace.support

     

torch.distributions.laplace.Laplace.variance

     

torch.distributions.log_normal.LogNormal

     

torch.distributions.log_normal.LogNormal.arg_constraints

     

torch.distributions.log_normal.LogNormal.entropy

     

torch.distributions.log_normal.LogNormal.expand

     

torch.distributions.log_normal.LogNormal.has_rsample

     

torch.distributions.log_normal.LogNormal.loc

     

torch.distributions.log_normal.LogNormal.mean

     

torch.distributions.log_normal.LogNormal.mode

     

torch.distributions.log_normal.LogNormal.scale

     

torch.distributions.log_normal.LogNormal.support

     

torch.distributions.log_normal.LogNormal.variance

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.arg_constraints

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.covariance_matrix

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.entropy

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.expand

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.has_rsample

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.log_prob

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.mean

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.mode

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.precision_matrix

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.rsample

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.scale_tril

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.support

     

torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal.variance

     

torch.distributions.mixture_same_family.MixtureSameFamily

     

torch.distributions.mixture_same_family.MixtureSameFamily.arg_constraints

     

torch.distributions.mixture_same_family.MixtureSameFamily.cdf

     

torch.distributions.mixture_same_family.MixtureSameFamily.component_distribution

     

torch.distributions.mixture_same_family.MixtureSameFamily.expand

     

torch.distributions.mixture_same_family.MixtureSameFamily.has_rsample

     

torch.distributions.mixture_same_family.MixtureSameFamily.log_prob

     

torch.distributions.mixture_same_family.MixtureSameFamily.mean

     

torch.distributions.mixture_same_family.MixtureSameFamily.mixture_distribution

     

torch.distributions.mixture_same_family.MixtureSameFamily.sample

     

torch.distributions.mixture_same_family.MixtureSameFamily.support

     

torch.distributions.mixture_same_family.MixtureSameFamily.variance

     

torch.distributions.multinomial.Multinomial

     

torch.distributions.multinomial.Multinomial.arg_constraints

     

torch.distributions.multinomial.Multinomial.entropy

     

torch.distributions.multinomial.Multinomial.expand

     

torch.distributions.multinomial.Multinomial.log_prob

     

torch.distributions.multinomial.Multinomial.logits

     

torch.distributions.multinomial.Multinomial.mean

     

torch.distributions.multinomial.Multinomial.param_shape

     

torch.distributions.multinomial.Multinomial.probs

     

torch.distributions.multinomial.Multinomial.sample

     

torch.distributions.multinomial.Multinomial.support

     

torch.distributions.multinomial.Multinomial.total_count

     

torch.distributions.multinomial.Multinomial.variance

     

torch.distributions.multivariate_normal.MultivariateNormal

     

torch.distributions.multivariate_normal.MultivariateNormal.arg_constraints

     

torch.distributions.multivariate_normal.MultivariateNormal.covariance_matrix

     

torch.distributions.multivariate_normal.MultivariateNormal.entropy

     

torch.distributions.multivariate_normal.MultivariateNormal.expand

     

torch.distributions.multivariate_normal.MultivariateNormal.has_rsample

     

torch.distributions.multivariate_normal.MultivariateNormal.log_prob

     

torch.distributions.multivariate_normal.MultivariateNormal.mean

     

torch.distributions.multivariate_normal.MultivariateNormal.mode

     

torch.distributions.multivariate_normal.MultivariateNormal.precision_matrix

     

torch.distributions.multivariate_normal.MultivariateNormal.rsample

     

torch.distributions.multivariate_normal.MultivariateNormal.scale_tril

     

torch.distributions.multivariate_normal.MultivariateNormal.support

     

torch.distributions.multivariate_normal.MultivariateNormal.variance

     

torch.distributions.negative_binomial.NegativeBinomial

     

torch.distributions.negative_binomial.NegativeBinomial.arg_constraints

     

torch.distributions.negative_binomial.NegativeBinomial.expand

     

torch.distributions.negative_binomial.NegativeBinomial.log_prob

     

torch.distributions.negative_binomial.NegativeBinomial.logits

     

torch.distributions.negative_binomial.NegativeBinomial.mean

     

torch.distributions.negative_binomial.NegativeBinomial.mode

     

torch.distributions.negative_binomial.NegativeBinomial.param_shape

     

torch.distributions.negative_binomial.NegativeBinomial.probs

     

torch.distributions.negative_binomial.NegativeBinomial.sample

     

torch.distributions.negative_binomial.NegativeBinomial.support

     

torch.distributions.negative_binomial.NegativeBinomial.variance

     

torch.distributions.normal.Normal

     

torch.distributions.normal.Normal.arg_constraints

     

torch.distributions.normal.Normal.cdf

     

torch.distributions.normal.Normal.entropy

     

torch.distributions.normal.Normal.expand

     

torch.distributions.normal.Normal.has_rsample

     

torch.distributions.normal.Normal.icdf

     

torch.distributions.normal.Normal.log_prob

     

torch.distributions.normal.Normal.mean

     

torch.distributions.normal.Normal.mode

     

torch.distributions.normal.Normal.rsample

     

torch.distributions.normal.Normal.sample

     

torch.distributions.normal.Normal.stddev

     

torch.distributions.normal.Normal.support

     

torch.distributions.normal.Normal.variance

     

torch.distributions.one_hot_categorical.OneHotCategorical

     

torch.distributions.one_hot_categorical.OneHotCategorical.arg_constraints

     

torch.distributions.one_hot_categorical.OneHotCategorical.entropy

     

torch.distributions.one_hot_categorical.OneHotCategorical.enumerate_support

     

torch.distributions.one_hot_categorical.OneHotCategorical.expand

     

torch.distributions.one_hot_categorical.OneHotCategorical.has_enumerate_support

     

torch.distributions.one_hot_categorical.OneHotCategorical.log_prob

     

torch.distributions.one_hot_categorical.OneHotCategorical.logits

     

torch.distributions.one_hot_categorical.OneHotCategorical.mean

     

torch.distributions.one_hot_categorical.OneHotCategorical.mode

     

torch.distributions.one_hot_categorical.OneHotCategorical.param_shape

     

torch.distributions.one_hot_categorical.OneHotCategorical.probs

     

torch.distributions.one_hot_categorical.OneHotCategorical.sample

     

torch.distributions.one_hot_categorical.OneHotCategorical.support

     

torch.distributions.one_hot_categorical.OneHotCategorical.variance

     

torch.distributions.pareto.Pareto

     

torch.distributions.pareto.Pareto.arg_constraints

     

torch.distributions.pareto.Pareto.entropy

     

torch.distributions.pareto.Pareto.expand

     

torch.distributions.pareto.Pareto.mean

     

torch.distributions.pareto.Pareto.mode

     

torch.distributions.pareto.Pareto.support

     

torch.distributions.pareto.Pareto.variance

     

torch.distributions.poisson.Poisson

     

torch.distributions.poisson.Poisson.arg_constraints

     

torch.distributions.poisson.Poisson.expand

     

torch.distributions.poisson.Poisson.log_prob

     

torch.distributions.poisson.Poisson.mean

     

torch.distributions.poisson.Poisson.mode

     

torch.distributions.poisson.Poisson.sample

     

torch.distributions.poisson.Poisson.support

     

torch.distributions.poisson.Poisson.variance

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.arg_constraints

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.expand

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.has_rsample

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.logits

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.probs

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.support

     

torch.distributions.relaxed_bernoulli.RelaxedBernoulli.temperature

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.arg_constraints

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.expand

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.log_prob

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.logits

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.param_shape

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.probs

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.rsample

     

torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli.support

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.arg_constraints

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.expand

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.has_rsample

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.logits

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.probs

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.support

     

torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.temperature

     

torch.distributions.studentT.StudentT

     

torch.distributions.studentT.StudentT.arg_constraints

     

torch.distributions.studentT.StudentT.entropy

     

torch.distributions.studentT.StudentT.expand

     

torch.distributions.studentT.StudentT.has_rsample

     

torch.distributions.studentT.StudentT.log_prob

     

torch.distributions.studentT.StudentT.mean

     

torch.distributions.studentT.StudentT.mode

     

torch.distributions.studentT.StudentT.rsample

     

torch.distributions.studentT.StudentT.support

     

torch.distributions.studentT.StudentT.variance

     

torch.distributions.transformed_distribution.TransformedDistribution

     

torch.distributions.transformed_distribution.TransformedDistribution.arg_constraints

     

torch.distributions.transformed_distribution.TransformedDistribution.cdf

     

torch.distributions.transformed_distribution.TransformedDistribution.expand

     

torch.distributions.transformed_distribution.TransformedDistribution.has_rsample

     

torch.distributions.transformed_distribution.TransformedDistribution.icdf

     

torch.distributions.transformed_distribution.TransformedDistribution.log_prob

     

torch.distributions.transformed_distribution.TransformedDistribution.rsample

     

torch.distributions.transformed_distribution.TransformedDistribution.sample

     

torch.distributions.transformed_distribution.TransformedDistribution.support

     

torch.distributions.uniform.Uniform

     

torch.distributions.uniform.Uniform.arg_constraints

     

torch.distributions.uniform.Uniform.cdf

     

torch.distributions.uniform.Uniform.entropy

     

torch.distributions.uniform.Uniform.expand

     

torch.distributions.uniform.Uniform.has_rsample

     

torch.distributions.uniform.Uniform.icdf

     

torch.distributions.uniform.Uniform.log_prob

     

torch.distributions.uniform.Uniform.mean

     

torch.distributions.uniform.Uniform.mode

     

torch.distributions.uniform.Uniform.rsample

     

torch.distributions.uniform.Uniform.stddev

     

torch.distributions.uniform.Uniform.support

     

torch.distributions.uniform.Uniform.variance

     

torch.distributions.von_mises.VonMises

     

torch.distributions.von_mises.VonMises.arg_constraints

     

torch.distributions.von_mises.VonMises.expand

     

torch.distributions.von_mises.VonMises.has_rsample

     

torch.distributions.von_mises.VonMises.log_prob

     

torch.distributions.von_mises.VonMises.mean

     

torch.distributions.von_mises.VonMises.mode

     

torch.distributions.von_mises.VonMises.sample

     

torch.distributions.von_mises.VonMises.support

     

torch.distributions.von_mises.VonMises.variance

     

torch.distributions.weibull.Weibull

     

torch.distributions.weibull.Weibull.arg_constraints

     

torch.distributions.weibull.Weibull.entropy

     

torch.distributions.weibull.Weibull.expand

     

torch.distributions.weibull.Weibull.mean

     

torch.distributions.weibull.Weibull.mode

     

torch.distributions.weibull.Weibull.support

     

torch.distributions.weibull.Weibull.variance

     

torch.distributions.wishart.Wishart

     

torch.distributions.wishart.Wishart.arg_constraints

     

torch.distributions.wishart.Wishart.covariance_matrix

     

torch.distributions.wishart.Wishart.entropy

     

torch.distributions.wishart.Wishart.expand

     

torch.distributions.wishart.Wishart.has_rsample

     

torch.distributions.wishart.Wishart.log_prob

     

torch.distributions.wishart.Wishart.mean

     

torch.distributions.wishart.Wishart.mode

     

torch.distributions.wishart.Wishart.precision_matrix

     

torch.distributions.wishart.Wishart.rsample

     

torch.distributions.wishart.Wishart.scale_tril

     

torch.distributions.wishart.Wishart.support

     

torch.distributions.wishart.Wishart.variance

     

torch.distributions.kl.kl_divergence

     

torch.distributions.kl.register_kl

     

torch.distributions.transforms.AbsTransform

     

torch.distributions.transforms.AffineTransform

     

torch.distributions.transforms.CatTransform

     

torch.distributions.transforms.ComposeTransform

     

torch.distributions.transforms.CorrCholeskyTransform

     

torch.distributions.transforms.CumulativeDistributionTransform

     

torch.distributions.transforms.ExpTransform

     

torch.distributions.transforms.IndependentTransform

     

torch.distributions.transforms.LowerCholeskyTransform

     

torch.distributions.transforms.PositiveDefiniteTransform

     

torch.distributions.transforms.PowerTransform

     

torch.distributions.transforms.ReshapeTransform

     

torch.distributions.transforms.SigmoidTransform

     

torch.distributions.transforms.SoftplusTransform

     

torch.distributions.transforms.TanhTransform

     

torch.distributions.transforms.SoftmaxTransform

     

torch.distributions.transforms.StackTransform

     

torch.distributions.transforms.StickBreakingTransform

     

torch.distributions.transforms.Transform

     

torch.distributions.transforms.Transform.inv

     

torch.distributions.transforms.Transform.sign

     

torch.distributions.transforms.Transform.log_abs_det_jacobian

     

torch.distributions.transforms.Transform.forward_shape

     

torch.distributions.transforms.Transform.inverse_shape

     

torch.distributions.constraints.Constraint

     

torch.distributions.constraints.Constraint.check

     

torch.distributions.constraints.cat

     

torch.distributions.constraints.dependent_property

     

torch.distributions.constraints.greater_than

     

torch.distributions.constraints.greater_than_eq

     

torch.distributions.constraints.independent

     

torch.distributions.constraints.integer_interval

     

torch.distributions.constraints.interval

     

torch.distributions.constraints.half_open_interval

     

torch.distributions.constraints.less_than

     

torch.distributions.constraints.multinomial

     

torch.distributions.constraints.stack

     

torch.distributions.constraint_registry.ConstraintRegistry

     

torch.distributions.constraint_registry.ConstraintRegistry.register

     

torch.fft

PyTorch API

Compatibility

Limitations

torch.fft.fft

     

torch.fft.ifft

     

torch.fft.fft2

     

torch.fft.ifft2

     

torch.fft.fftn

     

torch.fft.ifftn

     

torch.fft.rfft

     

torch.fft.irfft

     

torch.fft.rfft2

     

torch.fft.irfft2

     

torch.fft.rfftn

     

torch.fft.irfftn

     

torch.fft.hfft

     

torch.fft.ihfft

     

torch.fft.hfft2

     

torch.fft.ihfft2

     

torch.fft.hfftn

     

torch.fft.ihfftn

     

torch.fft.fftfreq

     

torch.fft.rfftfreq

     

torch.fft.fftshift

     

torch.fft.ifftshift

     

torch.func

PyTorch API

Compatibility

Limitations

torch.func.vmap

     

torch.func.grad

     

torch.func.grad_and_value

     

torch.func.vjp

     

torch.func.jvp

     

torch.func.linearize

     

torch.func.jacrev

     

torch.func.jacfwd

     

torch.func.hessian

     

torch.func.functionalize

     

torch.func.functional_call

     

torch.func.stack_module_state

     

torch.func.replace_all_batch_norm_modules_

     

torch.futures

PyTorch API

Compatibility

Limitations

torch.futures.Future

Y

  

torch.futures.Future.add_done_callback

Y

  

torch.futures.Future.done

Y

  

torch.futures.Future.set_exception

Y

  

torch.futures.Future.set_result

Y

  

torch.futures.Future.then

Y

  

torch.futures.Future.value

Y

  

torch.futures.Future.wait

Y

  

torch.futures.collect_all

Y

  

torch.futures.wait_all

Y

  

torch.fx

PyTorch API

Compatibility

Limitations

torch.fx.symbolic_trace

Y

  

torch.fx.wrap

Y

  

torch.fx.GraphModule

Y

  

torch.fx.GraphModule.__init__

Y

  

torch.fx.GraphModule.add_submodule

Y

  

torch.fx.GraphModule.code

     

torch.fx.GraphModule.delete_all_unused_submodules

Y

  

torch.fx.GraphModule.delete_submodule

Y

  

torch.fx.GraphModule.graph

     

torch.fx.GraphModule.print_readable

Y

  

torch.fx.GraphModule.recompile

Y

  

torch.fx.GraphModule.to_folder

Y

  

torch.fx.Graph

Y

  

torch.fx.Graph.__init__

Y

  

torch.fx.Graph.call_function

Y

  

torch.fx.Graph.call_method

Y

  

torch.fx.Graph.call_module

Y

  

torch.fx.Graph.create_node

Y

  

torch.fx.Graph.eliminate_dead_code

Y

  

torch.fx.Graph.erase_node

Y

  

torch.fx.Graph.get_attr

Y

  

torch.fx.Graph.graph_copy

Y

  

torch.fx.Graph.inserting_after

Y

  

torch.fx.Graph.inserting_before

Y

  

torch.fx.Graph.lint

Y

  

torch.fx.Graph.node_copy

Y

  

torch.fx.Graph.nodes

     

torch.fx.Graph.on_generate_code

Y

  

torch.fx.Graph.output

Y

  

torch.fx.Graph.placeholder

Y

  

torch.fx.Graph.print_tabular

Y

  

torch.fx.Graph.process_inputs

Y

  

torch.fx.Graph.process_outputs

Y

  

torch.fx.Graph.python_code

Y

  

torch.fx.Graph.set_codegen

Y

  

torch.fx.Node

Y

  

torch.fx.Node.all_input_nodes

     

torch.fx.Node.append

Y

  

torch.fx.Node.args

     

torch.fx.Node.format_node

Y

  

torch.fx.Node.is_impure

Y

  

torch.fx.Node.kwargs

     

torch.fx.Node.next

     

torch.fx.Node.normalized_arguments

Y

  

torch.fx.Node.prepend

Y

  

torch.fx.Node.prev

     

torch.fx.Node.replace_all_uses_with

Y

  

torch.fx.Node.replace_input_with

Y

  

torch.fx.Node.stack_trace

     

torch.fx.Node.update_arg

Y

  

torch.fx.Node.update_kwarg

Y

  

torch.fx.Tracer

Y

  

torch.fx.Tracer.call_module

Y

  

torch.fx.Tracer.create_arg

Y

  

torch.fx.Tracer.create_args_for_root

Y

  

torch.fx.Tracer.create_node

Y

  

torch.fx.Tracer.create_proxy

Y

  

torch.fx.Tracer.getattr

Y

  

torch.fx.Tracer.is_leaf_module

Y

  

torch.fx.Tracer.iter

Y

  

torch.fx.Tracer.keys

Y

  

torch.fx.Tracer.path_of_module

Y

  

torch.fx.Tracer.proxy

Y

  

torch.fx.Tracer.to_bool

Y

  

torch.fx.Tracer.trace

Y

  

torch.fx.Proxy

Y

  

torch.fx.Interpreter

Y

  

torch.fx.Interpreter.boxed_run

     

torch.fx.Interpreter.call_function

Y

  

torch.fx.Interpreter.call_method

Y

  

torch.fx.Interpreter.call_module

Y

  

torch.fx.Interpreter.fetch_args_kwargs_from_env

Y

  

torch.fx.Interpreter.fetch_attr

Y

  

torch.fx.Interpreter.get_attr

Y

  

torch.fx.Interpreter.map_nodes_to_values

Y

  

torch.fx.Interpreter.output

Y

  

torch.fx.Interpreter.placeholder

Y

  

torch.fx.Interpreter.run

Y

  

torch.fx.Interpreter.run_node

Y

  

torch.fx.Transformer

Y

  

torch.fx.Transformer.call_function

Y

  

torch.fx.Transformer.call_module

Y

  

torch.fx.Transformer.get_attr

Y

  

torch.fx.Transformer.placeholder

Y

  

torch.fx.Transformer.transform

Y

  

torch.fx.replace_pattern

Y

  

torch.hub

PyTorch API

Compatibility

Limitations

torch.hub.list

Y

  

torch.hub.help

Y

  

torch.hub.load

Y

  

torch.hub.download_url_to_file

Y

  

torch.hub.load_state_dict_from_url

Y

  

torch.hub.get_dir

Y

  

torch.hub.set_dir

Y

  

torch.jit

PyTorch API

Compatibility

Limitations

torch.jit.export

Y

  

torch.jit.is_scripting

Y

  

torch.jit.is_tracing

Y

  

torch.jit.script

     

torch.jit.trace

Y

  

torch.jit.script_if_tracing

Y

  

torch.jit.trace_module

Y

  

torch.jit.fork

     

torch.jit.wait

     

torch.jit.ScriptModule

Y

  

torch.jit.ScriptModule.add_module

     

torch.jit.ScriptModule.apply

Y

  

torch.jit.ScriptModule.bfloat16

     

torch.jit.ScriptModule.buffers

Y

  

torch.jit.ScriptModule.children

Y

  

torch.jit.ScriptModule.code

     

torch.jit.ScriptModule.code_with_constants

     

torch.jit.ScriptModule.compile

     

torch.jit.ScriptModule.cpu

Y

  

torch.jit.ScriptModule.cuda

Y

  

torch.jit.ScriptModule.double

Y

  

torch.jit.ScriptModule.eval

Y

  

torch.jit.ScriptModule.extra_repr

Y

  

torch.jit.ScriptModule.float

Y

  

torch.jit.ScriptModule.get_buffer

Y

  

torch.jit.ScriptModule.get_extra_state

     

torch.jit.ScriptModule.get_parameter

Y

  

torch.jit.ScriptModule.get_submodule

Y

  

torch.jit.ScriptModule.graph

     

torch.jit.ScriptModule.half

Y

  

torch.jit.ScriptModule.inlined_graph

     

torch.jit.ScriptModule.ipu

Y

  

torch.jit.ScriptModule.load_state_dict

Y

  

torch.jit.ScriptModule.modules

Y

  

torch.jit.ScriptModule.named_buffers

Y

  

torch.jit.ScriptModule.named_children

Y

  

torch.jit.ScriptModule.named_modules

Y

  

torch.jit.ScriptModule.named_parameters

Y

  

torch.jit.ScriptModule.parameters

Y

  

torch.jit.ScriptModule.register_backward_hook

Y

  

torch.jit.ScriptModule.register_buffer

     

torch.jit.ScriptModule.register_forward_hook

Y

  

torch.jit.ScriptModule.register_forward_pre_hook

Y

  

torch.jit.ScriptModule.register_full_backward_hook

Y

  

torch.jit.ScriptModule.register_full_backward_pre_hook

Y

  

torch.jit.ScriptModule.register_load_state_dict_post_hook

Y

  

torch.jit.ScriptModule.register_module

     

torch.jit.ScriptModule.register_parameter

     

torch.jit.ScriptModule.register_state_dict_pre_hook

Y

  

torch.jit.ScriptModule.requires_grad_

Y

  

torch.jit.ScriptModule.save

Y

  

torch.jit.ScriptModule.set_extra_state

     

torch.jit.ScriptModule.share_memory

     

torch.jit.ScriptModule.state_dict

Y

  

torch.jit.ScriptModule.to

Y

  

torch.jit.ScriptModule.to_empty

Y

  

torch.jit.ScriptModule.train

Y

  

torch.jit.ScriptModule.type

Y

  

torch.jit.ScriptModule.xpu

Y

  

torch.jit.ScriptModule.zero_grad

Y

  

torch.jit.ScriptFunction

Y

  

torch.jit.ScriptFunction.get_debug_state

Y

  

torch.jit.ScriptFunction.save

Y

  

torch.jit.ScriptFunction.save_to_buffer

Y

  

torch.jit.freeze

Y

  

torch.jit.optimize_for_inference

Y

  

torch.jit.enable_onednn_fusion

Y

  

torch.jit.onednn_fusion_enabled

Y

  

torch.jit.set_fusion_strategy

     

torch.jit.strict_fusion

     

torch.jit.save

Y

  

torch.jit.load

Y

  

torch.jit.ignore

Y

  

torch.jit.unused

Y

  

torch.jit.isinstance

Y

  

torch.jit.Attribute

Y

  

torch.jit.Attribute.count

Y

  

torch.jit.Attribute.index

Y

  

torch.jit.Attribute.type

     

torch.jit.Attribute.value

     

torch.jit.annotate

Y

  

torch.linalg

PyTorch API

Compatibility

Limitations

torch.linalg.norm

Y

Only support fp16,fp32,int64,bool

torch.linalg.vector_norm

     

torch.linalg.matrix_norm

     

torch.linalg.diagonal

     

torch.linalg.det

     

torch.linalg.slogdet

     

torch.linalg.cond

     

torch.linalg.matrix_rank

     

torch.linalg.cholesky

     

torch.linalg.qr

     

torch.linalg.lu

     

torch.linalg.lu_factor

     

torch.linalg.eig

     

torch.linalg.eigvals

     

torch.linalg.eigh

     

torch.linalg.eigvalsh

     

torch.linalg.svd

     

torch.linalg.svdvals

     

torch.linalg.solve

     

torch.linalg.solve_triangular

     

torch.linalg.lu_solve

     

torch.linalg.lstsq

     

torch.linalg.inv

     

torch.linalg.pinv

     

torch.linalg.matrix_exp

     

torch.linalg.matrix_power

     

torch.linalg.cross

     

torch.linalg.matmul

     

torch.linalg.vecdot

     

torch.linalg.multi_dot

     

torch.linalg.householder_product

     

torch.linalg.tensorinv

     

torch.linalg.tensorsolve

     

torch.linalg.vander

     

torch.linalg.cholesky_ex

     

torch.linalg.inv_ex

     

torch.linalg.solve_ex

     

torch.linalg.lu_factor_ex

     

torch.linalg.ldl_factor

     

torch.linalg.ldl_factor_ex

     

torch.linalg.ldl_solve

     

torch.monitor

PyTorch API

Compatibility

Limitations

torch.monitor.Aggregation

Y

  

torch.monitor.Aggregation.name

     

torch.monitor.Stat

Y

  

torch.monitor.Stat.__init__

Y

  

torch.monitor.Stat.add

Y

  

torch.monitor.Stat.count

     

torch.monitor.Stat.get

Y

  

torch.monitor.Stat.name

     

torch.monitor.data_value_t

Y

  

torch.monitor.Event

Y

  

torch.monitor.Event.__init__

Y

  

torch.monitor.Event.data

     

torch.monitor.Event.name

     

torch.monitor.Event.timestamp

     

torch.monitor.EventHandlerHandle

Y

  

torch.monitor.log_event

Y

  

torch.monitor.register_event_handler

Y

  

torch.monitor.unregister_event_handler

Y

  

torch.monitor.TensorboardEventHandler

Y

  

torch.monitor.TensorboardEventHandler.__init__

Y

  

torch.signal

PyTorch API

Compatibility

Limitations

torch.signal.windows.bartlett

     

torch.signal.windows.blackman

     

torch.signal.windows.cosine

     

torch.signal.windows.exponential

     

torch.signal.windows.gaussian

     

torch.signal.windows.general_cosine

     

torch.signal.windows.general_hamming

     

torch.signal.windows.hamming

     

torch.signal.windows.hann

     

torch.signal.windows.kaiser

     

torch.signal.windows.nuttall

     

torch.special

PyTorch API

Compatibility

Limitations

torch.special.airy_ai

     

torch.special.bessel_j0

     

torch.special.bessel_j1

     

torch.special.digamma

     

torch.special.entr

     

torch.special.erf

     

torch.special.erfc

     

torch.special.erfcx

     

torch.special.erfinv

     

torch.special.exp2

     

torch.special.expit

     

torch.special.expm1

     

torch.special.gammainc

     

torch.special.gammaincc

     

torch.special.gammaln

     

torch.special.i0

     

torch.special.i0e

     

torch.special.i1

     

torch.special.i1e

     

torch.special.log1p

     

torch.special.log_ndtr

     

torch.special.log_softmax

     

torch.special.logit

     

torch.special.logsumexp

Y

Only support fp16,fp32,int64,bool

torch.special.multigammaln

     

torch.special.ndtr

     

torch.special.ndtri

     

torch.special.polygamma

     

torch.special.psi

     

torch.special.round

     

torch.special.scaled_modified_bessel_k0

     

torch.special.scaled_modified_bessel_k1

     

torch.special.sinc

     

torch.special.softmax

     

torch.special.spherical_bessel_j0

     

torch.special.xlog1py

     

torch.special.xlogy

Y

Only support fp16,fp32,int64,bool

torch.special.zeta

     

torch.overrides

PyTorch API

Compatibility

Limitations

torch.overrides.get_ignored_functions

Y

  

torch.overrides.get_overridable_functions

Y

  

torch.overrides.resolve_name

Y

  

torch.overrides.get_testing_overrides

Y

  

torch.overrides.handle_torch_function

Y

  

torch.overrides.has_torch_function

Y

  

torch.overrides.is_tensor_like

Y

  

torch.overrides.is_tensor_method_or_property

Y

  

torch.overrides.wrap_torch_function

Y

  

torch.package

PyTorch API

Compatibility

Limitations

torch.package.PackagingError

Y

  

torch.package.EmptyMatchError

Y

  

torch.package.PackageExporter

Y

  

torch.package.PackageExporter.__init__

Y

  

torch.package.PackageExporter.add_dependency

Y

  

torch.package.PackageExporter.all_paths

Y

  

torch.package.PackageExporter.close

Y

  

torch.package.PackageExporter.denied_modules

Y

  

torch.package.PackageExporter.deny

Y

  

torch.package.PackageExporter.dependency_graph_string

Y

  

torch.package.PackageExporter.extern

Y

  

torch.package.PackageExporter.externed_modules

Y

  

torch.package.PackageExporter.get_rdeps

Y

  

torch.package.PackageExporter.get_unique_id

Y

  

torch.package.PackageExporter.intern

Y

  

torch.package.PackageExporter.interned_modules

Y

  

torch.package.PackageExporter.mock

Y

  

torch.package.PackageExporter.mocked_modules

Y

  

torch.package.PackageExporter.register_extern_hook

Y

  

torch.package.PackageExporter.register_intern_hook

Y

  

torch.package.PackageExporter.register_mock_hook

Y

  

torch.package.PackageExporter.save_binary

Y

  

torch.package.PackageExporter.save_module

Y

  

torch.package.PackageExporter.save_pickle

Y

  

torch.package.PackageExporter.save_source_file

Y

  

torch.package.PackageExporter.save_source_string

Y

  

torch.package.PackageExporter.save_text

Y

  

torch.package.PackageImporter

Y

  

torch.package.PackageImporter.__init__

Y

  

torch.package.PackageImporter.file_structure

Y

  

torch.package.PackageImporter.id

Y

  

torch.package.PackageImporter.import_module

Y

  

torch.package.PackageImporter.load_binary

Y

  

torch.package.PackageImporter.load_pickle

Y

  

torch.package.PackageImporter.load_text

Y

  

torch.package.PackageImporter.python_version

Y

  

torch.package.Directory

Y

  

torch.package.Directory.has_file

Y

  

torch.profiler

PyTorch API

Compatibility

Limitations

torch.profiler._KinetoProfile

     

torch.profiler._KinetoProfile.add_metadata

     

torch.profiler._KinetoProfile.add_metadata_json

     

torch.profiler._KinetoProfile.events

     

torch.profiler._KinetoProfile.export_chrome_trace

     

torch.profiler._KinetoProfile.export_memory_timeline

     

torch.profiler._KinetoProfile.export_stacks

     

torch.profiler._KinetoProfile.key_averages

     

torch.profiler.profile

Y

  

torch.profiler.profile.step

Y

  

torch.profiler.ProfilerAction

Y

  

torch.profiler.ProfilerActivity

Y

  

torch.profiler.ProfilerActivity.name

     

torch.profiler.schedule

Y

  

torch.profiler.tensorboard_trace_handler

Y

  

torch.profiler.itt.is_available

     

torch.profiler.itt.mark

     

torch.profiler.itt.range_push

     

torch.profiler.itt.range_pop

     

torch.nn.init

PyTorch API

Compatibility

Limitations

torch.nn.init.calculate_gain

Y

  

torch.nn.init.uniform_

Y

  

torch.nn.init.normal_

Y

  

torch.nn.init.constant_

Y

  

torch.nn.init.ones_

Y

  

torch.nn.init.zeros_

Y

  

torch.nn.init.eye_

Y

  

torch.nn.init.dirac_

Y

  

torch.nn.init.xavier_uniform_

Y

  

torch.nn.init.xavier_normal_

Y

  

torch.nn.init.kaiming_uniform_

Y

  

torch.nn.init.kaiming_normal_

Y

  

torch.nn.init.trunc_normal_

     

torch.nn.init.orthogonal_

Y

  

torch.nn.init.sparse_

Y

  

torch.onnx

PyTorch API

Compatibility

Limitations

torch.onnx.dynamo_export

Y

  

torch.onnx.ExportOptions

Y

  

torch.onnx.enable_fake_mode

     

torch.onnx.ExportOutput

Y

  

torch.onnx.ExportOutput.adapt_torch_inputs_to_onnx

Y

  

torch.onnx.ExportOutput.adapt_torch_outputs_to_onnx

Y

  

torch.onnx.ExportOutput.diagnostic_context

     

torch.onnx.ExportOutput.fake_context

     

torch.onnx.ExportOutput.model_proto

     

torch.onnx.ExportOutput.save

Y

  

torch.onnx.ExportOutput.save_diagnostics

     

torch.onnx.ExportOutputSerializer

Y

  

torch.onnx.ExportOutputSerializer.serialize

Y

  

torch.onnx.OnnxExporterError

     

torch.onnx.OnnxRegistry

     

torch.onnx.OnnxRegistry.get_op_functions

     

torch.onnx.OnnxRegistry.is_registered_op

     

torch.onnx.OnnxRegistry.opset_version

     

torch.onnx.OnnxRegistry.register_op

     

torch.onnx.DiagnosticOptions

     

torch.onnx.is_onnxrt_backend_supported

     

torch.onnx.export

Y

  

torch.onnx.export_to_pretty_string

Y

  

torch.onnx.register_custom_op_symbolic

Y

  

torch.onnx.unregister_custom_op_symbolic

Y

  

torch.onnx.select_model_mode_for_export

Y

  

torch.onnx.is_in_onnx_export

Y

  

torch.onnx.enable_log

Y

  

torch.onnx.disable_log

Y

  

torch.onnx.verification.find_mismatch

Y

  

torch.onnx.JitScalarType

Y

  

torch.onnx.JitScalarType.dtype

Y

  

torch.onnx.JitScalarType.from_dtype

Y

  

torch.onnx.JitScalarType.from_value

Y

  

torch.onnx.JitScalarType.onnx_compatible

Y

  

torch.onnx.JitScalarType.onnx_type

Y

  

torch.onnx.JitScalarType.scalar_name

Y

  

torch.onnx.JitScalarType.torch_name

Y

  

torch.onnx.verification.GraphInfo

Y

  

torch.onnx.verification.GraphInfo.all_mismatch_leaf_graph_info

Y

  

torch.onnx.verification.GraphInfo.clear

Y

  

torch.onnx.verification.GraphInfo.essential_node_count

Y

  

torch.onnx.verification.GraphInfo.essential_node_kinds

Y

  

torch.onnx.verification.GraphInfo.export_repro

Y

  

torch.onnx.verification.GraphInfo.find_mismatch

Y

  

torch.onnx.verification.GraphInfo.find_partition

Y

  

torch.onnx.verification.GraphInfo.has_mismatch

Y

  

torch.onnx.verification.GraphInfo.pretty_print_mismatch

Y

  

torch.onnx.verification.GraphInfo.pretty_print_tree

Y

  

torch.onnx.verification.GraphInfo.verify_export

Y

  

torch.onnx.verification.VerificationOptions

Y

  

torch.optim

PyTorch API

Compatibility

Limitations

torch.optim.Optimizer

     

Optimizer.add_param_group

Y

  

Optimizer.load_state_dict

Y

  

Optimizer.state_dict

Y

  

Optimizer.step

Y

  

Optimizer.zero_grad

Y

  

torch.optim.Adadelta

Y

Only support fp16,fp32,int64,bool

torch.optim.Adadelta.add_param_group

Y

  

torch.optim.Adadelta.load_state_dict

Y

  

torch.optim.Adadelta.register_load_state_dict_post_hook

     

torch.optim.Adadelta.register_load_state_dict_pre_hook

     

torch.optim.Adadelta.register_state_dict_post_hook

     

torch.optim.Adadelta.register_state_dict_pre_hook

     

torch.optim.Adadelta.register_step_post_hook

     

torch.optim.Adadelta.register_step_pre_hook

     

torch.optim.Adadelta.state_dict

Y

  

torch.optim.Adadelta.step

     

torch.optim.Adadelta.zero_grad

Y

  

torch.optim.Adagrad

     

torch.optim.Adagrad.add_param_group

Y

  

torch.optim.Adagrad.load_state_dict

Y

  

torch.optim.Adagrad.register_load_state_dict_post_hook

     

torch.optim.Adagrad.register_load_state_dict_pre_hook

     

torch.optim.Adagrad.register_state_dict_post_hook

     

torch.optim.Adagrad.register_state_dict_pre_hook

     

torch.optim.Adagrad.register_step_post_hook

     

torch.optim.Adagrad.register_step_pre_hook

     

torch.optim.Adagrad.state_dict

Y

  

torch.optim.Adagrad.step

     

torch.optim.Adagrad.zero_grad

Y

  

torch.optim.Adam

     

torch.optim.Adam.add_param_group

Y

  

torch.optim.Adam.load_state_dict

Y

  

torch.optim.Adam.register_load_state_dict_post_hook

     

torch.optim.Adam.register_load_state_dict_pre_hook

     

torch.optim.Adam.register_state_dict_post_hook

     

torch.optim.Adam.register_state_dict_pre_hook

     

torch.optim.Adam.register_step_post_hook

     

torch.optim.Adam.register_step_pre_hook

     

torch.optim.Adam.state_dict

Y

  

torch.optim.Adam.step

     

torch.optim.Adam.zero_grad

Y

  

torch.optim.AdamW

     

torch.optim.AdamW.add_param_group

     

torch.optim.AdamW.load_state_dict

     

torch.optim.AdamW.register_load_state_dict_post_hook

     

torch.optim.AdamW.register_load_state_dict_pre_hook

     

torch.optim.AdamW.register_state_dict_post_hook

     

torch.optim.AdamW.register_state_dict_pre_hook

     

torch.optim.AdamW.register_step_post_hook

     

torch.optim.AdamW.register_step_pre_hook

     

torch.optim.AdamW.state_dict

     

torch.optim.AdamW.step

     

torch.optim.AdamW.zero_grad

     

torch.optim.SparseAdam

     

torch.optim.SparseAdam.add_param_group

     

torch.optim.SparseAdam.load_state_dict

     

torch.optim.SparseAdam.register_load_state_dict_post_hook

     

torch.optim.SparseAdam.register_load_state_dict_pre_hook

     

torch.optim.SparseAdam.register_state_dict_post_hook

     

torch.optim.SparseAdam.register_state_dict_pre_hook

     

torch.optim.SparseAdam.register_step_post_hook

     

torch.optim.SparseAdam.register_step_pre_hook

     

torch.optim.SparseAdam.state_dict

     

torch.optim.SparseAdam.step

     

torch.optim.SparseAdam.zero_grad

     

torch.optim.Adamax

     

torch.optim.Adamax.add_param_group

     

torch.optim.Adamax.load_state_dict

     

torch.optim.Adamax.register_load_state_dict_post_hook

     

torch.optim.Adamax.register_load_state_dict_pre_hook

     

torch.optim.Adamax.register_state_dict_post_hook

     

torch.optim.Adamax.register_state_dict_pre_hook

     

torch.optim.Adamax.register_step_post_hook

     

torch.optim.Adamax.register_step_pre_hook

     

torch.optim.Adamax.state_dict

     

torch.optim.Adamax.step

     

torch.optim.Adamax.zero_grad

     

torch.optim.ASGD

     

torch.optim.ASGD.add_param_group

     

torch.optim.ASGD.load_state_dict

     

torch.optim.ASGD.register_load_state_dict_post_hook

     

torch.optim.ASGD.register_load_state_dict_pre_hook

     

torch.optim.ASGD.register_state_dict_post_hook

     

torch.optim.ASGD.register_state_dict_pre_hook

     

torch.optim.ASGD.register_step_post_hook

     

torch.optim.ASGD.register_step_pre_hook

     

torch.optim.ASGD.state_dict

     

torch.optim.ASGD.step

     

torch.optim.ASGD.zero_grad

     

torch.optim.LBFGS

     

torch.optim.LBFGS.add_param_group

     

torch.optim.LBFGS.load_state_dict

     

torch.optim.LBFGS.register_load_state_dict_post_hook

     

torch.optim.LBFGS.register_load_state_dict_pre_hook

     

torch.optim.LBFGS.register_state_dict_post_hook

     

torch.optim.LBFGS.register_state_dict_pre_hook

     

torch.optim.LBFGS.register_step_post_hook

     

torch.optim.LBFGS.register_step_pre_hook

     

torch.optim.LBFGS.state_dict

     

torch.optim.LBFGS.step

     

torch.optim.LBFGS.zero_grad

     

torch.optim.NAdam

     

torch.optim.NAdam.add_param_group

     

torch.optim.NAdam.load_state_dict

     

torch.optim.NAdam.register_load_state_dict_post_hook

     

torch.optim.NAdam.register_load_state_dict_pre_hook

     

torch.optim.NAdam.register_state_dict_post_hook

     

torch.optim.NAdam.register_state_dict_pre_hook

     

torch.optim.NAdam.register_step_post_hook

     

torch.optim.NAdam.register_step_pre_hook

     

torch.optim.NAdam.state_dict

     

torch.optim.NAdam.step

     

torch.optim.NAdam.zero_grad

     

torch.optim.RAdam

     

torch.optim.RAdam.add_param_group

     

torch.optim.RAdam.load_state_dict

     

torch.optim.RAdam.register_load_state_dict_post_hook

     

torch.optim.RAdam.register_load_state_dict_pre_hook

     

torch.optim.RAdam.register_state_dict_post_hook

     

torch.optim.RAdam.register_state_dict_pre_hook

     

torch.optim.RAdam.register_step_post_hook

     

torch.optim.RAdam.register_step_pre_hook

     

torch.optim.RAdam.state_dict

     

torch.optim.RAdam.step

     

torch.optim.RAdam.zero_grad

     

torch.optim.RMSprop

     

torch.optim.RMSprop.add_param_group

     

torch.optim.RMSprop.load_state_dict

     

torch.optim.RMSprop.register_load_state_dict_post_hook

     

torch.optim.RMSprop.register_load_state_dict_pre_hook

     

torch.optim.RMSprop.register_state_dict_post_hook

     

torch.optim.RMSprop.register_state_dict_pre_hook

     

torch.optim.RMSprop.register_step_post_hook

     

torch.optim.RMSprop.register_step_pre_hook

     

torch.optim.RMSprop.state_dict

     

torch.optim.RMSprop.step

     

torch.optim.RMSprop.zero_grad

     

torch.optim.Rprop

Y

Only support fp16,fp32,int64,bool

torch.optim.Rprop.add_param_group

     

torch.optim.Rprop.load_state_dict

     

torch.optim.Rprop.register_load_state_dict_post_hook

     

torch.optim.Rprop.register_load_state_dict_pre_hook

     

torch.optim.Rprop.register_state_dict_post_hook

     

torch.optim.Rprop.register_state_dict_pre_hook

     

torch.optim.Rprop.register_step_post_hook

     

torch.optim.Rprop.register_step_pre_hook

     

torch.optim.Rprop.state_dict

     

torch.optim.Rprop.step

     

torch.optim.Rprop.zero_grad

     

torch.optim.SGD

     

torch.optim.SGD.add_param_group

     

torch.optim.SGD.load_state_dict

     

torch.optim.SGD.register_load_state_dict_post_hook

     

torch.optim.SGD.register_load_state_dict_pre_hook

     

torch.optim.SGD.register_state_dict_post_hook

     

torch.optim.SGD.register_state_dict_pre_hook

     

torch.optim.SGD.register_step_post_hook

     

torch.optim.SGD.register_step_pre_hook

     

torch.optim.SGD.state_dict

     

torch.optim.SGD.step

     

torch.optim.SGD.zero_grad

     

torch.optim.lr_scheduler.LambdaLR

Y

  

torch.optim.lr_scheduler.LambdaLR.get_last_lr

     

torch.optim.lr_scheduler.LambdaLR.load_state_dict

     

torch.optim.lr_scheduler.LambdaLR.print_lr

     

torch.optim.lr_scheduler.LambdaLR.state_dict

     

torch.optim.lr_scheduler.MultiplicativeLR

     

torch.optim.lr_scheduler.MultiplicativeLR.get_last_lr

     

torch.optim.lr_scheduler.MultiplicativeLR.load_state_dict

     

torch.optim.lr_scheduler.MultiplicativeLR.print_lr

     

torch.optim.lr_scheduler.MultiplicativeLR.state_dict

     

torch.optim.lr_scheduler.StepLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.StepLR.get_last_lr

     

torch.optim.lr_scheduler.StepLR.load_state_dict

     

torch.optim.lr_scheduler.StepLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.StepLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.MultiStepLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.MultiStepLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.MultiStepLR.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.MultiStepLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.MultiStepLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ConstantLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ConstantLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ConstantLR.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ConstantLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ConstantLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.LinearLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.LinearLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.LinearLR.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.LinearLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.LinearLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ExponentialLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ExponentialLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ExponentialLR.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ExponentialLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ExponentialLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.PolynomialLR

     

torch.optim.lr_scheduler.PolynomialLR.get_last_lr

     

torch.optim.lr_scheduler.PolynomialLR.load_state_dict

     

torch.optim.lr_scheduler.PolynomialLR.print_lr

     

torch.optim.lr_scheduler.PolynomialLR.state_dict

     

torch.optim.lr_scheduler.CosineAnnealingLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingLR.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingLR.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.ChainedScheduler

     

torch.optim.lr_scheduler.ChainedScheduler.get_last_lr

     

torch.optim.lr_scheduler.ChainedScheduler.load_state_dict

     

torch.optim.lr_scheduler.ChainedScheduler.print_lr

     

torch.optim.lr_scheduler.ChainedScheduler.state_dict

     

torch.optim.lr_scheduler.SequentialLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.SequentialLR.get_last_lr

     

torch.optim.lr_scheduler.SequentialLR.load_state_dict

     

torch.optim.lr_scheduler.SequentialLR.print_lr

     

torch.optim.lr_scheduler.SequentialLR.state_dict

     

torch.optim.lr_scheduler.ReduceLROnPlateau

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CyclicLR

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CyclicLR.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CyclicLR.get_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CyclicLR.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.OneCycleLR

Y

  

torch.optim.lr_scheduler.OneCycleLR.get_last_lr

     

torch.optim.lr_scheduler.OneCycleLR.load_state_dict

     

torch.optim.lr_scheduler.OneCycleLR.print_lr

     

torch.optim.lr_scheduler.OneCycleLR.state_dict

     

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts.get_last_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts.load_state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts.print_lr

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts.state_dict

Y

Only support fp16,fp32,int64,bool

torch.optim.lr_scheduler.CosineAnnealingWarmRestarts.step

Y

Only support fp16,fp32,int64,bool

torch.cpu

PyTorch API

Compatibility

Limitations

torch.cpu.current_stream

     

torch.cpu.is_available

     

torch.cpu.synchronize

     

torch.cpu.stream

     

torch.cpu.device_count

     

torch.cpu.StreamContext

     

torch._logging

PyTorch API

Compatibility

Limitations

torch._logging.set_logs

     

torch.compiler

PyTorch API

Compatibility

Limitations

torch.compiler.compile

     

torch.compiler.reset

     

torch.compiler.allow_in_graph

     

torch.compiler.assume_constant_result

     

torch.compiler.list_backends

     

torch.compiler.disable

     

torch.export

PyTorch API

Compatibility

Limitations

torch.export.export

     

torch.export.dynamic_dim

     

torch.export.constrain_as_size

     

torch.export.constrain_as_value

     

torch.export.save

     

torch.export.load

     

torch.export.Constraint

     

torch.export.ExportedProgram

     

torch.export.ExportedProgram.module

     

torch.export.ExportBackwardSignature

     

torch.export.ExportGraphSignature

     

torch.export.ArgumentKind

     

torch.export.ArgumentSpec

     

torch.export.ModuleCallSignature

     

torch.export.ModuleCallEntry

     

DDP Communication Hooks

PyTorch API

Compatibility

Limitations

torch.distributed.GradBucket

Y

  

torch.distributed.GradBucket.index

Y

  

torch.distributed.GradBucket.buffer

Y

  

torch.distributed.GradBucket.gradients

Y

  

torch.distributed.GradBucket.is_last

Y

  

torch.distributed.GradBucket.set_buffer

Y

  

torch.distributed.GradBucket.parameters

Y

  

torch.distributed.algorithms.ddp_comm_hooks.default_hooks.allreduce_hook

Y

  

torch.distributed.algorithms.ddp_comm_hooks.default_hooks.fp16_compress_hook

Y

  

torch.distributed.algorithms.ddp_comm_hooks.default_hooks.bf16_compress_hook

     

torch.distributed.algorithms.ddp_comm_hooks.default_hooks.fp16_compress_wrapper

Y

  

torch.distributed.algorithms.ddp_comm_hooks.default_hooks.bf16_compress_wrapper

     

torch.distributed.algorithms.ddp_comm_hooks.powerSGD_hook.PowerSGDState

Y

  

torch.distributed.algorithms.ddp_comm_hooks.powerSGD_hook.powerSGD_hook

Y

  

torch.distributed.algorithms.ddp_comm_hooks.powerSGD_hook.batched_powerSGD_hook

Y

  

torch.distributed.algorithms.ddp_comm_hooks.debugging_hooks.noop_hook

Y

  

__getstate__

     

__setstate__

     

Pipeline Parallelism

PyTorch API

Compatibility

Limitations

torch.distributed.pipeline.sync.Pipe

     

torch.distributed.pipeline.sync.Pipe.forward

     

torch.distributed.pipeline.sync.skip.skippable.skippable

     

torch.distributed.pipeline.sync.skip.skippable.stash

     

torch.distributed.pipeline.sync.skip.skippable.pop

     

torch.distributed.pipeline.sync.skip.skippable.verify_skippables

     

Quantization

PyTorch API

Compatibility

Limitations

torch.ao.quantization.quantize

     

torch.ao.quantization.quantize_dynamic

     

torch.ao.quantization.quantize_qat

     

torch.ao.quantization.prepare

     

torch.ao.quantization.prepare_qat

     

torch.ao.quantization.convert

     

torch.ao.quantization.fuse_modules

     

torch.ao.quantization.QuantStub

     

torch.ao.quantization.DeQuantStub

     

torch.ao.quantization.QuantWrapper

     

torch.ao.quantization.add_quant_dequant

     

torch.ao.quantization.swap_module

     

torch.ao.quantization.propagate_qconfig_

     

torch.ao.quantization.default_eval_fn

     

torch.ao.quantization.quantize_fx.prepare_fx

     

torch.ao.quantization.quantize_fx.prepare_qat_fx

     

torch.ao.quantization.quantize_fx.convert_fx

     

torch.ao.quantization.quantize_fx.fuse_fx

     

torch.ao.quantization.qconfig_mapping.QConfigMapping

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.from_dict

     

torch.ao.quantization.qconfig_mapping.QConfigMapping.set_global

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.set_module_name

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.set_module_name_object_type_order

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.set_module_name_regex

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.set_object_type

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.QConfigMapping.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.get_default_qconfig_mapping

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.qconfig_mapping.get_default_qat_qconfig_mapping

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendConfig.configs

     

torch.ao.quantization.backend_config.BackendConfig.from_dict

     

torch.ao.quantization.backend_config.BackendConfig.set_backend_pattern_config

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendConfig.set_backend_pattern_configs

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendConfig.set_name

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.add_dtype_config

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.from_dict

     

torch.ao.quantization.backend_config.BackendPatternConfig.set_dtype_configs

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_fused_module

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_fuser_method

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_observation_type

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_pattern

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_qat_module

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_reference_quantized_module

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.set_root_module

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.BackendPatternConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.DTypeConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.DTypeConfig.from_dict

     

torch.ao.quantization.backend_config.DTypeConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.DTypeWithConstraints

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.ObservationType

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.backend_config.ObservationType.INPUT_OUTPUT_NOT_OBSERVED

     

torch.ao.quantization.backend_config.ObservationType.OUTPUT_SHARE_OBSERVER_WITH_INPUT

     

torch.ao.quantization.backend_config.ObservationType.OUTPUT_USE_DIFFERENT_OBSERVER_AS_INPUT

     

torch.ao.quantization.fx.custom_config.FuseCustomConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.FuseCustomConfig.from_dict

     

torch.ao.quantization.fx.custom_config.FuseCustomConfig.set_preserved_attributes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.FuseCustomConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.from_dict

     

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_float_to_observed_mapping

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_input_quantized_indexes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_non_traceable_module_classes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_non_traceable_module_names

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_output_quantized_indexes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_preserved_attributes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_standalone_module_class

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.set_standalone_module_name

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.PrepareCustomConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.ConvertCustomConfig

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.ConvertCustomConfig.from_dict

     

torch.ao.quantization.fx.custom_config.ConvertCustomConfig.set_observed_to_quantized_mapping

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.ConvertCustomConfig.set_preserved_attributes

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.ConvertCustomConfig.to_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.fx.custom_config.StandaloneModuleConfigEntry

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.ObserverBase

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.ObserverBase.with_args

     

torch.ao.quantization.observer.ObserverBase.with_callable_args

     

torch.ao.quantization.observer.MinMaxObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.MinMaxObserver.calculate_qparams

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.MinMaxObserver.forward

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.MinMaxObserver.reset_min_max_vals

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.MovingAverageMinMaxObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.PerChannelMinMaxObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.PerChannelMinMaxObserver.reset_min_max_vals

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.MovingAveragePerChannelMinMaxObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.HistogramObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.PlaceholderObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.RecordingObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.NoopObserver

Y

Only support fp16,fp32,int64,bool

torch.ao.quantization.observer.get_observer_state_dict

     

torch.ao.quantization.observer.load_observer_state_dict

     

torch.ao.quantization.observer.default_observer

     

torch.ao.quantization.observer.default_placeholder_observer

     

torch.ao.quantization.observer.default_debug_observer

     

torch.ao.quantization.observer.default_weight_observer

     

torch.ao.quantization.observer.default_histogram_observer

     

torch.ao.quantization.observer.default_per_channel_weight_observer

     

torch.ao.quantization.observer.default_dynamic_quant_observer

     

torch.ao.quantization.observer.default_float_qparams_observer

     

torch.ao.quantization.fake_quantize.FakeQuantizeBase

     

torch.ao.quantization.fake_quantize.FakeQuantize

     

torch.ao.quantization.fake_quantize.FixedQParamsFakeQuantize

     

torch.ao.quantization.fake_quantize.FusedMovingAvgObsFakeQuantize

     

torch.ao.quantization.fake_quantize.default_fake_quant

     

torch.ao.quantization.fake_quantize.default_weight_fake_quant

     

torch.ao.quantization.fake_quantize.default_per_channel_weight_fake_quant

     

torch.ao.quantization.fake_quantize.default_histogram_fake_quant

     

torch.ao.quantization.fake_quantize.default_fused_act_fake_quant

     

torch.ao.quantization.fake_quantize.default_fused_wt_fake_quant

     

torch.ao.quantization.fake_quantize.default_fused_per_channel_wt_fake_quant

     

torch.ao.quantization.fake_quantize.disable_fake_quant

     

torch.ao.quantization.fake_quantize.enable_fake_quant

     

torch.ao.quantization.fake_quantize.disable_observer

     

torch.ao.quantization.fake_quantize.enable_observer

     

torch.ao.quantization.qconfig.QConfig

     

torch.ao.quantization.qconfig.default_qconfig

     

torch.ao.quantization.qconfig.default_debug_qconfig

     

torch.ao.quantization.qconfig.default_per_channel_qconfig

     

torch.ao.quantization.qconfig.default_dynamic_qconfig

     

torch.ao.quantization.qconfig.float16_dynamic_qconfig

     

torch.ao.quantization.qconfig.float16_static_qconfig

     

torch.ao.quantization.qconfig.per_channel_dynamic_qconfig

     

torch.ao.quantization.qconfig.float_qparams_weight_only_qconfig

     

torch.ao.quantization.qconfig.default_qat_qconfig

     

torch.ao.quantization.qconfig.default_weight_only_qconfig

     

torch.ao.quantization.qconfig.default_activation_only_qconfig

     

torch.ao.quantization.qconfig.default_qat_qconfig_v2

     

torch.ao.nn.intrinsic.ConvReLU1d

     

torch.ao.nn.intrinsic.ConvReLU2d

     

torch.ao.nn.intrinsic.ConvReLU3d

     

torch.ao.nn.intrinsic.LinearReLU

     

torch.ao.nn.intrinsic.ConvBn1d

     

torch.ao.nn.intrinsic.ConvBn2d

     

torch.ao.nn.intrinsic.ConvBn3d

     

torch.ao.nn.intrinsic.ConvBnReLU1d

     

torch.ao.nn.intrinsic.ConvBnReLU2d

     

torch.ao.nn.intrinsic.ConvBnReLU3d

     

torch.ao.nn.intrinsic.BNReLU2d

     

torch.ao.nn.intrinsic.BNReLU3d

     

torch.ao.nn.intrinsic.qat.LinearReLU

     

torch.ao.nn.intrinsic.qat.ConvBn1d

     

torch.ao.nn.intrinsic.qat.ConvBnReLU1d

     

torch.ao.nn.intrinsic.qat.ConvBn2d

     

torch.ao.nn.intrinsic.qat.ConvBnReLU2d

     

torch.ao.nn.intrinsic.qat.ConvReLU2d

     

torch.ao.nn.intrinsic.qat.ConvBn3d

     

torch.ao.nn.intrinsic.qat.ConvBnReLU3d

     

torch.ao.nn.intrinsic.qat.ConvReLU3d

     

torch.ao.nn.intrinsic.qat.update_bn_stats

     

torch.ao.nn.intrinsic.qat.freeze_bn_stats

     

torch.ao.nn.intrinsic.quantized.BNReLU2d

     

torch.ao.nn.intrinsic.quantized.BNReLU3d

     

torch.ao.nn.intrinsic.quantized.ConvReLU1d

     

torch.ao.nn.intrinsic.quantized.ConvReLU2d

     

torch.ao.nn.intrinsic.quantized.ConvReLU3d

     

torch.ao.nn.intrinsic.quantized.LinearReLU

     

torch.ao.nn.intrinsic.quantized.dynamic.LinearReLU

     

torch.ao.nn.qat.Conv2d

     

torch.ao.nn.qat.Conv3d

     

torch.ao.nn.qat.Linear

     

torch.ao.nn.qat.Linear.from_float

     

torch.ao.nn.qat.dynamic.Linear

     

torch.ao.nn.quantized.ReLU6

     

torch.ao.nn.quantized.Hardswish

     

torch.ao.nn.quantized.ELU

     

torch.ao.nn.quantized.LeakyReLU

     

torch.ao.nn.quantized.Sigmoid

     

torch.ao.nn.quantized.BatchNorm2d

     

torch.ao.nn.quantized.BatchNorm3d

     

torch.ao.nn.quantized.Conv1d

     

torch.ao.nn.quantized.Conv1d.from_float

     

torch.ao.nn.quantized.Conv2d

     

torch.ao.nn.quantized.Conv2d.from_float

     

torch.ao.nn.quantized.Conv3d

     

torch.ao.nn.quantized.Conv3d.from_float

     

torch.ao.nn.quantized.ConvTranspose1d

     

torch.ao.nn.quantized.ConvTranspose2d

     

torch.ao.nn.quantized.ConvTranspose3d

     

torch.ao.nn.quantized.Embedding

     

torch.ao.nn.quantized.Embedding.from_float

     

torch.ao.nn.quantized.EmbeddingBag

     

torch.ao.nn.quantized.EmbeddingBag.from_float

     

torch.ao.nn.quantized.FloatFunctional

     

torch.ao.nn.quantized.FXFloatFunctional

     

torch.ao.nn.quantized.QFunctional

     

torch.ao.nn.quantized.Linear

     

torch.ao.nn.quantized.Linear.from_float

     

torch.ao.nn.quantized.Linear.from_reference

     

torch.ao.nn.quantized.LayerNorm

     

torch.ao.nn.quantized.GroupNorm

     

torch.ao.nn.quantized.InstanceNorm1d

     

torch.ao.nn.quantized.InstanceNorm2d

     

torch.ao.nn.quantized.InstanceNorm3d

     

torch.ao.nn.quantized.functional.avg_pool2d

     

torch.ao.nn.quantized.functional.avg_pool3d

     

torch.ao.nn.quantized.functional.adaptive_avg_pool2d

     

torch.ao.nn.quantized.functional.adaptive_avg_pool3d

     

torch.ao.nn.quantized.functional.conv1d

     

torch.ao.nn.quantized.functional.conv2d

     

torch.ao.nn.quantized.functional.conv3d

     

torch.ao.nn.quantized.functional.interpolate

     

torch.ao.nn.quantized.functional.linear

     

torch.ao.nn.quantized.functional.max_pool1d

     

torch.ao.nn.quantized.functional.max_pool2d

     

torch.ao.nn.quantized.functional.celu

     

torch.ao.nn.quantized.functional.leaky_relu

     

torch.ao.nn.quantized.functional.hardtanh

     

torch.ao.nn.quantized.functional.hardswish

     

torch.ao.nn.quantized.functional.threshold

     

torch.ao.nn.quantized.functional.elu

     

torch.ao.nn.quantized.functional.hardsigmoid

     

torch.ao.nn.quantized.functional.clamp

     

torch.ao.nn.quantized.functional.upsample

     

torch.ao.nn.quantized.functional.upsample_bilinear

     

torch.ao.nn.quantized.functional.upsample_nearest

     

torch.ao.nn.quantizable.LSTM

     

torch.ao.nn.quantizable.MultiheadAttention

     

torch.ao.nn.quantizable.MultiheadAttention.dequantize

     

torch.ao.nn.quantizable.MultiheadAttention.forward

     

torch.ao.nn.quantized.dynamic.Linear

     

torch.ao.nn.quantized.dynamic.Linear.from_float

     

torch.ao.nn.quantized.dynamic.Linear.from_reference

     

torch.ao.nn.quantized.dynamic.LSTM

Y

Only support fp16,fp32,int64,bool

torch.ao.nn.quantized.dynamic.GRU

Y

Only support fp16,fp32,int64,bool

torch.ao.nn.quantized.dynamic.RNNCell

Y

Only support fp16,fp32,int64,bool

torch.ao.nn.quantized.dynamic.LSTMCell

Y

Only support fp16,fp32,int64,bool

torch.ao.nn.quantized.dynamic.GRUCell

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.compare_weights

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.get_logger_dict

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Logger

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Logger.forward

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.ShadowLogger

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.ShadowLogger.forward

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.OutputLogger

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.OutputLogger.forward

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.forward

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.add

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.add_scalar

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.mul

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.mul_scalar

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.cat

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.Shadow.add_relu

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.prepare_model_with_stubs

     

torch.ao.ns._numeric_suite.compare_model_stub

     

torch.ao.ns._numeric_suite.get_matching_activations

Y

Only support fp16,fp32,int64,bool

torch.ao.ns._numeric_suite.prepare_model_outputs

     

torch.ao.ns._numeric_suite.compare_model_outputs

     

torch.ao.ns._numeric_suite_fx.OutputLogger

     

torch.ao.ns._numeric_suite_fx.OutputLogger.forward

     

torch.ao.ns._numeric_suite_fx.OutputComparisonLogger

     

torch.ao.ns._numeric_suite_fx.OutputComparisonLogger.forward

     

torch.ao.ns._numeric_suite_fx.NSTracer

     

torch.ao.ns._numeric_suite_fx.NSTracer.is_leaf_module

     

torch.ao.ns._numeric_suite_fx.extract_weights

     

torch.ao.ns._numeric_suite_fx.add_loggers

     

torch.ao.ns._numeric_suite_fx.extract_logger_info

     

torch.ao.ns._numeric_suite_fx.add_shadow_loggers

     

torch.ao.ns._numeric_suite_fx.extract_shadow_logger_info

     

torch.ao.ns._numeric_suite_fx.extend_logger_results_with_comparison

     

torch.ao.ns._numeric_suite_fx.prepare_n_shadows_model

     

torch.ao.ns._numeric_suite_fx.loggers_set_enabled

     

torch.ao.ns._numeric_suite_fx.loggers_set_save_activations

     

torch.ao.ns._numeric_suite_fx.convert_n_shadows_model

     

torch.ao.ns._numeric_suite_fx.extract_results_n_shadows_model

     

torch.ao.ns._numeric_suite_fx.print_comparisons_n_shadows_model

     

torch.ao.ns.fx.utils.compute_sqnr

     

torch.ao.ns.fx.utils.compute_normalized_l2_error

     

torch.ao.ns.fx.utils.compute_cosine_similarity

     

Distributed RPC Framework

PyTorch API

Compatibility

Limitations

torch.distributed.rpc.init_rpc

     

torch.distributed.rpc.rpc_sync

     

torch.distributed.rpc.rpc_async

     

torch.distributed.rpc.remote

     

torch.distributed.rpc.get_worker_info

     

torch.distributed.rpc.shutdown

     

torch.distributed.rpc.WorkerInfo

     

torch.distributed.rpc.WorkerInfo.id

     

torch.distributed.rpc.WorkerInfo.name

     

torch.distributed.rpc.functions.async_execution

     

torch.distributed.rpc.BackendType

     

torch.distributed.rpc.RpcBackendOptions

     

torch.distributed.rpc.RpcBackendOptions.init_method

     

torch.distributed.rpc.RpcBackendOptions.rpc_timeout

     

torch.distributed.rpc.TensorPipeRpcBackendOptions

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.device_maps

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.devices

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.init_method

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.num_worker_threads

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.rpc_timeout

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.set_device_map

     

torch.distributed.rpc.TensorPipeRpcBackendOptions.set_devices

     

torch.distributed.rpc.PyRRef

     

torch.distributed.rpc.PyRRef.backward

     

torch.distributed.rpc.PyRRef.confirmed_by_owner

     

torch.distributed.rpc.PyRRef.is_owner

     

torch.distributed.rpc.PyRRef.local_value

     

torch.distributed.rpc.PyRRef.owner

     

torch.distributed.rpc.PyRRef.owner_name

     

torch.distributed.rpc.PyRRef.remote

     

torch.distributed.rpc.PyRRef.rpc_async

     

torch.distributed.rpc.PyRRef.rpc_sync

     

torch.distributed.rpc.PyRRef.to_here

     

torch.distributed.nn.api.remote_module.RemoteModule

     

torch.distributed.nn.api.remote_module.RemoteModule.get_module_rref

     

torch.distributed.nn.api.remote_module.RemoteModule.remote_parameters

     

torch.distributed.autograd.backward

     

torch.distributed.autograd.context

     

torch.distributed.autograd.get_gradients

     

torch.random

PyTorch API

Compatibility

Limitations

torch.random.fork_rng

Y

  

torch.random.get_rng_state

Y

  

torch.random.initial_seed

Y

  

torch.random.manual_seed

Y

  

torch.random.seed

Y

  

torch.random.set_rng_state

Y

  

torch.nested

PyTorch API

Compatibility

Limitations

torch.nested.nested_tensor

     

torch.nested.as_nested_tensor

     

torch.nested.to_padded_tensor

     

torch.sparse

PyTorch API

Compatibility

Limitations

Tensor.is_sparse_csr

     

Tensor.to_sparse_coo

     

Tensor.coalesce

     

Tensor.sparse_resize_

     

Tensor.sparse_resize_and_clear_

     

Tensor.is_coalesced

     

Tensor.crow_indices

     

Tensor.col_indices

     

Tensor.row_indices

     

Tensor.ccol_indices

     

torch.sparse_compressed_tensor

     

torch.sparse.sum

     

torch.sparse.addmm

     

torch.sparse.sampled_addmm

     

torch.sparse.mm

     

torch.sspaddmm

     

torch.hspmm

     

torch.smm

     

torch.sparse.softmax

     

torch.sparse.log_softmax

     

torch.sparse.spdiags

     

torch.sparse.check_sparse_tensor_invariants

     

torch.sparse.check_sparse_tensor_invariants.disable

     

torch.sparse.check_sparse_tensor_invariants.enable

     

torch.sparse.check_sparse_tensor_invariants.is_enabled

     

torch.sparse.as_sparse_gradcheck

     

torch.Storage

PyTorch API

Compatibility

Limitations

torch.TypedStorage

Y

  

torch.TypedStorage.bfloat16

     

torch.TypedStorage.bool

Y

  

torch.TypedStorage.byte

Y

  

torch.TypedStorage.char

Y

  

torch.TypedStorage.clone

Y

  

torch.TypedStorage.complex_double

     

torch.TypedStorage.complex_float

     

torch.TypedStorage.copy_

Y

  

torch.TypedStorage.cpu

Y

  

torch.TypedStorage.cuda

Y

  

torch.TypedStorage.data_ptr

Y

  

torch.TypedStorage.device

     

torch.TypedStorage.double

Y

  

torch.TypedStorage.dtype

     

torch.TypedStorage.element_size

Y

  

torch.TypedStorage.fill_

Y

  

torch.TypedStorage.float

Y

  

torch.TypedStorage.from_buffer

Y

  

torch.TypedStorage.from_file

Y

  

torch.TypedStorage.get_device

Y

  

torch.TypedStorage.half

Y

  

torch.TypedStorage.hpu

     

torch.TypedStorage.int

Y

  

torch.TypedStorage.is_cuda

     

torch.TypedStorage.is_hpu

     

torch.TypedStorage.is_pinned

Y

  

torch.TypedStorage.is_shared

Y

  

torch.TypedStorage.is_sparse

     

torch.TypedStorage.long

Y

  

torch.TypedStorage.nbytes

Y

  

torch.TypedStorage.pickle_storage_type

Y

  

torch.TypedStorage.pin_memory

Y

  

torch.TypedStorage.resize_

Y

  

torch.TypedStorage.share_memory_

     

torch.TypedStorage.short

Y

  

torch.TypedStorage.size

Y

  

torch.TypedStorage.tolist

Y

  

torch.TypedStorage.type

Y

  

torch.TypedStorage.untyped

Y

  

torch.UntypedStorage

Y

  

torch.UntypedStorage.bfloat16

     

torch.UntypedStorage.bool

Y

  

torch.UntypedStorage.byte

Y

  

torch.UntypedStorage.byteswap

     

torch.UntypedStorage.char

Y

  

torch.UntypedStorage.clone

Y

  

torch.UntypedStorage.complex_double

     

torch.UntypedStorage.complex_float

     

torch.UntypedStorage.copy_

Y

  

torch.UntypedStorage.cpu

Y

  

torch.UntypedStorage.cuda

Y

  

torch.UntypedStorage.data_ptr

Y

  

torch.UntypedStorage.device

     

torch.UntypedStorage.double

Y

  

torch.UntypedStorage.element_size

Y

  

torch.UntypedStorage.fill_

Y

  

torch.UntypedStorage.float

Y

  

torch.UntypedStorage.from_buffer

     

torch.UntypedStorage.from_file

     

torch.UntypedStorage.get_device

Y

  

torch.UntypedStorage.half

Y

  

torch.UntypedStorage.hpu

     

torch.UntypedStorage.int

Y

  

torch.UntypedStorage.is_cuda

     

torch.UntypedStorage.is_hpu

     

torch.UntypedStorage.is_pinned

Y

  

torch.UntypedStorage.is_shared

Y

  

torch.UntypedStorage.is_sparse

     

torch.UntypedStorage.is_sparse_csr

     

torch.UntypedStorage.long

Y

  

torch.UntypedStorage.mps

     

torch.UntypedStorage.nbytes

Y

  

torch.UntypedStorage.new

Y

  

torch.UntypedStorage.pin_memory

Y

  

torch.UntypedStorage.resize_

Y

  

torch.UntypedStorage.share_memory_

     

torch.UntypedStorage.short

Y

  

torch.UntypedStorage.size

Y

  

torch.UntypedStorage.tolist

Y

  

torch.UntypedStorage.type

Y

  

torch.UntypedStorage.untyped

Y

  

torch.DoubleStorage

Y

  

torch.DoubleStorage.dtype

     

torch.FloatStorage

Y

  

torch.FloatStorage.dtype

     

torch.HalfStorage

Y

  

torch.HalfStorage.dtype

     

torch.LongStorage

Y

  

torch.LongStorage.dtype

     

torch.IntStorage

Y

  

torch.IntStorage.dtype

     

torch.ShortStorage

Y

  

torch.ShortStorage.dtype

     

torch.CharStorage

Y

  

torch.CharStorage.dtype

     

torch.ByteStorage

Y

  

torch.ByteStorage.dtype

     

torch.BoolStorage

Y

  

torch.BoolStorage.dtype

     

torch.BFloat16Storage

     

torch.BFloat16Storage.dtype

     

torch.ComplexDoubleStorage

     

torch.ComplexDoubleStorage.dtype

     

torch.ComplexFloatStorage

     

torch.ComplexFloatStorage.dtype

     

torch.QUInt8Storage

     

torch.QUInt8Storage.dtype

     

torch.QInt8Storage

     

torch.QInt8Storage.dtype

     

torch.QInt32Storage

     

torch.QInt32Storage.dtype

     

torch.QUInt4x2Storage

     

torch.QUInt4x2Storage.dtype

     

torch.QUInt2x4Storage

     

torch.QUInt2x4Storage.dtype

     

torch.testing

PyTorch API

Compatibility

Limitations

torch.testing.assert_close

Y

  

torch.testing.make_tensor

Y

  

torch.testing.assert_allclose

Y

  

torch.utils

PyTorch API

Compatibility

Limitations

torch.utils.rename_privateuse1_backend

     

torch.utils.generate_methods_for_privateuse1_backend

     

torch.utils.get_cpp_backtrace

     

torch.utils.set_module

     

torch.utils.benchmark

PyTorch API

Compatibility

Limitations

torch.utils.benchmark.Timer

     

torch.utils.benchmark.Timer.blocked_autorange

     

torch.utils.benchmark.Timer.collect_callgrind

Y

  

torch.utils.benchmark.Timer.timeit

     

torch.utils.benchmark.Measurement

Y

  

torch.utils.benchmark.Measurement.merge

     

torch.utils.benchmark.Measurement.significant_figures

     

torch.utils.benchmark.CallgrindStats

Y

  

torch.utils.benchmark.CallgrindStats.as_standardized

Y

  

torch.utils.benchmark.CallgrindStats.counts

Y

  

torch.utils.benchmark.CallgrindStats.delta

Y

  

torch.utils.benchmark.CallgrindStats.stats

Y

  

torch.utils.benchmark.FunctionCounts

Y

  

torch.utils.benchmark.FunctionCounts.denoise

Y

  

torch.utils.benchmark.FunctionCounts.filter

Y

  

torch.utils.benchmark.FunctionCounts.transform

Y

  

torch.utils.checkpoint

PyTorch API

Compatibility

Limitations

torch.utils.checkpoint.checkpoint

Y

  

torch.utils.checkpoint.checkpoint_sequential

Y

  

torch.utils.cpp_extension

PyTorch API

Compatibility

Limitations

torch.utils.cpp_extension.CppExtension

Y

  

torch.utils.cpp_extension.CUDAExtension

Y

  

torch.utils.cpp_extension.BuildExtension

Y

  

torch.utils.cpp_extension.load

     

torch.utils.cpp_extension.load_inline

Y

  

torch.utils.cpp_extension.include_paths

Y

  

torch.utils.cpp_extension.get_compiler_abi_compatibility_and_version

Y

  

torch.utils.cpp_extension.verify_ninja_availability

Y

  

torch.utils.cpp_extension.is_ninja_available

Y

  

torch.utils.data

PyTorch API

Compatibility

Limitations

torch.utils.data.DataLoader

Y

  

torch.utils.data.Dataset

Y

  

torch.utils.data.IterableDataset

Y

  

torch.utils.data.TensorDataset

Y

  

torch.utils.data.StackDataset

     

torch.utils.data.ConcatDataset

Y

  

torch.utils.data.ChainDataset

Y

  

torch.utils.data.Subset

Y

  

torch.utils.data._utils.collate.collate

Y

  

torch.utils.data.default_collate

Y

  

torch.utils.data.default_convert

Y

  

torch.utils.data.get_worker_info

Y

  

torch.utils.data.random_split

Y

  

torch.utils.data.Sampler

Y

  

torch.utils.data.SequentialSampler

Y

  

torch.utils.data.RandomSampler

Y

  

torch.utils.data.SubsetRandomSampler

Y

  

torch.utils.data.WeightedRandomSampler

Y

  

torch.utils.data.BatchSampler

Y

  

torch.utils.data.distributed.DistributedSampler

Y

  

torch.utils.dlpack

PyTorch API

Compatibility

Limitations

torch.utils.dlpack.from_dlpack

     

torch.utils.dlpack.to_dlpack

     

torch.utils.mobile_optimizer

PyTorch API

Compatibility

Limitations

torch.utils.mobile_optimizer.optimize_for_mobile

     

torch.utils.model_zoo

PyTorch API

Compatibility

Limitations

torch.utils.model_zoo.load_url

Y

  

torch.utils.tensorboard

PyTorch API

Compatibility

Limitations

torch.utils.tensorboard.writer.SummaryWriter

Y

  

torch.utils.tensorboard.writer.SummaryWriter.__init__

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_scalar

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_scalars

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_histogram

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_image

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_images

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_figure

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_video

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_audio

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_text

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_graph

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_embedding

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_pr_curve

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_custom_scalars

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_mesh

Y

  

torch.utils.tensorboard.writer.SummaryWriter.add_hparams

Y

  

torch.utils.tensorboard.writer.SummaryWriter.flush

Y

  

torch.utils.tensorboard.writer.SummaryWriter.close

Y

  

Type Info

PyTorch API

Compatibility

Limitations

torch.finfo

Y

  

torch.iinfo

Y

  

Named Tensors

PyTorch API

Compatibility

Limitations

torch.Tensor

Y

  

names

Y

  

rename

Y

  

rename_

Y

  

refine_names

Y

  

align_as

Y

  

align_to

Y

  

flatten

Y

  

torch.__config__

PyTorch API

Compatibility

Limitations

torch.__config__.show

Y

  

torch.__config__.parallel_info

Y

  

Understanding CUDA Memory Usage

PyTorch API

Compatibility

Limitations

torch.cuda.memory._record_memory_history

     

torch.cuda.memory._snapshot

     

torch.cuda.memory._dump_snapshot