You need to enable JavaScript to run this app.
\u200E
开始使用
特性
文档
API
使用指南
工具平台
工具
AutoDL
PaddleHub
PARL
ERNIE
全部
平台
AI Studio
EasyDL
EasyEdge
资源
模型和数据集
学习资料
应用案例
3.0beta
develop
2.6
2.5
2.4
2.3
2.2
2.1
2.0
1.8
中文(简)
English(En)
安装指南
Pip 安装
Linux 下的 PIP 安装
MacOS 下的 PIP 安装
Windows 下的 PIP 安装
Conda 安装
Linux 下的 Conda 安装
MacOS 下的 Conda 安装
Windows 下的 Conda 安装
Docker 安装
Linux 下的 Docker 安装
MacOS 下的 Docker 安装
Docker 列表
从源码编译
Linux 下使用 make 从源码编译
macOS 下使用 make 从源码编译
Linux 下使用 ninja 从源码编译
macOS 下使用 ninja 从源码编译
Windows 下从源码编译
飞腾/鲲鹏下从源码编译
申威下从源码编译
兆芯下从源码编译
龙芯下从源码编译
昆仑 XPU 芯片
海光 DCU 芯片运行飞桨
附录
使用指南
模型开发入门
10分钟快速上手飞桨
Tensor 介绍
数据集定义与加载
数据预处理
模型组网
模型训练、评估与推理
模型保存与加载
模型开发更多用法
使用 VisualDL 可视化模型,数据和训练
自动微分机制介绍
使用 paddle.nn.Layer 自定义网络
自定义Loss、Metric 及 Callback
梯度裁剪方式介绍
飞桨模型转 ONNX 模型
动态图转静态图
使用样例
转换原理
支持语法
案例解析
报错调试
Limitations
推理部署
服务器部署 — Paddle Inference
移动端/嵌入式部署 — Paddle Lite
模型自动化压缩工具(ACT)
分布式训练
分布式训练简介
环境部署
快速开始-数据并行
快速开始-参数服务器
数据并行
原理和实践案例
前向重计算
自动混合精度
Gradient Merge
参数服务器
参数服务器概述
纯 GPU 参数服务器
CPUPS 流式训练示例
张量模型并行
流水线并行
分组切分并行
MoE
静态图流水并行时序图可视化工具
性能调优
自动混合精度训练(AMP)
模型性能分析(Profiler)
训练全流程自动调优(Beta)
模型迁移
迁移指南
从 PyTorch 迁移到飞桨
CV - 快速上手
CV - 迁移经验汇总
NLP - 快速上手
NLP - 迁移经验汇总
解读网络结构转换
解读 Bert 模型权重转换
Pytorch 2.1.0 与 Paddle develop API 映射表
Pytorch 自定义算子转写教程
使用 X2Paddle 迁移推理模型
迁移飞桨旧版本
升级指南
版本迁移工具
兼容载入旧格式模型
Paddle 1.8 与 Paddle 2.0 API 映射表
附录: 飞桨框架 2.x
硬件支持
飞桨产品硬件支持表
昆仑 XPU 芯片
昆仑 XPU 安装说明
昆仑 XPU 运行示例
昆仑 XPU 支持模型
海光 DCU 芯片运行飞桨
飞桨框架 ROCm 版支持模型
飞桨框架 ROCm 版安装说明
飞桨框架 ROCm 版训练示例
飞桨框架 ROCm 版预测示例
昇腾 NPU 芯片运行飞桨
飞桨框架昇腾 NPU 版安装说明
飞桨框架昇腾 NPU 版训练示例
Graphcore IPU 芯片运行飞桨
飞桨框架 IPU 版安装说明
飞桨框架 IPU 版训练示例
飞桨框架 IPU 版预测示例
寒武纪 MLU 芯片运行飞桨
飞桨框架寒武纪 MLU 版安装说明
飞桨框架 MLU 版训练示例
飞桨框架寒武纪 MLU 版支持模型
自定义算子
自定义 C++算子
自定义 C++ 扩展
自定义 Python 算子
环境变量 FLAGS
cudnn
数值计算
调试
check nan inf 工具
设备管理
分布式
执行器
存储管理
昇腾 NPU
其他
应用实践
快速上手
hello paddle: 从普通程序走向机器学习程序
动态图
飞桨高层API使用指南
模型保存及加载
使用线性回归预测波士顿房价
计算机视觉
使用LeNet在MNIST数据集实现图像分类
使用卷积神经网络进行图像分类
CIFAR-100数据集上基于Vision Transformer 实现图片分类
GAMMA比赛多模态眼底图像数据集下基于EfficientNet和ResNet构造fundus_img和oct_img的分类模型
MosMedData: 新冠肺炎胸部 CT扫描数据集上基于3D-CNN实现二分类
基于图片相似度的图片搜索
基于U-Net卷积神经网络实现宠物图像分割
通过OCR实现验证码识别
通过Sub-Pixel实现图像超分辨率
人脸关键点检测
点云处理:实现PointNet点云分类
点云处理:实现PointNet点云分割
自然语言处理
用N-Gram模型在莎士比亚文集中训练word embedding
IMDB 数据集使用BOW网络的文本分类
使用预训练的词向量完成文本分类任务
使用注意力机制的LSTM的机器翻译
基于Transformer实现英语到西班牙语的翻译任务
使用序列到序列模型完成数字加法
推荐
使用协同过滤实现电影推荐
强化学习
强化学习——Actor Critic Method
强化学习——Advantage Actor-Critic(A2C)
强化学习——Deep Deterministic Policy Gradient (DDPG)
强化学习——DQN玩合成大西瓜
用飞桨框架2.0造一个会下五子棋的AI模型
时序数据
通过AutoEncoder实现时序数据异常检测
Jena Climate时间序列数据集上使用LSTM进行温度的预报
证券数据集下使用LSTM模型预测A股走势
动转静
使用动转静完成以图搜图
生成式对抗网络
图像风格迁移模型-CycleGAN
通过DCGAN实现人脸图像生成
MNIST数据集下用Paddle框架的动态图模式玩耍经典对抗生成网络(GAN)
城市街景分割数据集下使用对抗网络Pix2Pix根据掩码生成街景
API 文档
paddle
Overview
abs
acos
acosh
add
add_
add_n
addmm
all
allclose
amax
amin
angle
any
arange
argmax
argmin
argsort
as_complex
as_real
as_strided
asin
asinh
assign
atan
atan2
atanh
atleast_1d
atleast_2d
atleast_3d
batch
bernoulli
bincount
binomial
bitwise_and
bitwise_not
bitwise_or
bitwise_xor
bmm
broadcast_shape
broadcast_tensors
broadcast_to
bucketize
cast
cdist
ceil
ceil_
chunk
clip
clip_
clone
column_stack
combinations
complex
concat
conj
cos
cosh
count_nonzero
CPUPlace
create_parameter
crop
cross
CUDAPinnedPlace
CUDAPlace
cummax
cummin
cumprod
cumsum
cumulative_trapezoid
DataParallel
deg2rad
diag
diag_embed
diagflat
diagonal
diagonal_scatter
diff
digamma
disable_signal_handler
disable_static
dist
divide
dot
dsplit
dstack
einsum
empty
empty_like
enable_grad
enable_static
equal
equal_all
erf
erfinv
erfinv_
exp
exp_
expand
expand_as
expm1
eye
finfo
flatten
flatten_
flip
floor
floor_
floor_divide
flops
fmax
fmin
frac
frexp
full
full_like
gather
gather_nd
gcd
get_cuda_rng_state
get_default_dtype
get_flags
grad
greater_equal
greater_than
heaviside
histogram
histogramdd
hsplit
hstack
hypot
hypot_
i0
i0e
i1
i1e
iinfo
imag
in_dynamic_mode
increment
index_add
index_add_
index_fill
index_fill_
index_put
index_put_
index_sample
index_select
inner
is_complex
is_empty
is_floating_point
is_grad_enabled
is_integer
is_tensor
isclose
isfinite
isinf
isnan
kron
kthvalue
LazyGuard
lcm
ldexp
lerp
lerp_
less_equal
less_than
lgamma
linspace
load
log
log10
log1p
log2
logaddexp
logcumsumexp
logical_and
logical_not
logical_or
logical_xor
logit
logspace
logsumexp
masked_fill
masked_fill_
masked_scatter
masked_scatter_
masked_select
matmul
max
maximum
mean
median
meshgrid
min
minimum
mm
mod
mode
Model
moveaxis
multigammaln
multigammaln_
multinomial
multiplex
multiply
mv
nan_to_num
nanmean
nanmedian
nansum
neg
nextafter
no_grad
nonzero
normal
not_equal
NPUPlace
numel
ones
ones_like
outer
ParamAttr
pdist
poisson
polar
polygamma
pow
prod
put_along_axis
put_along_axis_
quantile
rad2deg
rand
randint
randint_like
randn
randperm
rank
real
reciprocal
reciprocal_
remainder
remainder_
repeat_interleave
reshape
reshape_
roll
rot90
round
round_
row_stack
rsqrt
rsqrt_
save
scale
scale_
scatter
scatter_
scatter_nd
scatter_nd_add
searchsorted
seed
select_scatter
set_cuda_rng_state
set_default_dtype
set_flags
set_grad_enabled
set_printoptions
sgn
shape
shard_index
sign
signbit
sin
sin_
sinh
slice
sort
split
sqrt
sqrt_
square
squeeze
squeeze_
stack
standard_normal
stanh
std
strided_slice
subtract
subtract_
sum
summary
t
take
take_along_axis
tan
tan_
tanh
tanh_
Tensor
tensor_split
tensordot
tile
to_tensor
tolist
topk
trace
transpose
trapezoid
tril
tril_indices
triu
triu_indices
trunc
unbind
unflatten
unfold
uniform
uniform_
unique
unique_consecutive
unsqueeze
unsqueeze_
unstack
vander
var
view
view_as
vsplit
vstack
where
zeros
zeros_like
paddle.amp
Overview
auto_cast
debugging
check_numerics
collect_operator_stats
compare_accuracy
DebugMode
disable_operator_stats_collection
disable_tensor_checker
enable_operator_stats_collection
enable_tensor_checker
TensorCheckerConfig
decorate
GradScaler
is_bfloat16_supported
is_float16_supported
paddle.audio
Overview
backends
get_current_backend
list_available_backends
set_backend
datasets
ESC50
TESS
features
LogMelSpectrogram
MelSpectrogram
MFCC
Spectrogram
functional
compute_fbank_matrix
create_dct
fft_frequencies
get_window
hz_to_mel
mel_frequencies
mel_to_hz
power_to_db
info
load
save
paddle.autograd
backward
hessian
jacobian
PyLayer
PyLayerContext
saved_tensors_hooks
paddle.callbacks
Overview
Callback
EarlyStopping
LRScheduler
ModelCheckpoint
ProgBarLogger
ReduceLROnPlateau
VisualDL
paddle.device
cuda
current_stream
device_count
empty_cache
Event
get_device_capability
get_device_name
get_device_properties
max_memory_allocated
max_memory_reserved
memory_allocated
memory_reserved
Stream
stream_guard
synchronize
current_stream
Event
get_all_custom_device_type
get_all_device_type
get_available_custom_device
get_available_device
get_cudnn_version
get_device
IPUPlace
is_compiled_with_cinn
is_compiled_with_cuda
is_compiled_with_custom_device
is_compiled_with_ipu
is_compiled_with_mlu
is_compiled_with_npu
is_compiled_with_rocm
is_compiled_with_xpu
MLUPlace
set_device
set_stream
Stream
stream_guard
synchronize
XPUPlace
paddle.distributed
Overview
all_gather
all_gather_object
all_reduce
alltoall
alltoall_single
barrier
broadcast
broadcast_object_list
destroy_process_group
DistAttr
dtensor_from_fn
fleet
DistributedStrategy
Fleet
PaddleCloudRoleMaker
UserDefinedRoleMaker
UtilBase
utils
HDFSClient
LocalFS
recompute
get_backend
get_group
get_rank
get_world_size
gloo_barrier
gloo_init_parallel_env
gloo_release
init_parallel_env
InMemoryDataset
irecv
is_available
is_initialized
isend
launch
new_group
ParallelEnv
Partial
Placement
ProcessMesh
QueueDataset
recv
reduce
reduce_scatter
ReduceOp
ReduceType
Replicate
reshard
rpc
get_all_worker_infos
get_current_worker_info
get_worker_info
init_rpc
rpc_async
rpc_sync
shutdown
scatter
scatter_object_list
send
Shard
shard_layer
shard_optimizer
shard_tensor
sharding
group_sharded_parallel
save_group_sharded_model
spawn
split
Strategy
stream
all_gather
all_reduce
alltoall
alltoall_single
broadcast
recv
reduce
reduce_scatter
scatter
send
to_static
utils
global_gather
global_scatter
wait
paddle.distribution
Overview
AbsTransform
AffineTransform
Bernoulli
Beta
Binomial
Categorical
Cauchy
ChainTransform
ContinuousBernoulli
Dirichlet
Distribution
ExponentialFamily
ExpTransform
Geometric
Gumbel
Independent
IndependentTransform
kl_divergence
Laplace
LogNormal
Multinomial
MultivariateNormal
Normal
Poisson
PowerTransform
register_kl
ReshapeTransform
SigmoidTransform
SoftmaxTransform
StackTransform
StickBreakingTransform
TanhTransform
Transform
TransformedDistribution
Uniform
paddle.fft
Overview
fft
fft2
fftfreq
fftn
fftshift
hfft
hfft2
hfftn
ifft
ifft2
ifftn
ifftshift
ihfft
ihfft2
ihfftn
irfft
irfft2
irfftn
rfft
rfft2
rfftfreq
rfftn
paddle.fluid
Overview
paddle.geometric
Overview
reindex_graph
reindex_heter_graph
sample_neighbors
segment_max
segment_mean
segment_min
segment_sum
send_u_recv
send_ue_recv
send_uv
paddle.hub
Overview
help
list
load
paddle.incubate
asp
add_supported_layer
calculate_density
decorate
prune_model
reset_excluded_layers
set_excluded_layers
autograd
Overview
disable_prim
enable_prim
forward_grad
grad
Hessian
Jacobian
jvp
vjp
autotune
set_config
graph_khop_sampler
graph_reindex
graph_sample_neighbors
graph_send_recv
identity_loss
LookAhead
ModelAverage
nn
functional
fused_ec_moe
fused_feedforward
fused_layer_norm
fused_matmul_bias
fused_multi_head_attention
fused_rms_norm
FusedEcMoe
FusedFeedForward
FusedLinear
FusedMultiHeadAttention
FusedTransformerEncoderLayer
optimizer
functional
minimize_bfgs
minimize_lbfgs
LBFGS
segment_max
segment_mean
segment_min
segment_sum
softmax_mask_fuse
softmax_mask_fuse_upper_triangle
xpu
resnet_block
ResNetBasicBlock
paddle.io
Overview
BatchSampler
ChainDataset
ComposeDataset
ConcatDataset
DataLoader
Dataset
DistributedBatchSampler
get_worker_info
IterableDataset
random_split
RandomSampler
Sampler
SequenceSampler
Subset
SubsetRandomSampler
TensorDataset
WeightedRandomSampler
paddle.jit
Overview
enable_to_static
ignore_module
load
not_to_static
save
set_code_level
set_verbosity
to_static
TranslatedLayer
paddle.linalg
Overview
cholesky
cholesky_solve
cond
corrcoef
cov
det
eig
eigh
eigvals
eigvalsh
householder_product
inv
lstsq
lu
lu_unpack
matrix_exp
matrix_power
matrix_rank
multi_dot
norm
pca_lowrank
pinv
qr
slogdet
solve
svd
triangular_solve
paddle.metric
Overview
accuracy
Accuracy
Auc
Metric
Precision
Recall
paddle.nn
Overview
AdaptiveAvgPool1D
AdaptiveAvgPool2D
AdaptiveAvgPool3D
AdaptiveMaxPool1D
AdaptiveMaxPool2D
AdaptiveMaxPool3D
AlphaDropout
AvgPool1D
AvgPool2D
AvgPool3D
BatchNorm
BatchNorm1D
BatchNorm2D
BatchNorm3D
BCELoss
BCEWithLogitsLoss
BeamSearchDecoder
Bilinear
BiRNN
CELU
ChannelShuffle
ClipGradByGlobalNorm
ClipGradByNorm
ClipGradByValue
Conv1D
Conv1DTranspose
Conv2D
Conv2DTranspose
Conv3D
Conv3DTranspose
CosineEmbeddingLoss
CosineSimilarity
CrossEntropyLoss
CTCLoss
Dropout
Dropout2D
Dropout3D
dynamic_decode
ELU
Embedding
Flatten
Fold
functional
adaptive_avg_pool1d
adaptive_avg_pool2d
adaptive_avg_pool3d
adaptive_max_pool1d
adaptive_max_pool2d
adaptive_max_pool3d
affine_grid
alpha_dropout
avg_pool1d
avg_pool2d
avg_pool3d
batch_norm
bilinear
binary_cross_entropy
binary_cross_entropy_with_logits
celu
channel_shuffle
class_center_sample
conv1d
conv1d_transpose
conv2d
conv2d_transpose
conv3d
conv3d_transpose
cosine_embedding_loss
cosine_similarity
cross_entropy
ctc_loss
dice_loss
dropout
dropout2d
dropout3d
elu
elu_
embedding
fold
gather_tree
gaussian_nll_loss
gelu
glu
grid_sample
gumbel_softmax
hardshrink
hardsigmoid
hardswish
hardtanh
hinge_embedding_loss
hsigmoid_loss
instance_norm
interpolate
kl_div
l1_loss
label_smooth
layer_norm
leaky_relu
linear
local_response_norm
log_loss
log_sigmoid
log_softmax
margin_cross_entropy
margin_ranking_loss
max_pool1d
max_pool2d
max_pool3d
max_unpool1d
max_unpool2d
max_unpool3d
maxout
mish
mse_loss
multi_label_soft_margin_loss
multi_margin_loss
nll_loss
normalize
npair_loss
one_hot
pad
pairwise_distance
pixel_shuffle
pixel_unshuffle
poisson_nll_loss
prelu
relu
relu6
relu_
rnnt_loss
rrelu
selu
sequence_mask
sigmoid
sigmoid_focal_loss
silu
smooth_l1_loss
soft_margin_loss
softmax
softmax_
softmax_with_cross_entropy
softplus
softshrink
softsign
sparse_attention
square_error_cost
swish
tanhshrink
temporal_shift
thresholded_relu
triplet_margin_loss
triplet_margin_with_distance_loss
unfold
upsample
zeropad2d
GaussianNLLLoss
GELU
GLU
GroupNorm
GRU
GRUCell
Hardshrink
Hardsigmoid
Hardswish
Hardtanh
HingeEmbeddingLoss
HSigmoidLoss
Identity
initializer
Assign
Bilinear
calculate_gain
Constant
Dirac
KaimingNormal
KaimingUniform
Normal
Orthogonal
set_global_initializer
TruncatedNormal
Uniform
XavierNormal
XavierUniform
InstanceNorm1D
InstanceNorm2D
InstanceNorm3D
KLDivLoss
L1Loss
Layer
LayerDict
LayerList
LayerNorm
LeakyReLU
Linear
LocalResponseNorm
LogSigmoid
LogSoftmax
LSTM
LSTMCell
MarginRankingLoss
Maxout
MaxPool1D
MaxPool2D
MaxPool3D
MaxUnPool1D
MaxUnPool2D
MaxUnPool3D
Mish
MSELoss
MultiHeadAttention
MultiLabelSoftMarginLoss
MultiMarginLoss
NLLLoss
Pad1D
Pad2D
Pad3D
PairwiseDistance
ParameterList
PixelShuffle
PixelUnshuffle
PoissonNLLLoss
PReLU
ReLU
ReLU6
RNN
RNNCellBase
RNNTLoss
RReLU
SELU
Sequential
Sigmoid
Silu
SimpleRNN
SimpleRNNCell
SmoothL1Loss
SoftMarginLoss
Softmax
Softmax2D
Softplus
Softshrink
Softsign
SpectralNorm
Swish
SyncBatchNorm
Tanh
Tanhshrink
ThresholdedReLU
Transformer
TransformerDecoder
TransformerDecoderLayer
TransformerEncoder
TransformerEncoderLayer
TripletMarginLoss
TripletMarginWithDistanceLoss
Unflatten
Unfold
Upsample
UpsamplingBilinear2D
UpsamplingNearest2D
utils
clip_grad_norm_
clip_grad_value_
parameters_to_vector
remove_weight_norm
spectral_norm
vector_to_parameters
weight_norm
ZeroPad2D
paddle.onnx
export
paddle.optimizer
Overview
Adadelta
Adagrad
Adam
Adamax
AdamW
Lamb
LBFGS
lr
CosineAnnealingDecay
CosineAnnealingWarmRestarts
CyclicLR
ExponentialDecay
InverseTimeDecay
LambdaDecay
LinearLR
LinearWarmup
LRScheduler
MultiplicativeDecay
MultiStepDecay
NaturalExpDecay
NoamDecay
OneCycleLR
PiecewiseDecay
PolynomialDecay
ReduceOnPlateau
StepDecay
Momentum
Optimizer
RMSProp
Rprop
SGD
paddle.profiler
Overview
export_chrome_tracing
export_protobuf
load_profiler_result
make_scheduler
Profiler
ProfilerState
ProfilerTarget
RecordEvent
SortedKeys
SummaryView
paddle.regularizer
L1Decay
L2Decay
paddle.signal
Overview
istft
stft
paddle.sparse
Overview
abs
add
addmm
asin
asinh
atan
atanh
cast
coalesce
deg2rad
divide
expm1
is_same_shape
isnan
log1p
masked_matmul
matmul
multiply
mv
neg
nn
BatchNorm
Conv3D
functional
attention
conv3d
leaky_relu
max_pool3d
relu
relu6
softmax
subm_conv3d
LeakyReLU
MaxPool3D
ReLU
ReLU6
Softmax
SubmConv3D
SyncBatchNorm
pca_lowrank
pow
rad2deg
reshape
sin
sinh
slice
sparse_coo_tensor
sparse_csr_tensor
sqrt
square
subtract
sum
tan
tanh
transpose
paddle.static
Overview
accuracy
append_backward
auc
BuildStrategy
CompiledProgram
cpu_places
create_global_var
cuda_places
data
default_main_program
default_startup_program
deserialize_persistables
deserialize_program
device_guard
Executor
ExponentialMovingAverage
global_scope
gradients
InputSpec
ipu_shard_guard
IpuCompiledProgram
IpuStrategy
load
load_from_file
load_inference_model
load_program_state
mlu_places
name_scope
nn
batch_norm
bilinear_tensor_product
case
cond
conv2d
conv2d_transpose
conv3d
conv3d_transpose
data_norm
deform_conv2d
embedding
fc
group_norm
instance_norm
layer_norm
multi_box_head
nce
prelu
row_conv
sequence_concat
sequence_conv
sequence_enumerate
sequence_expand
sequence_expand_as
sequence_first_step
sequence_last_step
sequence_pad
sequence_pool
sequence_reshape
sequence_reverse
sequence_scatter
sequence_slice
sequence_softmax
sequence_unpad
sparse_embedding
spectral_norm
static_pylayer
switch_case
while_loop
normalize_program
npu_places
Print
Program
program_guard
py_func
save
save_inference_model
save_to_file
scope_guard
serialize_persistables
serialize_program
set_ipu_shard
set_program_state
Variable
WeightNormParamAttr
xpu_places
paddle.sysconfig
get_include
get_lib
paddle.text
Overview
Conll05st
Imdb
Imikolov
Movielens
UCIHousing
viterbi_decode
ViterbiDecoder
WMT14
WMT16
paddle.utils
Overview
cpp_extension
CppExtension
CUDAExtension
get_build_directory
load
setup
deprecated
dlpack
from_dlpack
to_dlpack
download
get_weights_path_from_url
require_version
run_check
try_import
unique_name
generate
guard
switch
paddle.version
Overview
cuda
cudnn
show
xpu
xpu_xccl
paddle.vision
Overview
datasets
Cifar10
Cifar100
DatasetFolder
FashionMNIST
Flowers
ImageFolder
MNIST
VOC2012
get_image_backend
image_load
models
alexnet
AlexNet
DenseNet
densenet121
densenet161
densenet169
densenet201
densenet264
GoogLeNet
googlenet
inception_v3
InceptionV3
LeNet
mobilenet_v1
mobilenet_v2
mobilenet_v3_large
mobilenet_v3_small
MobileNetV1
MobileNetV2
MobileNetV3Large
MobileNetV3Small
ResNet
resnet101
resnet152
resnet18
resnet34
resnet50
resnext101_32x4d
resnext101_64x4d
resnext152_32x4d
resnext152_64x4d
resnext50_32x4d
resnext50_64x4d
shufflenet_v2_swish
shufflenet_v2_x0_25
shufflenet_v2_x0_33
shufflenet_v2_x0_5
shufflenet_v2_x1_0
shufflenet_v2_x1_5
shufflenet_v2_x2_0
ShuffleNetV2
SqueezeNet
squeezenet1_0
squeezenet1_1
VGG
vgg11
vgg13
vgg16
vgg19
wide_resnet101_2
wide_resnet50_2
ops
box_coder
decode_jpeg
deform_conv2d
DeformConv2D
distribute_fpn_proposals
generate_proposals
nms
prior_box
psroi_pool
PSRoIPool
read_file
roi_align
roi_pool
RoIAlign
RoIPool
yolo_box
yolo_loss
set_image_backend
transforms
adjust_brightness
adjust_contrast
adjust_hue
affine
BaseTransform
BrightnessTransform
center_crop
CenterCrop
ColorJitter
Compose
ContrastTransform
crop
erase
Grayscale
hflip
HueTransform
Normalize
normalize
Pad
pad
perspective
RandomAffine
RandomCrop
RandomErasing
RandomHorizontalFlip
RandomPerspective
RandomResizedCrop
RandomRotation
RandomVerticalFlip
Resize
resize
rotate
SaturationTransform
to_grayscale
to_tensor
ToTensor
Transpose
vflip
贡献指南
概述
代码贡献流程
新增 API 开发&提交流程
贡献前阅读
开发 API Python 端
开发 C++ 算子
API 设计和命名规范
API 文档书写规范
API 单测开发及验收规范
算子性能优化 提交流程
算子性能优化 方法介绍
算子性能优化 验收规范
Kernel Primitive API
API 介绍
API 介绍 - IO
API 介绍 - Compute
API 介绍 - OpFunc
API 示例
示例 - ElementwiseAdd
示例 - Reduce
示例 - Model
算子数据类型扩展 提交流程
算子数据类型扩展 验收规范
低精度算子开发贡献指南
低精度算子支持开发规范
低精度算子单测开发规范
曙光开发指南
曙光智算平台-Paddle 源码编译和单测执行
Paddle 适配 C86 加速卡详解
Paddle 框架下 ROCm(HIP)算子单测修复指导
硬件接入飞桨后端指南
硬件接入飞桨后端方案介绍
训练硬件 Custom Device 接入方案介绍
自定义 Runtime
数据类型
Device 接口
Memory 接口
Stream 接口
Event 接口
集合通讯接口
Profiler 接口
自定义 Kernel
Kernel 函数声明
Kernel 实现接口
Context API
Tensor API
Exception API
Kernel 注册接口
新硬件接入示例
文档贡献指南
规范和参考信息
代码规范
报错信息文案书写规范
代码风格检查指南
Paddle CI 测试详解
Python 文档示例代码书写规范
常见问题与解答
2.0 升级常见问题
安装常见问题
数据及其加载常见问题
组网、训练、评估常见问题
模型保存常见问题
参数调整常见问题
分布式训练常见问题
其他常见问题
2.6.0 Release Note
sqrt_
»
sqrt_
在 GitHub 上修改
sqrt_
¶
paddle.
sqrt_
(
x
,
name
=
None
)
¶
Inplace 版本的
sqrt
API,对输入
x
采用 Inplace 策略。
更多关于 inplace 操作的介绍请参考
3.1.3 原位(Inplace)操作和非原位操作的区别
了解详情。