torch.nn.functional
Functions
- layer_norm - Apply Layer Normalization.
- rms_norm - RMS Normalization: normalizes by root mean square of activations (LLaMA, modern LLMs).
- normalize - Lp Normalization: scales vectors to unit Lp norm along a dimension.
- group_norm - Group Normalization: divides channels into groups and normalizes each group independently.
- instance_norm - Instance Normalization: normalizes each channel independently for each sample.
- batch_norm - Batch Normalization: normalize activations using batch statistics during training, running statistic
- local_response_norm - Apply local response normalization.
- log_softmax - Log-Softmax activation function: numerically stable log(softmax(x)).
- nll_loss - Negative Log Likelihood (NLL) loss: standard loss for classification with pre-computed log-probabili
- cross_entropy - Cross Entropy Loss: standard loss function for multi-class classification from raw logits.
- mse_loss - Mean Squared Error (MSE) Loss: standard regression loss function.
- l1_loss - L1 Loss (Mean Absolute Error): robust regression loss function.
- dropout - Dropout regularization: randomly zeroes elements during training to prevent overfitting.
- alpha_dropout - Alpha Dropout: dropout for self-normalizing neural networks (SNNs) with SELU.
- feature_alpha_dropout - Randomly masks out entire channels with SELU-compatible dropout.
- dropout1d - Randomly zeros out entire channels in 1D input (sequences).
- dropout2d - Randomly zeros out entire channels in 2D input (spatial feature maps).
- dropout3d - Randomly zeros out entire channels in 3D input (volumetric feature maps).
- max_pool2d - 2D Max Pooling: downsamples feature maps by taking maximum values.
- avg_pool2d - 2D Average Pooling: downsamples feature maps by averaging values.
- max_pool3d - 3D Max Pooling: downsamples volumetric data by taking maximum values.
- avg_pool3d - 3D Average Pooling: downsamples volumetric data by averaging values.
- adaptive_avg_pool2d - 2D Adaptive Average Pooling: averages to fixed spatial size automatically.
- adaptive_max_pool2d - 2D Adaptive Max Pooling: takes max values to fixed spatial size automatically.
- adaptive_avg_pool3d - 3D Adaptive Average Pooling: averages to fixed volumetric size automatically.
- adaptive_max_pool3d - 3D Adaptive Max Pooling: takes max values to fixed volumetric size automatically.
- max_pool1d - 1D Max Pooling: downsamples sequences by taking maximum values.
- avg_pool1d - 1D Average Pooling: downsamples sequences by averaging values.
- adaptive_avg_pool1d - 1D Adaptive Average Pooling: averages to fixed output size automatically.
- adaptive_max_pool1d - 1D Adaptive Max Pooling: takes max values to fixed output size automatically.
- lp_pool1d - Apply 1D LP pooling.
- lp_pool2d - Apply 2D LP pooling.
- lp_pool3d - Apply 3D LP pooling.
- max_pool1d_with_indices - Apply 1D max pooling with indices.
- max_pool2d_with_indices - Applies 2D max pooling over an input signal and returns both the pooled values and their indices.
- max_pool3d_with_indices - Apply 3D max pooling with indices.
- max_unpool1d - Apply 1D max unpooling.
- max_unpool2d - Apply 2D max unpooling.
- max_unpool3d - Apply 3D max unpooling.
- fractional_max_pool2d - Apply 2D fractional max pooling over an input signal.
- fractional_max_pool3d - Apply 3D fractional max pooling over an input signal.
- fractional_max_pool2d_with_indices - Apply 2D fractional max pooling and return both output and indices.
- fractional_max_pool3d_with_indices - Apply 3D fractional max pooling and return both output and indices.
- conv2d - 2D Convolution: applies learned filters to extract spatial features from images.
- conv3d - Applies a 3D convolution over an input signal composed of several input planes.
- conv_transpose1d - 1D transposed convolution: applies a 1D transposed convolution operator over an input signal.
- conv_transpose2d - 2D Transposed Convolution (Deconvolution): upsamples spatial dimensions with learned parameters.
- conv_transpose3d - 3D transposed convolution: applies a 3D transposed convolution operator over an input volume.
- embedding - Generate embeddings by looking up indices in a weight matrix (dense vector representation table).
- embedding_bag - Compute sums, means, or maxes of bags of embeddings.
- one_hot - Convert class indices to one-hot encoded vectors (categorical representation).
- binary_cross_entropy - Binary cross entropy loss.
- binary_cross_entropy_with_logits - Binary cross entropy with logits loss.
- smooth_l1_loss - Smooth L1 Loss: hybrid regression loss combining benefits of L1 and MSE.
- huber_loss - Huber loss: robust regression loss with smooth L2 near zero, linear for outliers.
- kl_div - Kullback-Leibler Divergence Loss: measures dissimilarity between probability distributions.
- cosine_embedding_loss - Cosine Embedding Loss: learns similarity relationships using angular distance.
- triplet_margin_loss - Triplet Margin Loss: learns embeddings with relative distances between triplets.
- margin_ranking_loss - Margin ranking loss for learning relative ordering between pairs of samples.
- hinge_embedding_loss - Hinge embedding loss for learning embeddings with margin constraints.
- soft_margin_loss - Soft margin loss for binary classification with logistic regression.
- unfold - Unfold: extracts sliding local blocks from image-like tensors (im2col operation).
- fold - Combines sliding local blocks into a large tensor (col2im, inverse of unfold).
- conv1d - 1D Convolution: applies learned filters to extract temporal or sequential features.
- pad - Pad a tensor with various padding modes: constant fill, reflection, replication, or circular wrappin
- interpolate - Resample (upsample or downsample) spatial dimensions of a tensor to new sizes using interpolation.
- ctc_loss - CTC (Connectionist Temporal Classification) loss.
- poisson_nll_loss - Poisson negative log likelihood loss for count data and event prediction.
- gaussian_nll_loss - Gaussian (normal distribution) negative log likelihood loss for continuous predictions.
- multilabel_margin_loss - Multi-label margin loss for ranking multiple positive classes with margin separation.
- multilabel_soft_margin_loss - Multi-label soft margin loss using logistic loss for each class independently.
- multi_margin_loss - Multi-class margin loss function for classification with custom margin.
- triplet_margin_with_distance_loss - Triplet margin loss with custom distance function for flexible metric learning.
- pixel_shuffle - Pixel Shuffle: rearranges channels into spatial dimensions for super-resolution upsampling.
- pixel_unshuffle - Pixel Unshuffle: inverse of pixel shuffle, reorganizing channels into spatial dimensions.
- channel_shuffle - Channel Shuffle: rearranges channels by dividing into groups and shuffling order.
- pairwise_distance - Pairwise Distance: computes Lp distance between corresponding vector pairs.
- cosine_similarity - Cosine Similarity: measures angular distance between vectors, invariant to magnitude.
- pdist - Pairwise Distance: computes Lp distances between all pairs of vectors in a batch.
- upsample - Upsample: increases spatial dimensions via interpolation (alias for interpolate).
- upsample_nearest - Upsamples the input tensor using nearest neighbor interpolation.
- upsample_bilinear - Upsamples the input tensor using bilinear interpolation.
- grid_sample - Spatial Transformer Grid Sampling: samples input using learned spatial transformation grids.
- affine_grid - Affine Grid Generation: converts affine transformation matrices to coordinate grids for grid_sample.
- adaptive_max_pool1d_with_indices - Applies 1D adaptive max pooling and returns both output and indices.
- adaptive_max_pool2d_with_indices - Applies 2D adaptive max pooling and returns both output and indices.
- adaptive_max_pool3d_with_indices - Applies 3D adaptive max pooling and returns both output and indices.
- multi_head_attention_forward - Multi-head attention forward pass.
- grouped_mm - Performs grouped (multi-headed) matrix multiplication with optional bias and dtype casting.
- scaled_grouped_mm - Performs batched (grouped) matrix multiplication with per-batch scaling.
- scaled_mm - Performs scaled matrix multiplication with optional quantization and mixed-precision support.
- linear - Applies a linear transformation: y = x @ weight.T + bias
Types
- PoolWithIndicesResult - Result type for pooling operations that return both values and indices.
- ReluFunctionalOptions - Options for relu functional operation.
- RreluFunctionalOptions - Options for rrelu functional operation.
- LeakyReluFunctionalOptions - Options for leaky_relu functional operation.
- EluFunctionalOptions - Options for elu functional operation.
- GluFunctionalOptions - Options for glu functional operation.
- HardtanhFunctionalOptions - Options for hardtanh functional operation.
- CeluFunctionalOptions - Options for celu functional operation.
- HardshrinkFunctionalOptions - Options for hardshrink functional operation.
- SoftshrinkFunctionalOptions - Options for softshrink functional operation.
- SoftplusFunctionalOptions - Options for softplus functional operation.
- SoftminFunctionalOptions - Options for softmin functional operation.
- ScaledDotProductAttentionFunctionalOptions - Options for scaled_dot_product_attention functional operation.
- MultiHeadAttentionFunctionalOptions - Options for multi_head_attention_forward functional operation.
- UnfoldFunctionalOptions - Options for unfold functional operation.
- FoldFunctionalOptions - Options for fold functional operation.
- PadFunctionalOptions - Options for pad functional operation.
- PairwiseDistanceFunctionalOptions - Options for pairwise_distance functional operation.
- CosineSimilarityFunctionalOptions - Options for cosine_similarity functional operation.
- PdistFunctionalOptions - Options for pdist functional operation.
- NormalizeFunctionalOptions - Options for normalize functional operation.
- LocalResponseNormFunctionalOptions - Options for local_response_norm functional operation.
- BatchNormFunctionalOptions - Options for batch_norm functional operation.
- LayerNormFunctionalOptions - Options for layer_norm functional operation.
- RmsNormFunctionalOptions - Options for rms_norm functional operation.
- GroupNormFunctionalOptions - Options for group_norm functional operation.
- InstanceNormFunctionalOptions - Options for instance_norm functional operation.
- NllLossFunctionalOptions - Options for nll_loss functional operation.
- CrossEntropyFunctionalOptions - Options for cross_entropy functional operation.
- MseLossFunctionalOptions - Options for mse_loss functional operation.
- L1LossFunctionalOptions - Options for l1_loss functional operation.
- BinaryCrossEntropyFunctionalOptions - Options for binary_cross_entropy functional operation.
- BinaryCrossEntropyWithLogitsFunctionalOptions - Options for binary_cross_entropy_with_logits functional operation.
- SmoothL1LossFunctionalOptions - Options for smooth_l1_loss functional operation.
- HuberLossFunctionalOptions - Options for huber_loss functional operation.
- KlDivFunctionalOptions - Options for kl_div functional operation.
- CosineEmbeddingLossFunctionalOptions - Options for cosine_embedding_loss functional operation.
- TripletMarginLossFunctionalOptions - Options for triplet_margin_loss functional operation.
- MarginRankingLossFunctionalOptions - Options for margin_ranking_loss functional operation.
- HingeEmbeddingLossFunctionalOptions - Options for hinge_embedding_loss functional operation.
- SoftMarginLossFunctionalOptions - Options for soft_margin_loss functional operation.
- DropoutFunctionalOptions - Options for dropout functional operations.
- AlphaDropoutFunctionalOptions - Options for alpha_dropout and feature_alpha_dropout functional operations.
- Conv1dFunctionalOptions - Options for 1D convolution functional operation.
- Conv2dFunctionalOptions - Options for 2D convolution functional operation.
- Conv3dFunctionalOptions - Options for 3D convolution functional operation.
- ConvTranspose1dFunctionalOptions - Options for 1D transposed convolution functional operation.
- ConvTranspose2dFunctionalOptions - Options for 2D transposed convolution functional operation.
- ConvTranspose3dFunctionalOptions - Options for 3D transposed convolution functional operation.
- MaxPool1dFunctionalOptions - Options for max_pool1d functional operation.
- MaxPool2dFunctionalOptions - Options for max_pool2d functional operation.
- MaxPool3dFunctionalOptions - Options for max_pool3d functional operation.
- AvgPool1dFunctionalOptions - Options for avg_pool1d functional operation.
- AvgPool2dFunctionalOptions - Options for avg_pool2d functional operation.
- AvgPool3dFunctionalOptions - Options for avg_pool3d functional operation.
- AdaptiveMaxPoolFunctionalOptions - Options for adaptive_max_pool*d functional operations.
- LPPoolFunctionalOptions - Options for lp_pool*d functional operations.
- MaxUnpoolFunctionalOptions - Options for max_unpool*d functional operations.
- FractionalMaxPoolFunctionalOptions - Options for fractional_max_pool*d functional operations.
- EmbeddingFunctionalOptions - Options for embedding functional operation.
- EmbeddingBagFunctionalOptions - Options for embedding_bag functional operation.
- InterpolateFunctionalOptions - Options for interpolate functional operation.
- GridSampleFunctionalOptions - Options for grid_sample functional operation.
- AffineGridFunctionalOptions - Options for affine_grid functional operation.
- GroupedMMFunctionalOptions - Options for grouped_mm functional operation.
- ScaledGroupedMMFunctionalOptions - Options for scaled_grouped_mm functional operation.
- ScaledMMFunctionalOptions - Options for scaled_mm functional operation.
- SoftmaxOptions -
- KLDivOptions - Options for kl_div loss.
- CTCLossOptions -
- UpsampleOptions - Options for upsample operations.
- UpsampleNearestOptions - Options for upsample_nearest.
- UpsampleBilinearOptions - Options for upsample_bilinear.