Performs a batch matrix-matrix product of matrices stored in batch1
and batch2
, with a reduced add step (all matrix multiplications get accumulated along the first dimension). input
is added to the final result.
batch1
and batch2
must be 3-D tensors each containing the same number of matrices.
If batch1
is a ( b × n × m ) (b \times n \times m) (b×n×m) tensor, batch2
is a ( b × m × p ) (b \times m \times p) (b×m×p) tensor, input
must be broadcastable with a ( n × p ) (n \times p) (n×p) tensor and out
will be a ( n × p ) (n \times p) (n×p) tensor.
o u t = β input + α ( ∑ i = 0 b − 1 batch1 i @ batch2 i ) out = \beta\ \text{input} + \alpha\ (\sum_{i=0}^{b-1} \text{batch1}_i \mathbin{@} \text{batch2}_i) out=β input+α (i=0∑b−1batch1i@batch2i)
If beta
is 0, then the content of input
will be ignored, and nan and inf in it will not be propagated.
For inputs of type FloatTensor or DoubleTensor, arguments beta
and alpha
must be real numbers, otherwise they should be integers.
This operator supports TensorFloat32.
On certain ROCm devices, when using float16 inputs this module will use different precision for backward.
beta (Number, optional) – multiplier for input
( β \beta β)
alpha (Number, optional) – multiplier for batch1 @ batch2 ( α \alpha α)
out (Tensor, optional) – the output tensor.
Example:
>>> M = torch.randn(3, 5) >>> batch1 = torch.randn(10, 3, 4) >>> batch2 = torch.randn(10, 4, 5) >>> torch.addbmm(M, batch1, batch2) tensor([[ 6.6311, 0.0503, 6.9768, -12.0362, -2.1653], [ -4.8185, -1.4255, -6.6760, 8.9453, 2.5743], [ -3.8202, 4.3691, 1.0943, -1.1109, 5.4730]])
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4