DML_ACTIVATION_RELU_GRAD_OPERATOR_DESC structure (directml.h)
Computes backpropagation gradients for a rectified linear unit (ReLU). This operator performs the following element-wise computation.
X = InputTensor
dY = InputGradientTensor
OutputGradientTensor = (X > 0 ? dY : 0)
The corresponding forward-pass operator is DML_ACTIVATION_RELU_OPERATOR_DESC.
Syntax
struct DML_ACTIVATION_RELU_GRAD_OPERATOR_DESC {
const DML_TENSOR_DESC *InputTensor;
const DML_TENSOR_DESC *InputGradientTensor;
const DML_TENSOR_DESC *OutputGradientTensor;
};
Members
InputTensor
Type: const DML_TENSOR_DESC*
The input (feature) tensor. This is typically the same input as was provided during the forward pass (see DML_ACTIVATION_RELU_OPERATOR_DESC).
InputGradientTensor
Type: const DML_TENSOR_DESC*
The incoming gradient tensor. This is typically obtained from the output of backpropagation of a preceding layer. The Sizes and DataType of this tensor must exactly match those of the InputTensor.
OutputGradientTensor
Requirements
Requirement | Value |
---|---|
Minimum supported client | Windows 10 Build 20348 |
Minimum supported server | Windows 10 Build 20348 |
Header | directml.h |