/* ****************************************************************************** * * * This program and the accompanying materials are made available under the * terms of the Apache License, Version 2.0 which is available at * https://www.apache.org/licenses/LICENSE-2.0. * * See the NOTICE file distributed with this work for additional * information regarding copyright ownership. * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. * * SPDX-License-Identifier: Apache-2.0 ******************************************************************************/ // // @author raver119@gmail.com // #ifndef LIBND4J_HEADERS_LOSS_H #define LIBND4J_HEADERS_LOSS_H #include namespace sd { namespace ops { ////////////////////////////////////////////////////////////////////////// /** * Implementation of hinge loss function max(0, 1 - labels*logits) * * Input arrays: * 0: logits - logits, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, expected to be 0. or 1., type float. * Must have the same shape as logits. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as logits. * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Output array: * 0: loss values, type float. * Can be an array with the same shape as logits or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_hinge_loss) DECLARE_CUSTOM_OP(hinge_loss, 3, 1, false, 0, 1); DECLARE_CUSTOM_OP(hinge_loss_grad, 3, 3, false, 0, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of Huber loss function: * 0.5 * (labels-predictions)^2 if |labels-predictions| <= delta * 0.5 * delta^2 + delta * (|labels-predictions| - delta) if |labels-predictions| > delta * * Input arrays: * 0: predictions - the predicted values, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels, and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as predictions * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Input float arguments: * 0: point where the huber loss function changes from a quadratic to linear. * * Output array: * 0: loss values, type float. * Can be an array with the same shape as predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_huber_loss) DECLARE_CUSTOM_OP(huber_loss, 3, 1, false, 1, 1); DECLARE_CUSTOM_OP(huber_loss_grad, 3, 1, false, 1, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of logarithmic loss function ( y_i * log(p_i) + (1 - y_i) * log(1 - p_i) ) * * Input arrays: * 0: predictions - the predicted values, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels, and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as predictions * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Input float arguments: * 0: a small increment to add to avoid taking a log of zero. * * Output array: * 0: loss values, type float. * Can be an array with the same shape as predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_log_loss) DECLARE_CUSTOM_OP(log_loss, 3, 1, false, 1, 1); DECLARE_CUSTOM_OP(log_loss_grad, 3, 3, false, 1, 1); #endif /** * l2_loss op. * compute a l2 norm for given array. * * input param - an array (tensor) * output value - a real number with given type (e.g. float or double) */ #if NOT_EXCLUDED(OP_l2_loss) DECLARE_CUSTOM_OP(l2_loss, 1, 1, false, 0, 0); #endif /** * This op calculates logarithmic loss of poisson distributed input. * Input arrays: * 0: log_predictions - must be already pre-transformed to log(x) * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, expected to be 0. or 1., type float. * Must have the same shape as logits. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as logits. * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * 1: optional - boolean value compute_full_loss: 0 (default) or 1 (compute) * * Output array: * 0: loss values, type float. * Can be an array with the same shape as log_predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_log_poisson_loss) DECLARE_CUSTOM_OP(log_poisson_loss, 3, 1, true, 0, 1); DECLARE_CUSTOM_OP(log_poisson_loss_grad, 3, 3, true, 0, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of pairwise-errors-squared loss function * * Input arrays: * 0: predictions - the predicted values, type float. * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Output array: * 0: loss value, it is just single scalar, type float. */ #if NOT_EXCLUDED(OP_mean_pairwssqerr_loss) DECLARE_CUSTOM_OP(mean_pairwssqerr_loss, 3, 1, false, 0, 0); DECLARE_CUSTOM_OP(mean_pairwssqerr_loss_grad, 3, 3, false, 0, 0); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of Sum-of-Squares loss function 1/N * sum_{i}^{N}(predictions_i - labels_i)^2 * * Input arrays: * 0: predictions - the predicted values, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as predictions * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Output array: * 0: loss values, type float. * Can be an array with the same shape as predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_mean_sqerr_loss) DECLARE_CUSTOM_OP(mean_sqerr_loss, 3, 1, false, 0, 1); DECLARE_CUSTOM_OP(mean_sqerr_loss_grad, 3, 3, false, 0, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of sigmoid cross-entropy loss function max(logits, 0.) - logits * labels + log(1. + exp(-abs(logits))); * * Input arrays: * 0: logits - logits, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels, and must be broadcastable to labels. * 2: labels - ground truth vales, expected to be 0. or 1., type float. * Must have the same shape as logits. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as logits. * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Input float arguments: * 0: smoothing value, if it is greater than 0 then apply smoothing to the labels (smooth the labels towards 1/2): new_labels = labels * (1 - labelsSmoothing)+ 0.5 * labelsSmoothing * * Output array: * 0: loss values, type float. * Can be an array with the same shape as logits or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_sigm_cross_entropy_loss) DECLARE_CUSTOM_OP(sigm_cross_entropy_loss, 3, 1, false, 1, 1); DECLARE_CUSTOM_OP(sigm_cross_entropy_loss_grad, 3, 3, false, 1, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of softmax cross-entropy loss function max(logits, 0.) - logits * labels + log(1. + exp(-abs(logits))); * * Input arrays: * 0: logits - logits, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels, and must be broadcastable to labels. * 2: labels - ground truth vales, expected to be 0. or 1., type float. * Must have the same shape as logits. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as logits. * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Input float arguments: * 0: smoothing value, if it is greater than 0 then apply smoothing to the labels (smooth the labels towards 1/numClasses): new_labels = labels * (1 - labelsSmoothing) + labelsSmoothing / numClasses * * Output array: * 0: loss values, type float. * Can be an array with shape as in logits except last dimension is equal to unity or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_softmax_cross_entropy_loss) DECLARE_CUSTOM_OP(softmax_cross_entropy_loss, 3, 1, false, 1, 1); DECLARE_CUSTOM_OP(softmax_cross_entropy_loss_grad, 3, 3, false, 1, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of Absolute Difference loss function |predictions - labels| * * Input arrays: * 0: predictions - the predicted values, type float. * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as predictions * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * * Output array: * 0: loss values, type float. * Can be an array with the same shape as predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_absolute_difference_loss) DECLARE_CUSTOM_OP(absolute_difference_loss, 3, 1, false, 0, 1); DECLARE_CUSTOM_OP(absolute_difference_loss_grad, 3, 3, false, 0, 1); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of cosine-distance loss function 1. - (predictions * labels).reduce_sum_along(dimension) * * Input arrays: * 0: predictions - the predicted values, type float * 1: weights - is used for weighting (multiplying) of loss values, type float. * Can be single scalar or has the same rank as labels and must be broadcastable to labels. * 2: labels - ground truth vales, type float. * Must have the same shape as predictions. * * Input integer arguments: * 0: type of reduction to apply to loss * 0 - "none", unreduced weighted losses with the same shape as predictions * 1 - "weighted_sum", output is scalar and equal to sum of all elements of weightedLosses array * 2 - "weighted_mean", output is scalar and equal to sum of all elements of weightedLosses array divided by sum of all elements of weightsBroad array * 3 - "weighted_sum_by_nonzero_weights", output is scalar and equal to scalar sum of all elements of weightedLosses array divided by number of non-zero weights * 1: dimension along which the cosine distance is computed. * * Output array: * 0: loss values, type float. * Can be an array with the same shape as predictions or just single scalar, depending on reduction mode (see input integer argument) */ #if NOT_EXCLUDED(OP_cosine_distance_loss) DECLARE_CUSTOM_OP(cosine_distance_loss, 3, 1, false, 0, 2); DECLARE_CUSTOM_OP(cosine_distance_loss_grad, 3, 3, false, 0, 2); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of softmax cross-entropy loss function * * Input arrays: * 0: logits - logits, type float * 1: labels - ground truth vales, expected to be 0. or 1., type float. * Must have the same shape as logits. * * Input integer arguments: * 0: optional (default is last dimension) dimension with classes * * Output array: * 0: loss values, type float. An array with shape resulting from reducing of logits shape along dimension with classes */ #if NOT_EXCLUDED(OP_softmax_cross_entropy_loss_with_logits) DECLARE_CUSTOM_OP(softmax_cross_entropy_loss_with_logits, 2, 1, false, 0, 0); DECLARE_CUSTOM_OP(softmax_cross_entropy_loss_with_logits_grad, 2, 2, false, 0, 0); #endif ////////////////////////////////////////////////////////////////////////// /** * Implementation of sparse softmax cross-entropy loss function * * Input arrays: * 0: labels - ground truth vales, expected to be within range [0, num_classes), type float. * Must have rank equal logits rank minus 1. * 1: logits - logits, type float * * Output array: * 0: loss values, type float. Has the same shape as labels */ #if NOT_EXCLUDED(OP_sparse_softmax_cross_entropy_loss_with_logits) DECLARE_CUSTOM_OP(sparse_softmax_cross_entropy_loss_with_logits, 2, 1, false, 0, 0); DECLARE_CUSTOM_OP(sparse_softmax_cross_entropy_loss_with_logits_grad, 2, 1, false, 0, 0); #endif /** * Implementation of CTC loss function * * Input arrays: * 0: labels - labels NDArray {BATCH_LEN, MAX_TARGET_LEN}, type integer * 1: logits - logits NDArray {BATCH_LEN, FRAME_LEN, CLASS_LEN }. log softmax of rnn output. It should include a blank label as well, type float * 2: targetLabelLengths - Length of label sequence in labels NDArray {BATCH_LEN}, type integer * 3: logitsLengths - Length of input sequence in logits NDArray {BATCH_LEN}, type integer * * * Input integer arguments: * 0: blank index - index of the blank label in logits * * Output array: * 0: loss values, type float. NDArray {BATCH_LEN} negative log probabilities of loss */ #if NOT_EXCLUDED(OP_ctc_loss) DECLARE_CUSTOM_OP(ctc_loss, 4, 1, false, 0, 1); DECLARE_CUSTOM_OP(ctc_loss_grad, 4, 1, false, 0, 1); #endif } } #endif