* - start working on implementation of sqrtm op Signed-off-by: Yurii <iuriish@yahoo.com> * - improving householder procedure Signed-off-by: Yurii <iuriish@yahoo.com> * - further polishing householder stuff Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing hh pivoting qr procedure Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing BiDiagonalUp procedure Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing householder sequence class Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing jacobi svd class Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing svd stuff 1 Signed-off-by: Yurii <iuriish@yahoo.com> * - polishing svd stuff 2 Signed-off-by: Yurii <iuriish@yahoo.com> * - implementation and testing class which performs Hessenberg decomposition of square matrix Signed-off-by: Yurii <iuriish@yahoo.com> * - add static method to JacobiSVD class which makes the continuous Givens rotation generation algorithm Signed-off-by: Yurii <iuriish@yahoo.com> * - implementation and testing auxiliary methods of Schur decomp class Signed-off-by: Yurii <iuriish@yahoo.com> * some references here and there Signed-off-by: raver119 <raver119@gmail.com> * - trying figure out difference between eigen and our Schur alg Signed-off-by: Yurii <iuriish@yahoo.com> * - testing fixing bugs in Schur decomposition op Signed-off-by: Yurii <iuriish@yahoo.com> * - start to implement class which performs calculation of eigen values and vectors Signed-off-by: Yurii <iuriish@yahoo.com> * - add to EigenValsAndVecs method which calculates complex eigen vectors Signed-off-by: Yurii <iuriish@yahoo.com> * - testing and fixing bugs in EigenValsAndVecs class Signed-off-by: Yurii <iuriish@yahoo.com> * - implementation and testing triangularSolver class Signed-off-by: Yurii <iuriish@yahoo.com> * Added a 2D routine for triangular systems solve. Signed-off-by: shugeo <sgazeos@gmail.com> * Refactored triangularSolve2D routine and tests. Signed-off-by: shugeo <sgazeos@gmail.com> * Refactored another test for triangularSolve2D. Signed-off-by: shugeo <sgazeos@gmail.com> * Refactored test for triangularSolve for vector-bar case. Signed-off-by: shugeo <sgazeos@gmail.com> * Refactored triangularSolve2D routine and tests. Signed-off-by: shugeo <sgazeos@gmail.com> * - implementation of FullPivLU class Signed-off-by: Yurii <iuriish@yahoo.com> * - fix bugs in FullPivLU::solve method Signed-off-by: Yurii <iuriish@yahoo.com> * - correct permutation vector in FullPivLU::solve Signed-off-by: Yurii <iuriish@yahoo.com> * - correct include headers Signed-off-by: Yurii <iuriish@yahoo.com> * - implementation of Sqrtm class Signed-off-by: Yurii <iuriish@yahoo.com> * - testing and fixing bugs in Sqrtm class Signed-off-by: Yurii <iuriish@yahoo.com> * - include sqrtm classes to cuda folder, investigate in what places synchronization doesn't work Signed-off-by: Yurii <iuriish@yahoo.com> * Added implementation for cuda triangularSolve2D and also refactored triangularSolve2D for cpu. Signed-off-by: shugeo <sgazeos@gmail.com> * Eliminated waste implementations. Signed-off-by: shugeo <sgazeos@gmail.com> * - make offset calculation faster in t<> methods Signed-off-by: Yurii <iuriish@yahoo.com> * - rename refference T& NDArray::t<> method Signed-off-by: Yurii <iuriish@yahoo.com> * - further work on cuda sqrtm Signed-off-by: Yurii <iuriish@yahoo.com> * - provide correct synchronization to device in Sqrtm class Signed-off-by: Yurii <iuriish@yahoo.com> * - add tests for sqrtm op Signed-off-by: Yurii <iuriish@yahoo.com> * - correct fails which appeared while testing on jenkins Signed-off-by: Yurii <iuriish@yahoo.com> * - trying to find out mistake in svd::deflation method Signed-off-by: Yurii <iuriish@yahoo.com> * Revert "- trying to find out mistake in svd::deflation method" This reverts commit 19d37baddbc509028e4bc67bc932fe7449becdb6. * Revert "- trying to find out mistake in svd::deflation method" This reverts commit 19d37baddbc509028e4bc67bc932fe7449becdb6. Signed-off-by: Yurii <iuriish@yahoo.com> * - change call semantic of r<> and t<> methods Signed-off-by: Yurii <iuriish@yahoo.com> * - ged rid of ambiguity in * operator overloads for windows buikd Signed-off-by: Yurii <iuriish@yahoo.com> * - get rid of ambiguity in * operator overloads for windows build 2 Signed-off-by: Yurii <iuriish@yahoo.com> * - get rid of ambiguity in * operator overloads for windows build 3 Signed-off-by: Yurii <iuriish@yahoo.com> * - resolve conflicts with master Signed-off-by: Yurii <iuriish@yahoo.com> * cmakelists updated Signed-off-by: raver119@gmail.com <raver119@gmail.com> * - minor fix in merge cpu helper - make use of reference getter Signed-off-by: Yurii <iuriish@yahoo.com> Co-authored-by: raver119 <raver119@gmail.com> Co-authored-by: shugeo <sgazeos@gmail.com>
279 lines
9.1 KiB
C++
279 lines
9.1 KiB
C++
/*******************************************************************************
|
|
* Copyright (c) 2015-2018 Skymind, Inc.
|
|
* Copyright (c) 2019-2020 Konduit K.K.
|
|
*
|
|
* This program and the accompanying materials are made available under the
|
|
* terms of the Apache License, Version 2.0 which is available at
|
|
* https://www.apache.org/licenses/LICENSE-2.0.
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
* License for the specific language governing permissions and limitations
|
|
* under the License.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
******************************************************************************/
|
|
|
|
//
|
|
// @author Yurii Shyrma (iuriish@yahoo.com), created on 20.04.2018
|
|
// @author Oleh Semeniv (oleg.semeniv@gmail.com)
|
|
//
|
|
|
|
#include <ops/declarable/helpers/transforms.h>
|
|
#include <helpers/Loops.h>
|
|
|
|
namespace sd {
|
|
namespace ops {
|
|
namespace helpers {
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename X, typename Z>
|
|
static void mergeMaxIndex_(const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
|
|
const Nd4jLong numArgs = inArrs.size();
|
|
auto x = inArrs[0];
|
|
|
|
auto func = PRAGMA_THREADS_FOR {
|
|
for (auto e = start; e < stop; e++) {
|
|
X max = -DataTypeUtils::max<X>();
|
|
Z idx = static_cast<Z>(0);
|
|
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
X v = inArrs[i]->t<X>(e);
|
|
if (v > max) {
|
|
max = v;
|
|
idx = static_cast<Z>(i);
|
|
}
|
|
}
|
|
|
|
output.r<Z>(e) = static_cast<Z>(idx);
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, x->lengthOf());
|
|
}
|
|
|
|
void mergeMaxIndex(sd::LaunchContext * context, const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
BUILD_DOUBLE_SELECTOR(inArrs[0]->dataType(), output.dataType(), mergeMaxIndex_, (inArrs, output), LIBND4J_TYPES, INDEXING_TYPES);
|
|
}
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeMax_(const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
|
|
const Nd4jLong numArgs = inArrs.size();
|
|
auto x = inArrs[0];
|
|
|
|
auto func = PRAGMA_THREADS_FOR {
|
|
for (auto e = start; e < stop; e++) {
|
|
T max = -DataTypeUtils::max<T>();
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
T v = inArrs[i]->e<T>(e);
|
|
if (v > max)
|
|
max = v;
|
|
}
|
|
output.p(e, max);
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, x->lengthOf());
|
|
}
|
|
|
|
void mergeMax(sd::LaunchContext * context, const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
BUILD_SINGLE_SELECTOR(output.dataType(), mergeMax_, (inArrs, output), LIBND4J_TYPES);
|
|
}
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeMaxBp_(const std::vector<const NDArray*>& inArrs, std::vector<NDArray*>& outArrs) {
|
|
|
|
// outArrs.size() == inArrs.size() - 1
|
|
const Nd4jLong numArgs = outArrs.size();
|
|
// last array is gradient
|
|
const auto gradient = inArrs[numArgs]->bufferAsT<T>();
|
|
auto length = inArrs[numArgs]->lengthOf();
|
|
|
|
bool bSameOrderAndEws1 = (1 == inArrs[numArgs]->ews());
|
|
|
|
if (bSameOrderAndEws1) {
|
|
auto gradOrdering = inArrs[numArgs]->ordering();
|
|
|
|
for (int i = 0; i < numArgs; ++i) {
|
|
bSameOrderAndEws1 &= (gradOrdering == inArrs[i]->ordering());
|
|
bSameOrderAndEws1 &= (1 == inArrs[i]->ews());
|
|
bSameOrderAndEws1 &= (gradOrdering == outArrs[i]->ordering());
|
|
bSameOrderAndEws1 &= (1 == outArrs[i]->ews());
|
|
}
|
|
}
|
|
|
|
|
|
if(bSameOrderAndEws1){
|
|
auto func = PRAGMA_THREADS_FOR{
|
|
for (auto e = start; e < stop; e++) {
|
|
T max = -DataTypeUtils::max<T>();
|
|
Nd4jLong nMaxIndex = 0;
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
const T* v = inArrs[i]->bufferAsT<T>();
|
|
if (v[e] > max) {
|
|
max = v[e];
|
|
nMaxIndex = i;
|
|
}
|
|
}
|
|
T* z = outArrs[nMaxIndex]->bufferAsT<T>();
|
|
z[e] = gradient[e];
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, length);
|
|
return;
|
|
}
|
|
|
|
auto gradShape = inArrs[numArgs]->shapeInfo();
|
|
std::vector<bool> vbSameShaepeAndStrides(numArgs);
|
|
for (int i = 0; i < numArgs; ++i) {
|
|
vbSameShaepeAndStrides[i] = shape::haveSameShapeAndStrides(gradShape, inArrs[i]->shapeInfo());
|
|
}
|
|
|
|
auto func = PRAGMA_THREADS_FOR{
|
|
|
|
int coords[MAX_RANK];
|
|
for (auto e = start; e < stop; e++) {
|
|
|
|
shape::index2coordsCPU(start, e, gradShape, coords);
|
|
|
|
const auto gradOffset = shape::getOffset(gradShape, coords);
|
|
|
|
T max = -DataTypeUtils::max<T>();
|
|
Nd4jLong nMaxIndex = 0;
|
|
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
|
|
const auto xOffset = vbSameShaepeAndStrides[i] ? gradOffset : shape::getOffset(inArrs[i]->shapeInfo(), coords);
|
|
const T* v = inArrs[i]->bufferAsT<T>();
|
|
if (v[xOffset] > max) {
|
|
max = v[xOffset];
|
|
nMaxIndex = i;
|
|
}
|
|
}
|
|
|
|
const auto zOffset = vbSameShaepeAndStrides[nMaxIndex] ? gradOffset : shape::getOffset(outArrs[nMaxIndex]->shapeInfo(), coords);
|
|
|
|
T* z = outArrs[nMaxIndex]->bufferAsT<T>();
|
|
z[zOffset] = gradient[gradOffset];
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, length);
|
|
return;
|
|
}
|
|
|
|
void mergeMaxBp(sd::LaunchContext* context, const std::vector<const NDArray*>& inArrs, std::vector<NDArray*>& outArrs) {
|
|
BUILD_SINGLE_SELECTOR(outArrs[0]->dataType(), mergeMaxBp_, (inArrs, outArrs), LIBND4J_TYPES);
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeAvg_(const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
const Nd4jLong numArgs = inArrs.size();
|
|
const T factor = 1.f / numArgs;
|
|
auto x = inArrs[0];
|
|
|
|
auto func = PRAGMA_THREADS_FOR {
|
|
for (auto e = start; e < stop; e++) {
|
|
T sum = 0.;
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
T v = inArrs[i]->e<T>(e);
|
|
sum += v;
|
|
}
|
|
output.p<T>(e, sum * factor);
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, x->lengthOf());
|
|
}
|
|
|
|
void mergeAvg(sd::LaunchContext * context, const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
BUILD_SINGLE_SELECTOR(output.dataType(), mergeAvg_, (inArrs, output), LIBND4J_TYPES);
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeAvgBp_(const NDArray& gradient, std::vector<NDArray*>& outArrs) {
|
|
|
|
const Nd4jLong numArgs = outArrs.size();
|
|
|
|
auto func = PRAGMA_THREADS_FOR{
|
|
for (auto e = start; e < stop; e++) {
|
|
|
|
T v = gradient.e<T>(e) / numArgs;
|
|
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
outArrs[i]->p<T>(e, v);
|
|
}
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, gradient.lengthOf());
|
|
}
|
|
|
|
void mergeAvgBp(sd::LaunchContext* context, const NDArray& gradient, std::vector<NDArray*>& outArrs) {
|
|
BUILD_SINGLE_SELECTOR(gradient.dataType(), mergeAvgBp_, (gradient, outArrs), LIBND4J_TYPES);
|
|
}
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeAdd_(const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
|
|
const Nd4jLong numArgs = inArrs.size();
|
|
auto x = inArrs[0];
|
|
|
|
auto func = PRAGMA_THREADS_FOR {
|
|
for (auto e = start; e < stop; e++) {
|
|
T sum = (T) 0.f;
|
|
for (Nd4jLong i = 0; i < numArgs; i++)
|
|
sum += inArrs[i]->e<T>(e);
|
|
|
|
output.p(e, sum);
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, x->lengthOf());
|
|
}
|
|
void mergeAdd(sd::LaunchContext * context, const std::vector<const NDArray*>& inArrs, NDArray& output) {
|
|
BUILD_SINGLE_SELECTOR(output.dataType(), mergeAdd_, (inArrs, output), LIBND4J_TYPES);
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
static void mergeAddBp_(const NDArray& gradient, std::vector<NDArray*>& outArrs) {
|
|
|
|
const Nd4jLong numArgs = outArrs.size();
|
|
|
|
auto func = PRAGMA_THREADS_FOR{
|
|
for (auto e = start; e < stop; e++) {
|
|
|
|
T v = gradient.e<T>(e);
|
|
|
|
for (Nd4jLong i = 0; i < numArgs; i++) {
|
|
outArrs[i]->p<T>(e, v);
|
|
}
|
|
}
|
|
};
|
|
|
|
samediff::Threads::parallel_for(func, 0, gradient.lengthOf());
|
|
}
|
|
|
|
void mergeAddBp(sd::LaunchContext* context, const NDArray& gradient, std::vector<NDArray*>& outArrs) {
|
|
BUILD_SINGLE_SELECTOR(gradient.dataType(), mergeAddBp_, (gradient, outArrs), LIBND4J_TYPES);
|
|
}
|
|
|
|
|
|
}
|
|
}
|
|
}
|