2019-06-06 14:21:15 +02:00
|
|
|
/*******************************************************************************
|
|
|
|
* Copyright (c) 2015-2018 Skymind, Inc.
|
|
|
|
*
|
|
|
|
* This program and the accompanying materials are made available under the
|
|
|
|
* terms of the Apache License, Version 2.0 which is available at
|
|
|
|
* https://www.apache.org/licenses/LICENSE-2.0.
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
|
|
* License for the specific language governing permissions and limitations
|
|
|
|
* under the License.
|
|
|
|
*
|
|
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
|
|
******************************************************************************/
|
|
|
|
|
|
|
|
//
|
|
|
|
// Created by raver119 on 04.08.17.
|
|
|
|
//
|
|
|
|
|
|
|
|
#include "testlayers.h"
|
|
|
|
#include <memory>
|
|
|
|
#include <NDArray.h>
|
|
|
|
#include <MmulHelper.h>
|
|
|
|
|
|
|
|
using namespace nd4j;
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
class NDArrayTest : public testing::Test {
|
|
|
|
public:
|
|
|
|
int alpha = 0;
|
|
|
|
|
|
|
|
Nd4jLong *cShape = new Nd4jLong[8]{2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong *fShape = new Nd4jLong[8]{2, 2, 2, 1, 2, 8192, 1, 102};
|
|
|
|
|
|
|
|
float arr1[6] = {1,2,3,4,5,6};
|
|
|
|
Nd4jLong shape1[8] = {2,2,3,3,1,8192,1,99};
|
|
|
|
float arr2[48] = {1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6};
|
|
|
|
Nd4jLong shape2[10] = {3,2,4,6,24,6,1,8192,1,99};
|
|
|
|
const std::vector<Nd4jLong> tileShape1 = {2,2,2};
|
|
|
|
|
|
|
|
|
|
|
|
~NDArrayTest() {
|
|
|
|
delete[] cShape;
|
|
|
|
delete[] fShape;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestDup1) {
|
|
|
|
|
|
|
|
NDArray array(arr1, shape1);
|
|
|
|
|
|
|
|
auto arrC = array.dup('c');
|
|
|
|
auto arrF = array.dup('f');
|
|
|
|
|
|
|
|
ASSERT_TRUE(array.equalsTo(arrF));
|
|
|
|
ASSERT_TRUE(array.equalsTo(arrC));
|
|
|
|
|
|
|
|
ASSERT_TRUE(arrF->equalsTo(arrC));
|
|
|
|
|
|
|
|
delete arrC;
|
|
|
|
delete arrF;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, AssignScalar1) {
|
|
|
|
auto array = NDArrayFactory::create_<float>('c', {1, 10});
|
|
|
|
|
|
|
|
array->assign(2.0f);
|
|
|
|
|
|
|
|
for (int i = 0; i < array->lengthOf(); i++) {
|
|
|
|
ASSERT_EQ(2.0f, array->e<float>(i));
|
|
|
|
}
|
|
|
|
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, NDArrayOrder1) {
|
|
|
|
// original part
|
|
|
|
auto c = new float[4] {1, 2, 3, 4};
|
|
|
|
|
|
|
|
// expected part
|
|
|
|
auto f = new float[4] {1, 3, 2, 4};
|
|
|
|
|
|
|
|
auto arrayC = new NDArray(c, cShape);
|
|
|
|
auto arrayF = arrayC->dup('f');
|
|
|
|
auto arrayC2 = arrayF->dup('c');
|
|
|
|
|
|
|
|
ASSERT_EQ('c', arrayC->ordering());
|
|
|
|
ASSERT_EQ('f', arrayF->ordering());
|
|
|
|
ASSERT_EQ('c', arrayC2->ordering());
|
|
|
|
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
|
|
ASSERT_NEAR(f[i], arrayF->bufferAsT<float>()[i], 1e-5f);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
|
|
ASSERT_EQ(fShape[i], arrayF->getShapeInfo()[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
|
|
ASSERT_NEAR(c[i], arrayC2->bufferAsT<float>()[i], 1e-5f);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < 8; i++) {
|
|
|
|
ASSERT_EQ(cShape[i], arrayC2->getShapeInfo()[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete[] f;
|
|
|
|
delete arrayC;
|
|
|
|
delete arrayF;
|
|
|
|
delete arrayC2;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestGetScalar1) {
|
|
|
|
auto c = new float[4] {1, 2, 3, 4};
|
|
|
|
auto cShape = new Nd4jLong[8]{2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
auto arrayC = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
ASSERT_NEAR(3.0f, arrayC->e<float>(1, 0), 1e-5f);
|
|
|
|
ASSERT_NEAR(4.0f, arrayC->e<float>(1, 1), 1e-5f);
|
|
|
|
|
|
|
|
auto arrayF = arrayC->dup('f');
|
|
|
|
|
|
|
|
ASSERT_NEAR(3.0f, arrayF->e<float>(1, 0), 1e-5f);
|
|
|
|
ASSERT_NEAR(4.0f, arrayF->e<float>(1, 1), 1e-5f);
|
|
|
|
|
|
|
|
|
|
|
|
arrayF->p(1, 0, 7.0f);
|
|
|
|
ASSERT_NEAR(7.0f, arrayF->e<float>(1, 0), 1e-5f);
|
|
|
|
|
|
|
|
|
|
|
|
arrayC->p(1, 1, 9.0f);
|
|
|
|
ASSERT_NEAR(9.0f, arrayC->e<float>(1, 1), 1e-5f);
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete[] cShape;
|
|
|
|
|
|
|
|
delete arrayC;
|
|
|
|
delete arrayF;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, EqualityTest1) {
|
|
|
|
auto arrayA = NDArrayFactory::create_<float>('f', {3, 5});
|
|
|
|
auto arrayB = NDArrayFactory::create_<float>('f', {3, 5});
|
|
|
|
auto arrayC = NDArrayFactory::create_<float>('f', {3, 5});
|
|
|
|
|
|
|
|
auto arrayD = NDArrayFactory::create_<float>('f', {2, 4});
|
|
|
|
auto arrayE = NDArrayFactory::create_<float>('f', {1, 15});
|
|
|
|
|
|
|
|
for (int i = 0; i < arrayA->rows(); i++) {
|
|
|
|
for (int k = 0; k < arrayA->columns(); k++) {
|
|
|
|
arrayA->p(i, k, (float) i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < arrayB->rows(); i++) {
|
|
|
|
for (int k = 0; k < arrayB->columns(); k++) {
|
|
|
|
arrayB->p(i, k, (float) i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int i = 0; i < arrayC->rows(); i++) {
|
|
|
|
for (int k = 0; k < arrayC->columns(); k++) {
|
|
|
|
arrayC->p(i, k, (float) i+1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
//nd4j_printf("A B\n","");
|
|
|
|
ASSERT_TRUE(arrayA->equalsTo(arrayB, 1e-5));
|
|
|
|
|
|
|
|
//nd4j_printf("C B\n","");
|
|
|
|
ASSERT_FALSE(arrayC->equalsTo(arrayB, 1e-5));
|
|
|
|
|
|
|
|
//nd4j_printf("D B\n","");
|
|
|
|
ASSERT_FALSE(arrayD->equalsTo(arrayB, 1e-5));
|
|
|
|
|
|
|
|
//nd4j_printf("E B\n","");
|
|
|
|
ASSERT_FALSE(arrayE->equalsTo(arrayB, 1e-5));
|
|
|
|
|
|
|
|
delete arrayA;
|
|
|
|
delete arrayB;
|
|
|
|
delete arrayC;
|
|
|
|
delete arrayD;
|
|
|
|
delete arrayE;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestTad1) {
|
|
|
|
auto array = NDArrayFactory::create_<float>('c', {3, 3});
|
|
|
|
|
|
|
|
auto row2 = array->tensorAlongDimension(1, {1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(row2->isView());
|
|
|
|
ASSERT_EQ(3, row2->lengthOf());
|
|
|
|
|
|
|
|
row2->assign(1.0);
|
|
|
|
|
|
|
|
ASSERT_NEAR(3.0f, array->sumNumber().e<float>(0), 1e-5);
|
|
|
|
|
|
|
|
delete row2;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTad2) {
|
|
|
|
auto array = NDArrayFactory::create_<float>('c', {3, 3});
|
|
|
|
|
|
|
|
ASSERT_EQ(3, array->tensorsAlongDimension({1}));
|
|
|
|
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTad3) {
|
|
|
|
auto array = NDArrayFactory::create_<float>('c', {4, 3});
|
|
|
|
|
|
|
|
auto row2 = array->tensorAlongDimension(1, {1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(row2->isView());
|
|
|
|
ASSERT_EQ(3, row2->lengthOf());
|
|
|
|
|
|
|
|
row2->p(1, 1.0);
|
|
|
|
|
|
|
|
//array->printBuffer();
|
|
|
|
|
|
|
|
row2->p(2, 1.0);
|
|
|
|
|
|
|
|
//array->printBuffer();
|
|
|
|
|
|
|
|
delete row2;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshape1) {
|
|
|
|
|
|
|
|
NDArray array('c', {2, 2, 5, 5}, nd4j::DataType::FLOAT32);
|
|
|
|
int pShape[] = {4, 2, 5, 5, 2, 25, 5, 1, 50, 8192, 0, 99};
|
|
|
|
int rShape[] = {3, 2, 25, 2, 25, 1, 50, 8192, 0, 99};
|
|
|
|
|
|
|
|
array.permutei({1, 2, 3, 0});
|
|
|
|
|
|
|
|
for (int e = 0; e < shape::shapeInfoLength(array.getShapeInfo()); e++)
|
|
|
|
ASSERT_EQ(pShape[e], array.getShapeInfo()[e]);
|
|
|
|
|
|
|
|
array.reshapei('c', {2, 25, 2});
|
|
|
|
|
|
|
|
for (int e = 0; e < shape::shapeInfoLength(array.getShapeInfo()); e++)
|
|
|
|
ASSERT_EQ(rShape[e], array.getShapeInfo()[e]);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshape2) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {2, 2, 5, 5, 6, 6});
|
|
|
|
int pShape[] = {6, 2, 2, 6, 6, 5, 5, 900, 1800, 6, 1, 180, 36, 8192, 0, 99};
|
|
|
|
int rShape[] = {3, 2, 72, 25, 1800, 25, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
|
|
|
|
// array.printShapeInfo("before");
|
|
|
|
|
|
|
|
array.permutei({1, 0, 4, 5, 2, 3});
|
|
|
|
|
|
|
|
// array.printShapeInfo("after ");
|
|
|
|
|
|
|
|
auto aShape = array.getShapeInfo();
|
|
|
|
|
|
|
|
for (int e = 0; e < shape::shapeInfoLength(array.getShapeInfo()); e++)
|
|
|
|
ASSERT_EQ(pShape[e], aShape[e]);
|
|
|
|
|
|
|
|
array.reshapei('c', {2, 72, 25});
|
|
|
|
|
|
|
|
for (int e = 0; e < shape::shapeInfoLength(array.getShapeInfo()); e++)
|
|
|
|
ASSERT_EQ(rShape[e], array.getShapeInfo()[e]);
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestRepeat1) {
|
|
|
|
|
|
|
|
auto eBuffer = new float[8] {1.0,2.0,1.0,2.0,3.0,4.0,3.0,4.0};
|
|
|
|
auto eShape = new Nd4jLong[8]{2, 4, 2, 2, 1, 8192, 1, 99};
|
|
|
|
NDArray array('c', {2, 2}, nd4j::DataType::FLOAT32);
|
|
|
|
auto exp = new NDArray(eBuffer, eShape);
|
|
|
|
for (int e = 0; e < array.lengthOf(); e++)
|
|
|
|
array.p(e, e + 1);
|
|
|
|
|
|
|
|
// array.printBuffer();
|
|
|
|
|
|
|
|
auto rep = array.repeat(0, {2});
|
|
|
|
|
|
|
|
ASSERT_EQ(4, rep->sizeAt(0));
|
|
|
|
ASSERT_EQ(2, rep->sizeAt(1));
|
|
|
|
|
|
|
|
// rep->printIndexedBuffer("Repeated");
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp->equalsTo(rep));
|
|
|
|
|
|
|
|
delete[] eBuffer;
|
|
|
|
delete[] eShape;
|
|
|
|
delete exp;
|
|
|
|
delete rep;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestRepeat2) {
|
|
|
|
auto eBuffer = new float[8] {1.0,2.0,1.0,2.0,3.0,4.0,3.0,4.0};
|
|
|
|
auto eShape = new Nd4jLong[8]{2, 4, 2, 2, 1, 8192, 1, 99};
|
|
|
|
auto array = NDArrayFactory::create_<float>('c', {2, 2});
|
|
|
|
auto exp = new NDArray(eBuffer, eShape);
|
|
|
|
for (int e = 0; e < array->lengthOf(); e++)
|
|
|
|
array->p(e, e + 1);
|
|
|
|
|
|
|
|
//array->printBuffer();
|
|
|
|
|
|
|
|
auto rep = exp->dup();
|
|
|
|
rep->assign(0.);
|
2019-08-21 20:10:29 +02:00
|
|
|
array->repeat(0, {2}, *rep);
|
2019-06-06 14:21:15 +02:00
|
|
|
//rep->printIndexedBuffer("Repeated");
|
|
|
|
|
|
|
|
ASSERT_EQ(4, rep->sizeAt(0));
|
|
|
|
ASSERT_EQ(2, rep->sizeAt(1));
|
|
|
|
|
|
|
|
//rep->printBuffer();
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp->equalsTo(rep));
|
|
|
|
|
|
|
|
delete[] eBuffer;
|
|
|
|
delete[] eShape;
|
|
|
|
delete array;
|
|
|
|
delete exp;
|
|
|
|
delete rep;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestIndexedPut1) {
|
|
|
|
auto array = NDArrayFactory::create_<float>('f', {3, 3});
|
|
|
|
|
|
|
|
array->p(4, 1.0f);
|
|
|
|
ASSERT_EQ(1.0f, array->e<float>(4));
|
|
|
|
//array->printBuffer();
|
|
|
|
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestSum1) {
|
|
|
|
// Nd4jLong *cShape = new Nd4jLong[8]{2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
float *c = new float[4] {1, 2, 3, 4};
|
|
|
|
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
ASSERT_EQ(10.0f, array->sumNumber().e<float>(0));
|
|
|
|
ASSERT_EQ(2.5f, array->meanNumber().e<float>(0));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestAddiRowVector) {
|
|
|
|
float *c = new float[4] {1, 2, 3, 4};
|
|
|
|
float *e = new float[4] {2, 3, 4, 5};
|
|
|
|
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
auto row = NDArrayFactory::create_<float>('c', {1, 2});
|
|
|
|
auto exp = new NDArray(e, cShape);
|
|
|
|
row->assign(1.0f);
|
|
|
|
|
|
|
|
array->addiRowVector(row);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp->equalsTo(array));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete[] e;
|
|
|
|
|
|
|
|
delete array;
|
|
|
|
delete row;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestAddiColumnVector) {
|
|
|
|
float arr1[] = {1, 2, 3, 4};
|
|
|
|
float arr2[] = {5, 6};
|
|
|
|
float arr3[] = {6, 7, 9, 10};
|
|
|
|
Nd4jLong shape1[] = {2,2,2,2,1,8192,1,99};
|
|
|
|
Nd4jLong shape2[] = {2,2,1,1,1,8192,1,99};
|
|
|
|
NDArray matrix(arr1, shape1);
|
|
|
|
NDArray column(arr2, shape2);
|
|
|
|
NDArray exp(arr3, shape1);
|
|
|
|
|
|
|
|
matrix.addiColumnVector(&column);
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(&matrix));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&matrix));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMuliColumnVector) {
|
|
|
|
float arr1[] = {1, 2, 3, 4};
|
|
|
|
float arr2[] = {5, 6};
|
|
|
|
float arr3[] = {5, 10, 18, 24};
|
|
|
|
Nd4jLong shape1[] = {2,2,2,2,1,8192,1,99};
|
|
|
|
Nd4jLong shape2[] = {2,2,1,1,1,8192,1,99};
|
|
|
|
NDArray matrix(arr1, shape1);
|
|
|
|
NDArray column(arr2, shape2);
|
|
|
|
NDArray exp(arr3, shape1);
|
|
|
|
|
|
|
|
matrix.muliColumnVector(&column);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(&matrix));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&matrix));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test3D_1) {
|
|
|
|
auto arrayC = NDArrayFactory::create_<double>('c', {2, 5, 10});
|
|
|
|
auto arrayF = NDArrayFactory::create_<double>('f', {2, 5, 10});
|
|
|
|
|
|
|
|
ASSERT_EQ(100, arrayC->lengthOf());
|
|
|
|
ASSERT_EQ(100, arrayF->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_EQ('c', arrayC->ordering());
|
|
|
|
ASSERT_EQ('f', arrayF->ordering());
|
|
|
|
|
|
|
|
delete arrayC;
|
|
|
|
delete arrayF;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTranspose1) {
|
|
|
|
auto arrayC = NDArrayFactory::create_<double>('c', {2, 5, 10});
|
|
|
|
|
|
|
|
auto expC = new Nd4jLong[10] {3, 2, 5, 10, 50, 10, 1, 16384, 1, 99};
|
|
|
|
auto expT = new Nd4jLong[10] {3, 10, 5, 2, 1, 10, 50, 16384, 1, 102};
|
|
|
|
|
|
|
|
auto arrayT = arrayC->transpose();
|
|
|
|
|
|
|
|
for (int e = 0; e < arrayC->rankOf(); e++) {
|
|
|
|
ASSERT_EQ(shape::shapeOf(expC)[e], arrayC->sizeAt(e));
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
ASSERT_EQ(shape::shapeOf(expT)[e], arrayT.sizeAt(e));
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
delete arrayC;
|
|
|
|
delete[] expC;
|
|
|
|
delete[] expT;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTranspose2) {
|
|
|
|
auto arrayC = NDArrayFactory::create_<double>('c', {2, 5, 10});
|
|
|
|
|
|
|
|
auto expC = new Nd4jLong[10] {3, 2, 5, 10, 50, 10, 1, 16384, 1, 99};
|
|
|
|
auto expT = new Nd4jLong[10] {3, 10, 5, 2, 1, 10, 50, 16384, 1, 102};
|
|
|
|
|
|
|
|
arrayC->transposei();
|
|
|
|
|
|
|
|
|
|
|
|
for (int e = 0; e < arrayC->rankOf(); e++) {
|
|
|
|
ASSERT_EQ(shape::shapeOf(expT)[e], arrayC->sizeAt(e));
|
|
|
|
}
|
|
|
|
|
|
|
|
delete arrayC;
|
|
|
|
delete[] expC;
|
|
|
|
delete[] expT;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestSumAlongDimension1) {
|
|
|
|
|
|
|
|
NDArray array('c', {2,2}, {1,2,3,4}, nd4j::DataType::FLOAT32);
|
|
|
|
|
|
|
|
auto res = array.reduceAlongDims(reduce::Sum, {0});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, res.lengthOf());
|
|
|
|
|
|
|
|
ASSERT_EQ(4.0f, res.e<float>(0));
|
|
|
|
ASSERT_EQ(6.0f, res.e<float>(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestSumAlongDimension2) {
|
|
|
|
float *c = new float[4] {1, 2, 3, 4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
auto res = array->reduceAlongDimension(reduce::Sum, {1});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, res->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_EQ(3.0f, res->e<float>(0));
|
|
|
|
ASSERT_EQ(7.0f, res->e<float>(1));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
delete res;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestReduceAlongDimension1) {
|
|
|
|
float *c = new float[4] {1, 2, 3, 4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
auto exp = array->reduceAlongDimension(reduce::Sum, {1});
|
|
|
|
auto res = array->reduceAlongDimension(reduce::Sum, {1});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, res->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_EQ(3.0f, res->e<float>(0));
|
|
|
|
ASSERT_EQ(7.0f, res->e<float>(1));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
delete exp;
|
|
|
|
delete res;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTransform1) {
|
|
|
|
float *c = new float[4] {-1, -2, -3, -4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
float *e = new float[4] {1, 2, 3, 4};
|
|
|
|
auto exp = new NDArray(e, cShape);
|
|
|
|
|
|
|
|
array->applyTransform(transform::Abs, nullptr, nullptr);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp->equalsTo(array));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
delete[] e;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestReduceScalar1) {
|
|
|
|
float *c = new float[4] {-1, -2, -3, -4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
ASSERT_EQ(-4, array->reduceNumber(reduce::Min, nullptr).e<float>(0));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestReduceScalar2) {
|
|
|
|
float *c = new float[4] {-1, -2, -3, -4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
ASSERT_EQ(-10, array->reduceNumber(reduce::Sum, nullptr).e<float>(0));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestReduceScalar3) {
|
|
|
|
auto array = new NDArray(arr1, shape1);
|
|
|
|
|
|
|
|
ASSERT_EQ(21, array->reduceNumber(reduce::Sum, nullptr).e<float>(0));
|
|
|
|
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestApplyTransform1) {
|
|
|
|
float *c = new float[4] {-1, -2, -3, -4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
float *e = new float[4] {1, 2, 3, 4};
|
|
|
|
auto exp = new NDArray(e, cShape);
|
|
|
|
|
|
|
|
array->applyTransform(transform::Abs, nullptr, nullptr);
|
|
|
|
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp->equalsTo(array));
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
|
|
|
|
delete[] e;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestVectors1) {
|
|
|
|
float *c = new float[4]{-1, -2, -3, -4};
|
|
|
|
auto array = new NDArray(c, cShape);
|
|
|
|
|
|
|
|
|
|
|
|
auto vecShape = array->getShapeInfoAsVector();
|
|
|
|
auto vecBuffer = array->getBufferAsVector<float>();
|
|
|
|
|
|
|
|
ASSERT_EQ(8, vecShape.size());
|
|
|
|
ASSERT_EQ(4, vecBuffer.size());
|
|
|
|
|
|
|
|
for (int e = 0; e < vecBuffer.size(); e++) {
|
|
|
|
ASSERT_NEAR(c[e], vecBuffer[e], 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (int e = 0; e < vecShape.size(); e++) {
|
|
|
|
ASSERT_EQ(cShape[e], vecShape[e]);
|
|
|
|
}
|
|
|
|
|
|
|
|
delete[] c;
|
|
|
|
delete array;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestChecks1) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {1, 5});
|
|
|
|
|
|
|
|
ASSERT_FALSE(array.isMatrix());
|
|
|
|
ASSERT_FALSE(array.isScalar());
|
|
|
|
ASSERT_TRUE(array.isVector());
|
|
|
|
ASSERT_FALSE(array.isColumnVector());
|
|
|
|
ASSERT_TRUE(array.isRowVector());
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestChecks2) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {5, 5});
|
|
|
|
|
|
|
|
ASSERT_TRUE(array.isMatrix());
|
|
|
|
ASSERT_FALSE(array.isScalar());
|
|
|
|
ASSERT_FALSE(array.isVector());
|
|
|
|
ASSERT_FALSE(array.isColumnVector());
|
|
|
|
ASSERT_FALSE(array.isRowVector());
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestChecks3) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {5, 1});
|
|
|
|
|
|
|
|
ASSERT_FALSE(array.isMatrix());
|
|
|
|
ASSERT_FALSE(array.isScalar());
|
|
|
|
ASSERT_TRUE(array.isVector());
|
|
|
|
ASSERT_TRUE(array.isColumnVector());
|
|
|
|
ASSERT_FALSE(array.isRowVector());
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestChecks4) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {1, 1});
|
|
|
|
|
|
|
|
ASSERT_FALSE(array.isMatrix());
|
|
|
|
ASSERT_FALSE(array.isVector());
|
|
|
|
ASSERT_FALSE(array.isColumnVector());
|
|
|
|
ASSERT_FALSE(array.isRowVector());
|
|
|
|
ASSERT_TRUE(array.isScalar());
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestReductionAny1) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
array.p(0, 1.0f);
|
|
|
|
array.p(1, 1.0f);
|
|
|
|
array.p(2, 0.0f);
|
|
|
|
array.p(3, 0.0f);
|
|
|
|
array.syncToDevice();
|
|
|
|
auto result0 = array.reduceAlongDimension(reduce::Any, {0});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, result0->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_NEAR(1.0f, result0->e<float>(0), 1e-5f);
|
|
|
|
ASSERT_NEAR(1.0f, result0->e<float>(1), 1e-5f);
|
|
|
|
|
|
|
|
auto result1 = array.reduceAlongDimension(reduce::Any, {1});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, result1->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_NEAR(1.0f, result1->e<float>(0), 1e-5f);
|
|
|
|
ASSERT_NEAR(0.0f, result1->e<float>(1), 1e-5f);
|
|
|
|
|
|
|
|
delete result0;
|
|
|
|
delete result1;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestReductionAll1) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
array.p(0, 1.0f);
|
|
|
|
array.p(1, 1.0f);
|
|
|
|
array.p(2, 0.0f);
|
|
|
|
array.p(3, 0.0f);
|
|
|
|
|
|
|
|
auto result0 = array.reduceAlongDimension(reduce::All, {0});
|
|
|
|
auto result1 = array.reduceAlongDimension(reduce::All, {1});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, result0->lengthOf());
|
|
|
|
ASSERT_EQ(2, result1->lengthOf());
|
|
|
|
|
|
|
|
ASSERT_FALSE(result0->e<bool>(0));
|
|
|
|
ASSERT_FALSE(result0->e<bool>(1));
|
|
|
|
|
|
|
|
ASSERT_TRUE(result1->e<bool>(0));
|
|
|
|
ASSERT_FALSE(result1->e<bool>(1));
|
|
|
|
|
|
|
|
delete result0;
|
|
|
|
delete result1;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestChecks5) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {5, 5, 5});
|
|
|
|
|
|
|
|
ASSERT_FALSE(array.isMatrix());
|
|
|
|
ASSERT_FALSE(array.isVector());
|
|
|
|
ASSERT_FALSE(array.isColumnVector());
|
|
|
|
ASSERT_FALSE(array.isRowVector());
|
|
|
|
ASSERT_FALSE(array.isScalar());
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTile1) {
|
|
|
|
|
|
|
|
// float arr1[6] = {1,2,3,4,5,6};
|
|
|
|
// Nd4jLong shape1[8] = {2,2,3,3,1,8192,1,99};
|
|
|
|
// float arr2[48] = {1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6,1,2,3,1,2,3,4,5,6,4,5,6};
|
|
|
|
// Nd4jLong shape2[10] = {3,2,4,6,24,6,1,8192,1,99};
|
|
|
|
|
|
|
|
NDArray array1(arr1,shape1); // {2,3}
|
|
|
|
NDArray array2(arr2,shape2); // {2,4,6}
|
|
|
|
auto expA = array1.dup('c');
|
|
|
|
|
|
|
|
auto tiled = array1.tile(tileShape1);
|
|
|
|
|
|
|
|
// array2.printShapeInfo("Expct shape");
|
|
|
|
// tiled.printShapeInfo("Tiled shape");
|
|
|
|
// tiled.printBuffer();
|
|
|
|
|
|
|
|
ASSERT_TRUE(tiled.isSameShape(&array2));
|
|
|
|
ASSERT_TRUE(tiled.equalsTo(&array2));
|
|
|
|
|
|
|
|
ASSERT_TRUE(expA->isSameShape(&array1));
|
|
|
|
ASSERT_TRUE(expA->equalsTo(&array1));
|
|
|
|
|
|
|
|
// delete tiled;
|
|
|
|
delete expA;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestTile2) {
|
|
|
|
|
|
|
|
NDArray array1(arr1,shape1);
|
|
|
|
NDArray array2(arr2,shape2);
|
|
|
|
|
|
|
|
auto tiled = array1.tile(tileShape1);
|
|
|
|
|
|
|
|
ASSERT_TRUE(tiled.isSameShape(&array2));
|
|
|
|
ASSERT_TRUE(tiled.equalsTo(&array2));
|
|
|
|
// delete tiled;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTile3) {
|
|
|
|
|
|
|
|
NDArray array1(arr1,shape1);
|
|
|
|
NDArray array2(arr2,shape2);
|
|
|
|
|
|
|
|
array1.tilei(tileShape1);
|
|
|
|
|
|
|
|
ASSERT_TRUE(array1.isSameShapeStrict(&array2));
|
|
|
|
ASSERT_TRUE(array1.equalsTo(&array2));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTile4) {
|
|
|
|
|
|
|
|
float xBuff[] = {1,2,3,4,5,6};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {1.f,2.f, 1.f,2.f, 3.f,4.f, 3.f,4.f, 5.f,6.f, 5.f,6.f};
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>(xBuff, 'c', {3,1,2});
|
|
|
|
auto exp = NDArrayFactory::create<float>(expBuff, 'c', {3,2,2});
|
|
|
|
|
|
|
|
auto result = x.tile({2,1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(result.isSameShapeStrict(&exp));
|
|
|
|
ASSERT_TRUE(result.equalsTo(&exp));
|
|
|
|
}
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTile5) {
|
|
|
|
|
|
|
|
float xBuff[] = {1,2,3,4,5,6,7,8,9,10,11,12};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {1.f, 2.f, 3.f, 4.f, 1.f, 2.f, 3.f, 4.f, 5.f, 6.f, 7.f, 8.f, 5.f, 6.f, 7.f, 8.f, 9.f,10.f, 11.f,12.f, 9.f,10.f, 11.f,12.f};
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>(xBuff, 'c', {3,2,2});
|
|
|
|
auto exp = NDArrayFactory::create<float>(expBuff, 'c', {3,4,2});
|
|
|
|
|
|
|
|
auto result = x.tile({2,1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(result.isSameShapeStrict(&exp));
|
|
|
|
ASSERT_TRUE(result.equalsTo(&exp));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTile6)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,11., 10.,11., 10.,11., 10.,11., 12.,13., 12.,13., 12.,13., 12.,13., 14.,15., 14.,15., 14.,15., 14.,15.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
|
|
|
|
auto result = x.tile({1,4,1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper1) {
|
|
|
|
auto xBuffer = new float[3]{1.f, 2.f, 3.f};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 1, 3, 1, 1, 8192, 1, 99};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[3]{2.f, 4.f, 6.f};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 1, 3, 1, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = MmulHelper::mmul(x, y);
|
|
|
|
|
|
|
|
ASSERT_EQ(1, z->lengthOf());
|
|
|
|
ASSERT_NEAR(28, z->e<float>(0), 1e-5);
|
|
|
|
|
|
|
|
delete z;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] yShape;
|
|
|
|
delete y;
|
|
|
|
delete x;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshapeMmul1) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {6, 3});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {3, 6});
|
|
|
|
|
|
|
|
Nd4jLong _expS[] = {2, 3, 3, 1, 3, 8192, 1, 102};
|
2019-12-05 20:05:33 +01:00
|
|
|
float _expB[] = {231.0f, 252.0f, 273.0f, 537.0f, 594.0f, 651.0f, 843.0f, 936.0f, 1029.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
|
|
|
|
for (int e = 0; e < x.lengthOf(); e++)
|
|
|
|
x.p(e, e+1);
|
|
|
|
|
|
|
|
for (int e = 0; e < y.lengthOf(); e++)
|
|
|
|
y.p(e, e+1);
|
|
|
|
|
|
|
|
x.permutei({1, 0});
|
|
|
|
y.permutei({1, 0});
|
|
|
|
|
|
|
|
auto z = MmulHelper::mmul(&x, &y);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(z));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(z));
|
|
|
|
|
|
|
|
delete z;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshapeMmul2) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {6, 3});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {3, 6});
|
|
|
|
|
|
|
|
Nd4jLong _expS[] = {2, 3, 3, 1, 3, 8192, 1, 102};
|
2019-12-05 20:05:33 +01:00
|
|
|
float _expB[] = {231.0f, 252.0f, 273.0f, 537.0f, 594.0f, 651.0f, 843.0f, 936.0f, 1029.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
|
|
|
|
for (int e = 0; e < x.lengthOf(); e++)
|
|
|
|
x.p(e, e+1);
|
|
|
|
|
|
|
|
for (int e = 0; e < y.lengthOf(); e++)
|
|
|
|
y.p(e, e+1);
|
|
|
|
|
|
|
|
auto x_ = x.dup('f');
|
|
|
|
auto y_ = y.dup('f');
|
|
|
|
|
|
|
|
x_->permutei({1, 0});
|
|
|
|
y_->permutei({1, 0});
|
|
|
|
|
|
|
|
auto z = MmulHelper::mmul(x_, y_);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(z));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(z));
|
|
|
|
|
|
|
|
delete z;
|
|
|
|
delete x_;
|
|
|
|
delete y_;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshapeMmul3) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2, 2, 3, 2, 2});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {2, 3, 2 ,2});
|
|
|
|
|
|
|
|
Nd4jLong _expS[] = {2, 8, 2, 1, 8, 8192, 1, 102};
|
2019-12-05 20:05:33 +01:00
|
|
|
float _expB[] = {1624.0f, 1858.0f, 2092.0f, 2326.0f, 5368.0f, 5602.0f, 5836.0f, 6070.0f, 4504.0f, 5170.0f, 5836.0f, 6502.0f, 15160.0f, 15826.0f, 16492.0f, 17158.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
|
|
|
|
for (int e = 0; e < x.lengthOf(); e++)
|
|
|
|
x.p(e, e+1);
|
|
|
|
|
|
|
|
for (int e = 0; e < y.lengthOf(); e++)
|
|
|
|
y.p(e, e+1);
|
|
|
|
|
|
|
|
x.permutei({0, 3, 4, 5, 1, 2});
|
|
|
|
y.permutei({3, 2, 1, 0});
|
|
|
|
|
|
|
|
x.reshapei('c', {2 * 2 * 2, 3 * 2 * 2});
|
|
|
|
y.reshapei('c', {2 * 2 * 3, 2});
|
|
|
|
|
|
|
|
auto z = MmulHelper::mmul(&x, &y);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(z));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(z));
|
|
|
|
|
|
|
|
delete z;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestPermuteReshapeMmul4) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2, 2, 3, 2, 2});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {2, 3, 2 ,2});
|
|
|
|
|
|
|
|
Nd4jLong _expS[] = {2, 8, 2, 1, 8, 8192, 1, 102};
|
2019-12-05 20:05:33 +01:00
|
|
|
float _expB[] = {1624.0f, 1858.0f, 2092.0f, 2326.0f, 5368.0f, 5602.0f, 5836.0f, 6070.0f, 4504.0f, 5170.0f, 5836.0f, 6502.0f, 15160.0f, 15826.0f, 16492.0f, 17158.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
|
|
|
|
for (int e = 0; e < x.lengthOf(); e++)
|
|
|
|
x.p(e, e+1);
|
|
|
|
|
|
|
|
for (int e = 0; e < y.lengthOf(); e++)
|
|
|
|
y.p(e, e+1);
|
|
|
|
|
|
|
|
auto y_ = y.dup('f');
|
|
|
|
|
|
|
|
x.permutei({0, 3, 4, 5, 1, 2});
|
|
|
|
y_->permutei({3, 2, 1, 0});
|
|
|
|
|
|
|
|
x.reshapei('c', {2 * 2 * 2, 3 * 2 * 2});
|
|
|
|
y_->reshapei('c', {2 * 2 * 3, 2});
|
|
|
|
|
|
|
|
auto z = MmulHelper::mmul(&x, y_);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(z));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(z));
|
|
|
|
|
|
|
|
delete z;
|
|
|
|
delete y_;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper2) {
|
|
|
|
auto xBuffer = new float[15]{1.f, 2.f, 3.f, 4.f, 5.f, 6.f, 7.f, 8.f, 9.f, 10.f, 11.f, 12.f, 13.f, 14.f, 15.f};
|
|
|
|
Nd4jLong xShape[8] = {2, 5, 3, 3, 1, 8192, 1, 99};
|
|
|
|
auto x = new NDArray(xBuffer, xShape, nd4j::LaunchContext ::defaultContext(), true);
|
|
|
|
|
|
|
|
|
|
|
|
auto yBuffer = new float[3]{2.f, 4.f, 6.f};
|
|
|
|
Nd4jLong yShape[8] = {2, 3, 1, 1, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape, nd4j::LaunchContext ::defaultContext(), true);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {5, 1});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[5]{28.00f, 64.00f, 100.00f, 136.00f, 172.00f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo(), nd4j::LaunchContext ::defaultContext(), true);
|
|
|
|
|
|
|
|
//nd4j::blas::GEMV<float>::op('f', x->rows(), x->columns(), 1.0f, x->getBuffer(), y->rows(), y->getBuffer(), 1, 0.0, z->getBuffer(), 1);
|
|
|
|
|
|
|
|
MmulHelper::mmul(x, y, z);
|
|
|
|
|
|
|
|
//z->printBuffer();
|
|
|
|
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper3) {
|
|
|
|
auto xBuffer = new float[15]{1.f, 2.f, 3.f, 4.f, 5.f, 6.f, 7.f, 8.f, 9.f, 10.f, 11.f, 12.f, 13.f, 14.f, 15.f};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 5, 3, 1, 5, 8192, 1, 102};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[3]{2.f, 4.f, 6.f};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 3, 1, 1, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {5, 1});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[5]{92.00f, 104.00f, 116.00f, 128.00f, 140.00f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo());
|
|
|
|
|
|
|
|
//nd4j::blas::GEMV<float>::op('f', x->rows(), x->columns(), 1.0f, x->getBuffer(), y->rows(), y->getBuffer(), 1, 0.0, z->getBuffer(), 1);
|
|
|
|
|
|
|
|
MmulHelper::mmul(x, y, z);
|
|
|
|
|
|
|
|
//z->printBuffer();
|
|
|
|
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
delete[] expBuffer;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yShape;
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper4) {
|
|
|
|
auto xBuffer = new float[6]{1, 2, 3, 4, 5, 6};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 3, 2, 2, 1, 8192, 1, 99};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[6]{7, 8, 9, 0, 1, 2};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {3, 3});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[9]{7.0f, 21.0f, 35.0f, 10.0f, 28.0f, 46.0f, 13.0f, 35.0f, 57.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo());
|
|
|
|
|
|
|
|
MmulHelper::mmul(x, y, z);
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
delete[] expBuffer;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yShape;
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper5) {
|
|
|
|
auto xBuffer = new float[6]{1, 2, 3, 4, 5, 6};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 3, 2, 1, 3, 8192, 1, 102};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[6]{7, 8, 9, 0, 1, 2};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {3, 3});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[9]{7.0f, 14.0f, 21.0f, 12.0f, 21.0f, 30.0f, 17.0f, 28.0f, 39.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo());
|
|
|
|
|
|
|
|
MmulHelper::mmul(x, y, z);
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
delete[] expBuffer;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yShape;
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper6) {
|
|
|
|
auto xBuffer = new float[6]{1, 2, 3, 4, 5, 6};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 3, 2, 1, 3, 8192, 1, 102};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[6]{7, 8, 9, 0, 1, 2};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 2, 3, 1, 2, 8192, 1, 102};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {3, 3});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[9]{39.0f, 54.0f, 69.0f, 9.0f, 18.0f, 27.0f, 9.0f, 12.0f, 15.0f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo());
|
|
|
|
|
|
|
|
MmulHelper::mmul(x, y, z);
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
|
|
|
|
delete[] expBuffer;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yShape;
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper7) {
|
|
|
|
auto xBuffer = new float[15]{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15};
|
|
|
|
auto xShape = new Nd4jLong[8] {2, 5, 3, 1, 5, 8192, 1, 102};
|
|
|
|
auto x = new NDArray(xBuffer, xShape);
|
|
|
|
|
|
|
|
auto yBuffer = new float[5]{2, 4, 6, 8, 10};
|
|
|
|
auto yShape = new Nd4jLong[8] {2, 1, 5, 1, 1, 8192, 1, 99};
|
|
|
|
auto y = new NDArray(yBuffer, yShape);
|
|
|
|
|
|
|
|
auto z = NDArrayFactory::create_<float>('f', {1, 3});
|
|
|
|
|
2019-12-05 20:05:33 +01:00
|
|
|
auto expBuffer = new float[9]{110.00f, 260.00f, 410.00f};
|
2019-06-06 14:21:15 +02:00
|
|
|
auto exp = new NDArray(expBuffer, z->getShapeInfo());
|
|
|
|
|
|
|
|
MmulHelper::mmul(y, x, z);
|
|
|
|
|
|
|
|
//z->printBuffer();
|
|
|
|
ASSERT_TRUE(z->equalsTo(exp));
|
|
|
|
|
|
|
|
delete[] expBuffer;
|
|
|
|
delete[] xBuffer;
|
|
|
|
delete[] yBuffer;
|
|
|
|
delete[] xShape;
|
|
|
|
delete[] yShape;
|
|
|
|
|
|
|
|
delete x;
|
|
|
|
delete y;
|
|
|
|
delete z;
|
|
|
|
delete exp;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper_ND_1) {
|
|
|
|
Nd4jLong _expS[] = {3, 2, 3, 3, 9, 3, 1, 8192, 1, 99};
|
|
|
|
float _expB[] = {70.f, 80.f, 90.f, 158.f, 184.f, 210.f, 246.f, 288.f, 330.f, 1030.f, 1088.f, 1146.f, 1310.f, 1384.f, 1458.f, 1590.f, 1680.f, 1770.f};
|
|
|
|
|
|
|
|
auto a = NDArrayFactory::create<float>('c', {2, 3, 4});
|
|
|
|
for (int e = 0; e < a.lengthOf(); e++)
|
|
|
|
a.p(e, e+1);
|
|
|
|
|
|
|
|
auto b = NDArrayFactory::create<float>('c', {2, 4, 3});
|
|
|
|
for (int e = 0; e < b.lengthOf(); e++)
|
|
|
|
b.p(e, e+1);
|
|
|
|
|
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
auto c = MmulHelper::mmul(&a, &b);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(c));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(c));
|
|
|
|
|
|
|
|
delete c;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestMmulHelper_ND_2) {
|
|
|
|
Nd4jLong _expS[] = {3, 2, 72, 2, 144, 2, 1, 8192, 1, 99};
|
2019-12-05 20:05:33 +01:00
|
|
|
float _expB[] = {
|
|
|
|
1.07250000e+04f, 1.10500000e+04f, 2.63500000e+04f, 2.73000000e+04f, 4.19750000e+04f, 4.35500000e+04f,
|
|
|
|
5.76000000e+04f, 5.98000000e+04f, 7.32250000e+04f, 7.60500000e+04f, 8.88500000e+04f, 9.23000000e+04f,
|
|
|
|
1.04475000e+05f, 1.08550000e+05f, 1.20100000e+05f, 1.24800000e+05f, 1.35725000e+05f, 1.41050000e+05f,
|
|
|
|
1.51350000e+05f, 1.57300000e+05f, 1.66975000e+05f, 1.73550000e+05f, 1.82600000e+05f, 1.89800000e+05f,
|
|
|
|
1.98225000e+05f, 2.06050000e+05f, 2.13850000e+05f, 2.22300000e+05f, 2.29475000e+05f, 2.38550000e+05f,
|
|
|
|
2.45100000e+05f, 2.54800000e+05f, 2.60725000e+05f, 2.71050000e+05f, 2.76350000e+05f, 2.87300000e+05f,
|
|
|
|
2.91975000e+05f, 3.03550000e+05f, 3.07600000e+05f, 3.19800000e+05f, 3.23225000e+05f, 3.36050000e+05f,
|
|
|
|
3.38850000e+05f, 3.52300000e+05f, 3.54475000e+05f, 3.68550000e+05f, 3.70100000e+05f, 3.84800000e+05f,
|
|
|
|
3.85725000e+05f, 4.01050000e+05f, 4.01350000e+05f, 4.17300000e+05f, 4.16975000e+05f, 4.33550000e+05f,
|
|
|
|
4.32600000e+05f, 4.49800000e+05f, 4.48225000e+05f, 4.66050000e+05f, 4.63850000e+05f, 4.82300000e+05f,
|
|
|
|
4.79475000e+05f, 4.98550000e+05f, 4.95100000e+05f, 5.14800000e+05f, 5.10725000e+05f, 5.31050000e+05f,
|
|
|
|
5.26350000e+05f, 5.47300000e+05f, 5.41975000e+05f, 5.63550000e+05f, 5.57600000e+05f, 5.79800000e+05f,
|
|
|
|
5.73225000e+05f, 5.96050000e+05f, 5.88850000e+05f, 6.12300000e+05f, 6.04475000e+05f, 6.28550000e+05f,
|
|
|
|
6.20100000e+05f, 6.44800000e+05f, 6.35725000e+05f, 6.61050000e+05f, 6.51350000e+05f, 6.77300000e+05f,
|
|
|
|
6.66975000e+05f, 6.93550000e+05f, 6.82600000e+05f, 7.09800000e+05f, 6.98225000e+05f, 7.26050000e+05f,
|
|
|
|
7.13850000e+05f, 7.42300000e+05f, 7.29475000e+05f, 7.58550000e+05f, 7.45100000e+05f, 7.74800000e+05f,
|
|
|
|
7.60725000e+05f, 7.91050000e+05f, 7.76350000e+05f, 8.07300000e+05f, 7.91975000e+05f, 8.23550000e+05f,
|
|
|
|
8.07600000e+05f, 8.39800000e+05f, 8.23225000e+05f, 8.56050000e+05f, 8.38850000e+05f, 8.72300000e+05f,
|
|
|
|
8.54475000e+05f, 8.88550000e+05f, 8.70100000e+05f, 9.04800000e+05f, 8.85725000e+05f, 9.21050000e+05f,
|
|
|
|
9.01350000e+05f, 9.37300000e+05f, 9.16975000e+05f, 9.53550000e+05f, 9.32600000e+05f, 9.69800000e+05f,
|
|
|
|
9.48225000e+05f, 9.86050000e+05f, 9.63850000e+05f, 1.00230000e+06f, 9.79475000e+05f, 1.01855000e+06f,
|
|
|
|
9.95100000e+05f, 1.03480000e+06f, 1.01072500e+06f, 1.05105000e+06f, 1.02635000e+06f, 1.06730000e+06f,
|
|
|
|
1.04197500e+06f, 1.08355000e+06f, 1.05760000e+06f, 1.09980000e+06f, 1.07322500e+06f, 1.11605000e+06f,
|
|
|
|
1.08885000e+06f, 1.13230000e+06f, 1.10447500e+06f, 1.14855000e+06f, 1.12010000e+06f, 1.16480000e+06f,
|
|
|
|
1.13572500e+06f, 1.18105000e+06f, 1.15135000e+06f, 1.19730000e+06f, 1.16697500e+06f, 1.21355000e+06f,
|
|
|
|
3.54260000e+06f, 3.58980000e+06f, 3.58947500e+06f, 3.63730000e+06f, 3.63635000e+06f, 3.68480000e+06f,
|
|
|
|
3.68322500e+06f, 3.73230000e+06f, 3.73010000e+06f, 3.77980000e+06f, 3.77697500e+06f, 3.82730000e+06f,
|
|
|
|
3.82385000e+06f, 3.87480000e+06f, 3.87072500e+06f, 3.92230000e+06f, 3.91760000e+06f, 3.96980000e+06f,
|
|
|
|
3.96447500e+06f, 4.01730000e+06f, 4.01135000e+06f, 4.06480000e+06f, 4.05822500e+06f, 4.11230000e+06f,
|
|
|
|
4.10510000e+06f, 4.15980000e+06f, 4.15197500e+06f, 4.20730000e+06f, 4.19885000e+06f, 4.25480000e+06f,
|
|
|
|
4.24572500e+06f, 4.30230000e+06f, 4.29260000e+06f, 4.34980000e+06f, 4.33947500e+06f, 4.39730000e+06f,
|
|
|
|
4.38635000e+06f, 4.44480000e+06f, 4.43322500e+06f, 4.49230000e+06f, 4.48010000e+06f, 4.53980000e+06f,
|
|
|
|
4.52697500e+06f, 4.58730000e+06f, 4.57385000e+06f, 4.63480000e+06f, 4.62072500e+06f, 4.68230000e+06f,
|
|
|
|
4.66760000e+06f, 4.72980000e+06f, 4.71447500e+06f, 4.77730000e+06f, 4.76135000e+06f, 4.82480000e+06f,
|
|
|
|
4.80822500e+06f, 4.87230000e+06f, 4.85510000e+06f, 4.91980000e+06f, 4.90197500e+06f, 4.96730000e+06f,
|
|
|
|
4.94885000e+06f, 5.01480000e+06f, 4.99572500e+06f, 5.06230000e+06f, 5.04260000e+06f, 5.10980000e+06f,
|
|
|
|
5.08947500e+06f, 5.15730000e+06f, 5.13635000e+06f, 5.20480000e+06f, 5.18322500e+06f, 5.25230000e+06f,
|
|
|
|
5.23010000e+06f, 5.29980000e+06f, 5.27697500e+06f, 5.34730000e+06f, 5.32385000e+06f, 5.39480000e+06f,
|
|
|
|
5.37072500e+06f, 5.44230000e+06f, 5.41760000e+06f, 5.48980000e+06f, 5.46447500e+06f, 5.53730000e+06f,
|
|
|
|
5.51135000e+06f, 5.58480000e+06f, 5.55822500e+06f, 5.63230000e+06f, 5.60510000e+06f, 5.67980000e+06f,
|
|
|
|
5.65197500e+06f, 5.72730000e+06f, 5.69885000e+06f, 5.77480000e+06f, 5.74572500e+06f, 5.82230000e+06f,
|
|
|
|
5.79260000e+06f, 5.86980000e+06f, 5.83947500e+06f, 5.91730000e+06f, 5.88635000e+06f, 5.96480000e+06f,
|
|
|
|
5.93322500e+06f, 6.01230000e+06f, 5.98010000e+06f, 6.05980000e+06f, 6.02697500e+06f, 6.10730000e+06f,
|
|
|
|
6.07385000e+06f, 6.15480000e+06f, 6.12072500e+06f, 6.20230000e+06f, 6.16760000e+06f, 6.24980000e+06f,
|
|
|
|
6.21447500e+06f, 6.29730000e+06f, 6.26135000e+06f, 6.34480000e+06f, 6.30822500e+06f, 6.39230000e+06f,
|
|
|
|
6.35510000e+06f, 6.43980000e+06f, 6.40197500e+06f, 6.48730000e+06f, 6.44885000e+06f, 6.53480000e+06f,
|
|
|
|
6.49572500e+06f, 6.58230000e+06f, 6.54260000e+06f, 6.62980000e+06f, 6.58947500e+06f, 6.67730000e+06f,
|
|
|
|
6.63635000e+06f, 6.72480000e+06f, 6.68322500e+06f, 6.77230000e+06f, 6.73010000e+06f, 6.81980000e+06f,
|
|
|
|
6.77697500e+06f, 6.86730000e+06f, 6.82385000e+06f, 6.91480000e+06f, 6.87072500e+06f, 6.96230000e+06f,
|
|
|
|
6.91760000e+06f, 7.00980000e+06f, 6.96447500e+06f, 7.05730000e+06f, 7.01135000e+06f, 7.10480000e+06f,
|
|
|
|
1.17619750e+07f, 1.18560500e+07f, 1.18401000e+07f, 1.19348000e+07f, 1.19182250e+07f, 1.20135500e+07f,
|
|
|
|
1.19963500e+07f, 1.20923000e+07f, 1.20744750e+07f, 1.21710500e+07f, 1.21526000e+07f, 1.22498000e+07f, 1.22307250e+07f, 1.23285500e+07f, 1.23088500e+07f, 1.24073000e+07f, 1.23869750e+07f, 1.24860500e+07f, 1.24651000e+07f, 1.25648000e+07f, 1.25432250e+07f, 1.26435500e+07f, 1.26213500e+07f, 1.27223000e+07f, 1.26994750e+07f, 1.28010500e+07f, 1.27776000e+07f, 1.28798000e+07f, 1.28557250e+07f, 1.29585500e+07f, 1.29338500e+07f, 1.30373000e+07f, 1.30119750e+07f, 1.31160500e+07f, 1.30901000e+07f, 1.31948000e+07f, 1.31682250e+07f, 1.32735500e+07f, 1.32463500e+07f, 1.33523000e+07f, 1.33244750e+07f, 1.34310500e+07f, 1.34026000e+07f, 1.35098000e+07f, 1.34807250e+07f, 1.35885500e+07f, 1.35588500e+07f, 1.36673000e+07f, 1.36369750e+07f, 1.37460500e+07f, 1.37151000e+07f, 1.38248000e+07f, 1.37932250e+07f, 1.39035500e+07f, 1.38713500e+07f, 1.39823000e+07f, 1.39494750e+07f, 1.40610500e+07f, 1.40276000e+07f, 1.41398000e+07f, 1.41057250e+07f, 1.42185500e+07f, 1.41838500e+07f, 1.42973000e+07f, 1.42619750e+07f, 1.43760500e+07f, 1.43401000e+07f, 1.44548000e+07f, 1.44182250e+07f, 1.45335500e+07f, 1.44963500e+07f, 1.46123000e+07f, 1.45744750e+07f, 1.46910500e+07f, 1.46526000e+07f, 1.47698000e+07f, 1.47307250e+07f, 1.48485500e+07f, 1.48088500e+07f, 1.49273000e+07f, 1.48869750e+07f, 1.50060500e+07f, 1.49651000e+07f, 1.50848000e+07f, 1.50432250e+07f, 1.51635500e+07f, 1.51213500e+07f, 1.52423000e+07f, 1.51994750e+07f, 1.53210500e+07f, 1.52776000e+07f, 1.53998000e+07f, 1.53557250e+07f, 1.54785500e+07f, 1.54338500e+07f, 1.55573000e+07f, 1.55119750e+07f, 1.56360500e+07f, 1.55901000e+07f, 1.57148000e+07f, 1.56682250e+07f, 1.57935500e+07f, 1.57463500e+07f, 1.58723000e+07f, 1.58244750e+07f, 1.59510500e+07f, 1.59026000e+07f, 1.60298000e+07f, 1.59807250e+07f, 1.61085500e+07f, 1.60588500e+07f, 1.61873000e+07f, 1.61369750e+07f, 1.62660500e+07f, 1.62151000e+07f, 1.63448000e+07f, 1.62932250e+07f, 1.64235500e+07f, 1.63713500e+07f, 1.65023000e+07f, 1.64494750e+07f, 1.65810500e+07f, 1.65276000e+07f, 1.66598000e+07f, 1.66057250e+07f, 1.67385500e+07f, 1.66838500e+07f, 1.68173000e+07f, 1.67619750e+07f, 1.68960500e+07f, 1.68401000e+07f, 1.69748000e+07f, 1.69182250e+07f, 1.70535500e+07f, 1.69963500e+07f, 1.71323000e+07f, 1.70744750e+07f, 1.72110500e+07f, 1.71526000e+07f, 1.72898000e+07f, 1.72307250e+07f, 1.73685500e+07f, 1.73088500e+07f, 1.74473000e+07f, 1.73869750e+07f, 1.75260500e+07f, 1.74651000e+07f, 1.76048000e+07f, 1.75432250e+07f, 1.76835500e+07f, 2.46688500e+07f, 2.48098000e+07f, 2.47782250e+07f, 2.49198000e+07f, 2.48876000e+07f, 2.50298000e+07f, 2.49969750e+07f, 2.51398000e+07f, 2.51063500e+07f, 2.52498000e+07f, 2.52157250e+07f, 2.53598000e+07f, 2.53251000e+07f, 2.54698000e+07f, 2.54344750e+07f, 2.55798000e+07f, 2.55438500e+07f, 2.56898000e+07f, 2.56532250e+07f, 2.57998000e+07f, 2.57626000e+07f, 2.59098000e+07f, 2.58719750e+07f, 2.60198000e+07f, 2.59813500e+07f, 2.61298000e+07f, 2.60907250e+07f, 2.62398000e+07f, 2.62001000e+07f, 2.63498000e+07f, 2.63094750e+07f, 2.64598000e+07f, 2.64188500e+07f, 2.65698000e+07f, 2.65282250e+07f, 2.66798000e+07f, 2.66376000e+07f, 2.67898000e+07f, 2.67469750e+07f, 2.68998000e+07f, 2.68563500e+07f, 2.70098000e+07f, 2.69657250e+07f, 2.71198000e+07f, 2.70751000e+07f, 2.72298000e+07f, 2.71844750e+07f, 2.73398000e+07f, 2.72938500e+07f, 2.74498000e+07f, 2.74032250e+07f, 2.75598000e+07f, 2.75126000e+07f, 2.76698000e+07f, 2.76219750e+07f, 2.77798000e+07f, 2.77313500e+07f, 2.78898000e+07f, 2.78407250e+07f, 2.79998000e+07f, 2.79501000
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
auto a = NDArrayFactory::create<float>('c', {2, 72, 25});
|
|
|
|
for (int e = 0; e < a.lengthOf(); e++)
|
|
|
|
a.p(e, e+1);
|
|
|
|
|
|
|
|
auto b = NDArrayFactory::create<float>('c', {2, 25, 2});
|
|
|
|
for (int e = 0; e < b.lengthOf(); e++)
|
|
|
|
b.p(e, e+1);
|
|
|
|
|
|
|
|
NDArray exp(_expB, _expS);
|
|
|
|
|
|
|
|
auto c = MmulHelper::mmul(&a, &b);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(c));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(c, 1e1));
|
|
|
|
|
|
|
|
delete c;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestNegSize1) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {2, 5, 7});
|
|
|
|
|
|
|
|
ASSERT_EQ(7, array.sizeAt(-1));
|
|
|
|
ASSERT_EQ(5, array.sizeAt(-2));
|
|
|
|
ASSERT_EQ(2, array.sizeAt(-3));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
// not-in-place
|
|
|
|
TEST_F(NDArrayTest, Permute1) {
|
|
|
|
|
|
|
|
Nd4jLong shape1[] = {3, 5, 10, 15, 150, 15, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong shape2[] = {3, 15, 5, 10, 1, 150, 15, 8192, 0, 99};
|
|
|
|
const std::initializer_list<int> perm = {2, 0, 1};
|
|
|
|
|
|
|
|
NDArray arr1(shape1,true);
|
|
|
|
NDArray arr2(shape2,true);
|
|
|
|
|
|
|
|
auto result = arr1.permute(perm);
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
ASSERT_TRUE(result.isSameShapeStrict(&arr2));
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
// in-place
|
|
|
|
TEST_F(NDArrayTest, Permute2) {
|
|
|
|
|
|
|
|
Nd4jLong shape1[] = {3, 5, 10, 15, 150, 15, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong shape2[] = {3, 15, 5, 10, 1, 150, 15, 8192, 0, 99};
|
|
|
|
const std::initializer_list<int> perm = {2, 0, 1};
|
|
|
|
|
|
|
|
NDArray arr1(shape1,true);
|
|
|
|
NDArray arr2(shape2,true);
|
|
|
|
|
|
|
|
ASSERT_TRUE(arr1.permutei(perm));
|
|
|
|
ASSERT_TRUE(arr1.isSameShapeStrict(&arr2));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Broadcast1) {
|
|
|
|
|
|
|
|
Nd4jLong shape1[10] = {3, 5, 1, 10, 10, 10, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong shape2[8] = {2, 7, 10, 10, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong shape3[10] = {3, 5, 7, 10, 70, 10, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray arr1(shape1);
|
|
|
|
NDArray arr2(shape2);
|
|
|
|
NDArray arr3(shape3);
|
|
|
|
|
|
|
|
auto result = arr1.broadcast(arr2);
|
|
|
|
ASSERT_TRUE(result->isSameShapeStrict(&arr3));
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, RSubScalarTest1) {
|
|
|
|
auto array = NDArrayFactory::create<double>('c', {1, 4});
|
|
|
|
array.assign(2.0);
|
|
|
|
|
|
|
|
auto result = NDArrayFactory::create<double>('c', {1, 4});
|
|
|
|
|
|
|
|
array.applyScalar(scalar::ReverseSubtract, 1.0, &result);
|
|
|
|
|
|
|
|
ASSERT_NEAR(-1.0, result.meanNumber().e<double>(0), 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, BroadcastOpsTest1) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {5, 5});
|
|
|
|
auto row = NDArrayFactory::linspace(1.0f, 5.0f, 5);
|
|
|
|
float *brow = new float[5]{1,2,3,4,5};
|
|
|
|
auto bshape = new Nd4jLong[8]{2, 1, 5, 1, 1, 8192, 1, 99};
|
|
|
|
float *ebuf = new float[25] {1, 2, 3, 4, 5, 1, 2, 3, 4, 5, 1, 2, 3, 4, 5, 1, 2, 3, 4, 5, 1, 2, 3, 4, 5};
|
|
|
|
auto eshape = new Nd4jLong[8] {2, 5, 5, 5, 1, 8192, 1, 99};
|
|
|
|
NDArray expRow(brow, bshape);
|
|
|
|
NDArray exp(ebuf, eshape);
|
|
|
|
|
|
|
|
ASSERT_TRUE(row->equalsTo(&expRow));
|
|
|
|
|
|
|
|
|
|
|
|
x.applyBroadcast(broadcast::Add, {1}, row);
|
|
|
|
|
|
|
|
//x.printBuffer("Result");
|
|
|
|
|
|
|
|
ASSERT_TRUE(x.equalsTo(&exp));
|
|
|
|
|
|
|
|
delete[] brow;
|
|
|
|
delete[] bshape;
|
|
|
|
delete[] ebuf;
|
|
|
|
delete[] eshape;
|
|
|
|
delete row;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexedPut2) {
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2});
|
|
|
|
//x.printShapeInfo("x shape");
|
|
|
|
x.p(1, 1.0f);
|
|
|
|
|
|
|
|
//x.printBuffer("after");
|
|
|
|
ASSERT_NEAR(reinterpret_cast<float *>(x.getBuffer())[2], 1.0, 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexedPut3) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
x.p(1, 1.0f);
|
|
|
|
|
|
|
|
//x.printBuffer("after");
|
|
|
|
ASSERT_NEAR(reinterpret_cast<float *>(x.getBuffer())[1], 1.0, 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexedPut4) {
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2});
|
|
|
|
x.p(0, 1, 1.0f);
|
|
|
|
|
|
|
|
//x.printBuffer("after");
|
|
|
|
ASSERT_NEAR(reinterpret_cast<float *>(x.getBuffer())[2], 1.0, 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexedPut5) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
x.p(0, 1, 1.0f);
|
|
|
|
|
|
|
|
//x.printBuffer("after");
|
|
|
|
ASSERT_NEAR(x.bufferAsT<float>()[1], 1.0, 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestAllTensors1) {
|
|
|
|
auto matrix = NDArrayFactory::create<float>('c', {3, 5});
|
|
|
|
|
|
|
|
std::unique_ptr<ResultSet> rows(matrix.allTensorsAlongDimension({1}));
|
|
|
|
|
|
|
|
ASSERT_EQ(3, rows->size());
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexing1) {
|
|
|
|
auto matrix = NDArrayFactory::create<float>('c', {5, 5});
|
|
|
|
for (int e = 0; e < matrix.lengthOf(); e++)
|
|
|
|
matrix.p(e, (float) e);
|
|
|
|
|
|
|
|
auto sub = matrix({2,4, 0,0}, true);
|
|
|
|
|
|
|
|
ASSERT_EQ(2, sub.rows());
|
|
|
|
ASSERT_EQ(5, sub.columns());
|
|
|
|
|
|
|
|
ASSERT_NEAR(10, sub.e<float>(0), 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexing2) {
|
|
|
|
auto matrix = NDArrayFactory::create<float>('c', {2, 5, 4, 4});
|
|
|
|
matrix.linspace(0);
|
|
|
|
|
|
|
|
auto sub = matrix({0,0, 2,4, 0,0, 0,0}, true);
|
|
|
|
|
|
|
|
ASSERT_EQ(2, sub.sizeAt(0));
|
|
|
|
ASSERT_EQ(2, sub.sizeAt(1));
|
|
|
|
ASSERT_EQ(4, sub.sizeAt(2));
|
|
|
|
ASSERT_EQ(4, sub.sizeAt(3));
|
|
|
|
|
|
|
|
ASSERT_EQ(64, sub.lengthOf());
|
|
|
|
ASSERT_NEAR(32, sub.e<float>(0), 1e-5);
|
|
|
|
ASSERT_NEAR(112, sub.e<float>(32), 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexing3) {
|
|
|
|
auto matrix = NDArrayFactory::create<float>('c', {5, 5});
|
|
|
|
matrix.linspace(0);
|
|
|
|
|
|
|
|
auto sub = matrix({2,4, 0,0});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, sub.rows());
|
|
|
|
ASSERT_EQ(5, sub.columns());
|
|
|
|
|
|
|
|
ASSERT_NEAR(10, sub.e<float>(0), 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestIndexing4) {
|
|
|
|
auto matrix = NDArrayFactory::create<float>('c', {2, 5, 4, 4});
|
|
|
|
matrix.linspace(0);
|
|
|
|
|
|
|
|
auto sub = matrix({0,0, 2,4, 0,0, 0,0});
|
|
|
|
|
|
|
|
ASSERT_EQ(2, sub.sizeAt(0));
|
|
|
|
ASSERT_EQ(2, sub.sizeAt(1));
|
|
|
|
ASSERT_EQ(4, sub.sizeAt(2));
|
|
|
|
ASSERT_EQ(4, sub.sizeAt(3));
|
|
|
|
|
|
|
|
|
|
|
|
ASSERT_EQ(64, sub.lengthOf());
|
|
|
|
ASSERT_NEAR(32, sub.e<float>(0), 1e-5);
|
|
|
|
ASSERT_NEAR(112, sub.e<float>(32), 1e-5);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestReshapeNegative1) {
|
|
|
|
std::unique_ptr<NDArray> array(NDArrayFactory::create_<float>('c', {2, 3, 4, 64}));
|
|
|
|
|
|
|
|
array->reshapei('c', {-1, 64});
|
|
|
|
|
|
|
|
ASSERT_EQ(24, array->sizeAt(0));
|
|
|
|
ASSERT_EQ(64, array->sizeAt(1));
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestReshapeNegative2) {
|
|
|
|
std::unique_ptr<NDArray> array(NDArrayFactory::create_<float>('c', {2, 3, 4, 64}));
|
|
|
|
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
auto reshaped = array->reshape('c', {-1, 64});
|
2019-06-06 14:21:15 +02:00
|
|
|
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
ASSERT_EQ(24, reshaped.sizeAt(0));
|
|
|
|
ASSERT_EQ(64, reshaped.sizeAt(1));
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
// TEST_F(NDArrayTest, SVD1) {
|
|
|
|
|
|
|
|
// double arrA[8] = {1, 2, 3, 4, 5, 6, 7, 8};
|
|
|
|
// double arrU[8] = {-0.822647, 0.152483, -0.421375, 0.349918, -0.020103, 0.547354, 0.381169, 0.744789};
|
|
|
|
// double arrS[2] = {0.626828, 14.269095};
|
|
|
|
// double arrVt[4] = {0.767187,-0.641423, 0.641423, 0.767187};
|
|
|
|
|
|
|
|
// int shapeA[8] = {2, 4, 2, 2, 1, 0, 1, 99};
|
|
|
|
// int shapeS[8] = {2, 1, 2, 2, 1, 0, 1, 99};
|
|
|
|
// int shapeVt[8] = {2, 2, 2, 2, 1, 0, 1, 99};
|
|
|
|
|
|
|
|
// auto a(arrA, shapeA);
|
|
|
|
// auto u(arrU, shapeA);
|
|
|
|
// auto s(arrS, shapeS);
|
|
|
|
// auto vt(arrVt, shapeVt);
|
|
|
|
// auto expU, expS(shapeS), expVt(shapeVt);
|
|
|
|
|
|
|
|
// a.svd(expU, expS, expVt);
|
|
|
|
// ASSERT_TRUE(u.equalsTo(&expU));
|
|
|
|
// ASSERT_TRUE(s.equalsTo(&expS));
|
|
|
|
// ASSERT_TRUE(vt.equalsTo(&expVt));
|
|
|
|
|
|
|
|
// }
|
|
|
|
|
|
|
|
// //////////////////////////////////////////////////////////////////////
|
|
|
|
// TEST_F(NDArrayTest, SVD2) {
|
|
|
|
|
|
|
|
// double arrA[6] = {1, 2, 3, 4, 5, 6};
|
|
|
|
// double arrU[6] = {-0.386318, -0.922366, 0.000000, -0.922366, 0.386318, 0.000000};
|
|
|
|
// double arrS[3] = {9.508032, 0.77287, 0.000};
|
|
|
|
// double arrVt[9] = {-0.428667, -0.566307, -0.703947, 0.805964, 0.112382, -0.581199, 0.408248, -0.816497, 0.408248};
|
|
|
|
|
|
|
|
// int shapeA[8] = {2, 2, 3, 3, 1, 0, 1, 99};
|
|
|
|
// int shapeS[8] = {2, 1, 3, 3, 1, 0, 1, 99};
|
|
|
|
// int shapeVt[8] = {2, 3, 3, 3, 1, 0, 1, 99};
|
|
|
|
|
|
|
|
// auto a(arrA, shapeA);
|
|
|
|
// auto u(arrU, shapeA);
|
|
|
|
// auto s(arrS, shapeS);
|
|
|
|
// auto vt(arrVt, shapeVt);
|
|
|
|
// auto expU, expS(shapeS), expVt(shapeVt);
|
|
|
|
|
|
|
|
// a.svd(expU, expS, expVt);
|
|
|
|
// ASSERT_TRUE(u.equalsTo (&expU));
|
|
|
|
// ASSERT_TRUE(s.equalsTo(&expS));
|
|
|
|
// ASSERT_TRUE(vt.equalsTo(&expVt));
|
|
|
|
|
|
|
|
// }
|
|
|
|
|
|
|
|
// //////////////////////////////////////////////////////////////////////
|
|
|
|
// TEST_F(NDArrayTest, SVD3) {
|
|
|
|
|
|
|
|
// double arrA[8] = {1, 2, 3, 4, 5, 6, 7, 8};
|
|
|
|
// double arrU[8] = {-0.822647, 0.152483, -0.421375, 0.349918, -0.020103, 0.547354, 0.381169, 0.744789};
|
|
|
|
// double arrS[2] = {0.626828, 14.269095};
|
|
|
|
// double arrVt[4] = {0.767187,-0.641423, 0.641423, 0.767187};
|
|
|
|
|
|
|
|
// int shapeA[8] = {2, 4, 2, 2, 1, 0, 1, 99};
|
|
|
|
// int shapeS[8] = {2, 1, 2, 2, 1, 0, 1, 99};
|
|
|
|
// int shapeVt[8] = {2, 2, 2, 2, 1, 0, 1, 99};
|
|
|
|
|
|
|
|
// auto a(arrA, shapeA);
|
|
|
|
// auto u(arrU, shapeA);
|
|
|
|
// auto s(arrS, shapeS);
|
|
|
|
// auto vt(arrVt, shapeVt);
|
|
|
|
// auto expU, expS(shapeS), expVt(shapeVt);
|
|
|
|
|
|
|
|
// a.svd(expU, expS, expVt);
|
|
|
|
// ASSERT_TRUE(expU.hasOrthonormalBasis(1));
|
|
|
|
// ASSERT_TRUE(expVt.hasOrthonormalBasis(0));
|
|
|
|
// ASSERT_TRUE(expVt.hasOrthonormalBasis(1));
|
|
|
|
// ASSERT_TRUE(expVt.isUnitary());
|
|
|
|
// }
|
|
|
|
|
|
|
|
// //////////////////////////////////////////////////////////////////////
|
|
|
|
// TEST_F(NDArrayTest, SVD4) {
|
|
|
|
|
|
|
|
// double arrA[6] = {1, 2, 3, 4, 5, 6};
|
|
|
|
// double arrU[6] = {-0.386318, -0.922366, 0.000000, -0.922366, 0.386318, 0.000000};
|
|
|
|
// double arrS[3] = {9.508032, 0.77287, 0.000};
|
|
|
|
// double arrVt[9] = {-0.428667, -0.566307, -0.703947, 0.805964, 0.112382, -0.581199, 0.408248, -0.816497, 0.408248};
|
|
|
|
|
|
|
|
// int shapeA[8] = {2, 2, 3, 3, 1, 0, 1, 99};
|
|
|
|
// int shapeS[8] = {2, 1, 3, 3, 1, 0, 1, 99};
|
|
|
|
// int shapeVt[8] = {2, 3, 3, 3, 1, 0, 1, 99};
|
|
|
|
|
|
|
|
// auto a(arrA, shapeA);
|
|
|
|
// auto u(arrU, shapeA);
|
|
|
|
// auto s(arrS, shapeS);
|
|
|
|
// auto vt(arrVt, shapeVt);
|
|
|
|
// auto expU, expS(shapeS), expVt(shapeVt);
|
|
|
|
|
|
|
|
// a.svd(expU, expS, expVt);
|
|
|
|
// ASSERT_TRUE(expU.hasOrthonormalBasis(1));
|
|
|
|
// ASSERT_TRUE(expVt.hasOrthonormalBasis(0));
|
|
|
|
// ASSERT_TRUE(expVt.hasOrthonormalBasis(1));
|
|
|
|
// ASSERT_TRUE(expVt.isUnitary());
|
|
|
|
// }
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestStdDev1) {
|
|
|
|
auto array = NDArrayFactory::create<double>('c', {1, 5});
|
|
|
|
for (int e = 0; e < array.lengthOf(); e++)
|
|
|
|
array.p(e, e+1);
|
|
|
|
|
|
|
|
auto std = array.varianceNumber(variance::SummaryStatsStandardDeviation, true).e<double>(0);
|
|
|
|
ASSERT_NEAR(std, 1.58109, 1e-4);
|
|
|
|
}
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestStdDev2) {
|
|
|
|
auto array = NDArrayFactory::create<double>('c', {5, 6});
|
|
|
|
auto tad = array.tensorAlongDimension(0, {0});
|
|
|
|
|
|
|
|
ASSERT_EQ(5, tad->lengthOf());
|
|
|
|
|
|
|
|
for (int e = 0; e < tad->lengthOf(); e++)
|
|
|
|
tad->p(e, e+1);
|
|
|
|
|
|
|
|
ASSERT_NEAR(15, tad->sumNumber().e<double>(0), 1e-5);
|
|
|
|
|
|
|
|
auto std = tad->varianceNumber(variance::SummaryStatsStandardDeviation, true).e<double>(0);
|
|
|
|
ASSERT_NEAR(std, 1.58109, 1e-4);
|
|
|
|
|
|
|
|
delete tad;
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestStdDev3) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {1, 50000});
|
|
|
|
for (int e = 0; e < array.lengthOf(); e++)
|
|
|
|
array.p(e, 1.f + (e%2?0.5f:-0.5f));
|
|
|
|
|
|
|
|
auto std = array.varianceNumber(variance::SummaryStatsStandardDeviation, true).e<double>(0);
|
|
|
|
// nd4j_printf("Variance is %f\n", std);
|
|
|
|
ASSERT_NEAR(std, 0.5f, 1.0e-5f);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestStdDev4) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {1, 20000});
|
|
|
|
float const ethalon = 1 / 3.f;
|
|
|
|
float x = ethalon;
|
|
|
|
int total = array.lengthOf();
|
|
|
|
for (int e = 0; e < total; e++) {
|
|
|
|
array.p(e, 1.0f + (e % 2?ethalon:-ethalon));
|
|
|
|
x *= (e % 2? 2.f: 0.5f);
|
|
|
|
}
|
|
|
|
x = 0.f;
|
|
|
|
for (int e = 0; e < total; ++e) {
|
|
|
|
x += array.e<float>(e);
|
|
|
|
}
|
|
|
|
x /= array.lengthOf();
|
|
|
|
float y = 0;
|
|
|
|
double M2 = 0;
|
|
|
|
for (int e = 0; e < total; ++e) {
|
|
|
|
// y += nd4j::math::nd4j_abs(array(e) - x);
|
|
|
|
M2 += (array.e<float>(e) - x) * (array.e<float>(e) - x);
|
|
|
|
}
|
|
|
|
//y /= total;
|
|
|
|
M2 /= total;
|
|
|
|
|
|
|
|
y = M2;
|
|
|
|
auto a = array.varianceNumber(variance::SummaryStatsStandardDeviation, false);
|
|
|
|
auto std = a.e<float>(0);
|
|
|
|
// float bY = array.varianceNumber();
|
|
|
|
float bY = 0.3333333f;
|
|
|
|
// nd4j_printf("Variance is %f, res is %f, internal is %f\n, deviance is %f(%f)\n", std, x, bY, y, nd4j::math::nd4j_sqrt<double>(M2));
|
|
|
|
ASSERT_NEAR(std, 0.3333333f, 1.0e-5f);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestStdDev5) {
|
|
|
|
auto array = NDArrayFactory::create<float>('c', {1, 10000}); //00000});
|
|
|
|
auto arrayD = NDArrayFactory::create<double>('c', {1, 10000}); //00000});
|
|
|
|
for (int e = 0; e < array.lengthOf(); e++) {
|
|
|
|
array.p(e, 1.f + (e%2?1/5.f:-1/5.f));
|
|
|
|
arrayD.p(e, 1.0 + (e%2?1/5.:-1/5.));
|
|
|
|
}
|
|
|
|
float stdF = array.varianceNumber(variance::SummaryStatsStandardDeviation, false).e<float>(0);
|
|
|
|
double stdD = arrayD.varianceNumber(variance::SummaryStatsStandardDeviation, false).e<double>(0);
|
|
|
|
// nd4j_printf("Variance is %f(%f)\n", stdF, stdD);
|
|
|
|
ASSERT_NEAR(stdD, 0.2, 1.0e-8); // 1/5 = 0.2
|
|
|
|
ASSERT_NEAR(stdF, 0.2f, 1.0e-5f); // 1/5 = 0.2
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestApplyIndexReduce1) {
|
|
|
|
float xBuff[] = {1, 5, 2, 12, 9, 3, 10, 7, 4, 11, 6, 8};
|
|
|
|
Nd4jLong xShapeInfo[] = {3, 2, 3, 2, 6, 2, 1, 8192, 1, 99};
|
|
|
|
std::vector<int> dim = {0,1};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
auto exp = NDArrayFactory::create<Nd4jLong>({3, 1});
|
|
|
|
|
|
|
|
auto result = x.applyIndexReduce(indexreduce::IndexMax, dim);
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, applyReduce3Dot) {
|
|
|
|
float xBuff[] = {1, 2, 3, 4, 5, 6};
|
|
|
|
float yBuff[] = {2, 2, 2, 2, 2, 2};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, xShapeInfo);
|
|
|
|
|
|
|
|
auto result = x.applyReduce3(reduce3::Dot, &y);
|
|
|
|
ASSERT_TRUE(result->lengthOf() == 1);
|
|
|
|
ASSERT_NEAR(42, result->e<float>(0), 1e-5);
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, applyAllReduce3EuclideanDistance) {
|
|
|
|
float xBuff[] = {1, 2, 3, 4, 5, 6};
|
|
|
|
float yBuff[] = {2, 2, 2, 2, 2, 2};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {1.414214f, 1.414214f, 5.385165f, 5.385165f};
|
2019-06-06 14:21:15 +02:00
|
|
|
Nd4jLong expShapeInfo[] = {2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, xShapeInfo);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {2, 2}, {1.414214f, 1.414214f, 5.385165f, 5.385165f});
|
|
|
|
|
|
|
|
auto result = x.applyAllReduce3(reduce3::EuclideanDistance, &y,{1});
|
|
|
|
|
|
|
|
// result->printIndexedBuffer("result");
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, applyReduce3EuclideanDistance) {
|
|
|
|
float xBuff[] = {1, 2, 3, 4, 5, 6};
|
|
|
|
float yBuff[] = {2, 2, 2, 2, 2, 2};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {1.414214f, 1.414214f, 5.385165f, 5.385165f};
|
2019-06-06 14:21:15 +02:00
|
|
|
Nd4jLong expShapeInfo[] = {2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, xShapeInfo);
|
|
|
|
NDArray exp(expBuff, expShapeInfo);
|
|
|
|
|
|
|
|
auto result = x.applyAllReduce3(reduce3::EuclideanDistance, &y,{1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestVarianceAlongDimension1) {
|
|
|
|
|
|
|
|
float xBuff[] = {1, 2, 3, 4, 5, 6};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {0.816497f, 0.816497f};
|
2019-06-06 14:21:15 +02:00
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong expShapeInfo[] = {1, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray exp(expBuff, expShapeInfo);
|
|
|
|
|
|
|
|
auto result = x.varianceAlongDimension(variance::SummaryStatsStandardDeviation, false, {1});
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestVarianceAlongDimension2) {
|
|
|
|
float xBuff[] = {1, 2, 3, 4, 5, 6};
|
2019-12-05 20:05:33 +01:00
|
|
|
float expBuff[] = {0.666667f, 0.666667f};
|
2019-06-06 14:21:15 +02:00
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 3, 3, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong expShapeInfo[] = {1, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray exp(expBuff, expShapeInfo);
|
|
|
|
|
|
|
|
auto result = x.varianceAlongDimension(variance::SummaryStatsVariance, false, {1});
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestVarianceAlongDimension3) {
|
|
|
|
|
|
|
|
|
|
|
|
NDArray x = NDArrayFactory::create<double>('c', {10, 10});//(xBuff, xShapeInfo);
|
|
|
|
NDArray exp = NDArrayFactory::create<double>('c', {10});//(expBuff, expShapeInfo);
|
|
|
|
x.linspace(1); // 1, 2, 3, ..., 100
|
|
|
|
exp.assign(825.f);
|
|
|
|
auto result = x.varianceAlongDimension(variance::SummaryStatsVariance, false, {0});
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestVarianceAlongDimension4) {
|
|
|
|
|
|
|
|
|
|
|
|
NDArray x = NDArrayFactory::create<double>('c', {12, 1, 12});//(xBuff, xShapeInfo);
|
|
|
|
NDArray exp = NDArrayFactory::create<double>('c', {1,12});//(expBuff, expShapeInfo);
|
|
|
|
x.linspace(1); // 1, 2, 3, ..., 100
|
|
|
|
exp.assign(1716.);
|
|
|
|
auto result = x.varianceAlongDimension(variance::SummaryStatsVariance, false, {0});
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(result));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestSubRowVector1) {
|
|
|
|
float xBuff[] = {6, 7, 8, 9};
|
|
|
|
float yBuff[] = {1, 2};
|
|
|
|
float expBuff[] = {5, 5, 7, 7};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong yShapeInfo[] = {2, 1, 2, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, yShapeInfo);
|
|
|
|
NDArray target(x);
|
|
|
|
NDArray exp(expBuff, xShapeInfo);
|
|
|
|
|
|
|
|
x.subRowVector(&y,&target);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(&target));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&target));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestDivRowVector1) {
|
|
|
|
float xBuff[] = {6, 8, 10, 12};
|
|
|
|
float yBuff[] = {2, 4};
|
|
|
|
float expBuff[] = {3, 2, 5, 3};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong yShapeInfo[] = {2, 1, 2, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, yShapeInfo);
|
|
|
|
NDArray target(x);
|
|
|
|
NDArray exp(expBuff, xShapeInfo);
|
|
|
|
|
|
|
|
x.divRowVector(&y,&target);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(&target));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&target));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMulRowVector1) {
|
|
|
|
float xBuff[] = {6, 8, 10, 12};
|
|
|
|
float yBuff[] = {2, 4};
|
|
|
|
float expBuff[] = {12, 32, 20, 48};
|
|
|
|
Nd4jLong xShapeInfo[] = {2, 2, 2, 2, 1, 8192, 1, 99};
|
|
|
|
Nd4jLong yShapeInfo[] = {2, 1, 2, 2, 1, 8192, 1, 99};
|
|
|
|
|
|
|
|
NDArray x(xBuff, xShapeInfo);
|
|
|
|
NDArray y(yBuff, yShapeInfo);
|
|
|
|
NDArray target(x);
|
|
|
|
NDArray exp(expBuff, xShapeInfo);
|
|
|
|
|
|
|
|
x.mulRowVector(&y,&target);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShapeStrict(&target));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&target));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestTensorDotAgain_1) {
|
|
|
|
int sY = 1;
|
|
|
|
int sX = 1;
|
|
|
|
int pY = 0;
|
|
|
|
int pX = 0;
|
|
|
|
int iC = 2;
|
|
|
|
int oC = 2;
|
|
|
|
int kY = 3;
|
|
|
|
int kX = 3;
|
|
|
|
int iY = 2;
|
|
|
|
int iX = 2;
|
|
|
|
int oY = 6;
|
|
|
|
int oX = 6;
|
|
|
|
int eD = iC * oC;
|
|
|
|
int B = 2;
|
|
|
|
|
|
|
|
/*
|
|
|
|
input = np.linspace(1, B * iC * iY * iX, B * iC * iY * iX).reshape(B, iC, iY, iX)
|
|
|
|
weights = np.linspace(1, iC * oC * kY * kX, iC * oC * kY * kX).reshape(iC, oC, kY, kX)
|
|
|
|
*/
|
|
|
|
double _expB[] = {96.0, 116.0, 136.0, 156.0, 256.0, 276.0, 296.0, 316.0, 102.0, 124.0, 146.0, 168.0, 278.0, 300.0, 322.0, 344.0, 108.0, 132.0, 156.0, 180.0, 300.0, 324.0, 348.0, 372.0, 114.0, 140.0, 166.0, 192.0, 322.0, 348.0, 374.0, 400.0, 120.0, 148.0, 176.0, 204.0, 344.0, 372.0, 400.0, 428.0, 126.0, 156.0, 186.0, 216.0, 366.0, 396.0, 426.0, 456.0, 132.0, 164.0, 196.0, 228.0, 388.0, 420.0, 452.0, 484.0, 138.0, 172.0, 206.0, 240.0, 410.0, 444.0, 478.0, 512.0, 144.0, 180.0, 216.0, 252.0, 432.0, 468.0, 504.0, 540.0, 150.0, 188.0, 226.0, 264.0, 454.0, 492.0, 530.0, 568.0, 156.0, 196.0, 236.0, 276.0, 476.0, 516.0, 556.0, 596.0, 162.0, 204.0, 246.0, 288.0, 498.0, 540.0, 582.0, 624.0, 168.0, 212.0, 256.0, 300.0, 520.0, 564.0, 608.0, 652.0, 174.0, 220.0, 266.0, 312.0, 542.0, 588.0, 634.0, 680.0, 180.0, 228.0, 276.0, 324.0, 564.0, 612.0, 660.0, 708.0, 186.0, 236.0, 286.0, 336.0, 586.0, 636.0, 686.0, 736.0, 192.0, 244.0, 296.0, 348.0, 608.0, 660.0, 712.0, 764.0, 198.0, 252.0, 306.0, 360.0, 630.0, 684.0, 738.0, 792.0};
|
|
|
|
|
|
|
|
Nd4jLong _expS[] = {6, 2, 3, 3, 2, 2, 2, 72, 24, 8, 4, 2, 1, 16384, 1, 99};
|
|
|
|
NDArray exp(_expB, _expS, nd4j::LaunchContext ::defaultContext(), false);
|
|
|
|
|
|
|
|
auto input = NDArrayFactory::create<double>('c', {B, iC, iY, iX});
|
|
|
|
auto weights = NDArrayFactory::create<double>('c', {iC, oC, kY, kX});
|
|
|
|
|
|
|
|
input.linspace(1);
|
|
|
|
weights.linspace(1);
|
|
|
|
|
|
|
|
auto result = MmulHelper::tensorDot(&weights, &input, {0}, {1});
|
|
|
|
|
|
|
|
//result->printShapeInfo("result shape");
|
|
|
|
ASSERT_TRUE(exp.isSameShape(result));
|
|
|
|
|
|
|
|
// exp.printBuffer("Expctd buffer");
|
|
|
|
// result->printBuffer("Result buffer");
|
|
|
|
ASSERT_TRUE(exp.equalsTo(result));
|
|
|
|
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestBroadcast_1) {
|
|
|
|
double _expB[] = {1.000000, 1.000000, 1.000000, 1.000000, 2.000000, 2.000000, 2.000000, 2.000000, 3.000000, 3.000000, 3.000000, 3.000000, 1.000000, 1.000000, 1.000000, 1.000000, 2.000000, 2.000000, 2.000000, 2.000000, 3.000000, 3.000000, 3.000000, 3.000000};
|
|
|
|
Nd4jLong _expS[] = {4, 2, 3, 2, 2, 12, 4, 2, 1, 16384, 1, 99};
|
|
|
|
NDArray exp(_expB, _expS, nd4j::LaunchContext ::defaultContext(), false);
|
|
|
|
|
|
|
|
auto input = NDArrayFactory::create<double>('c',{ 2, 3, 2, 2});
|
|
|
|
auto bias = NDArrayFactory::create<double>('c', {1, 3});
|
|
|
|
|
|
|
|
bias.linspace(1);
|
|
|
|
|
|
|
|
input.applyBroadcast(broadcast::Add, {1}, &bias);
|
|
|
|
|
|
|
|
//input.printBuffer("result");
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&input));
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestTranspose_11) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 3, 4});
|
|
|
|
x.transposei();
|
|
|
|
|
|
|
|
ASSERT_EQ(4, x.sizeAt(0));
|
|
|
|
ASSERT_EQ(3, x.sizeAt(1));
|
|
|
|
ASSERT_EQ(2, x.sizeAt(2));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, TestTranspose_12) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 3, 4});
|
|
|
|
auto y = x.transpose();
|
|
|
|
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
ASSERT_EQ(4, y.sizeAt(0));
|
|
|
|
ASSERT_EQ(3, y.sizeAt(1));
|
|
|
|
ASSERT_EQ(2, y.sizeAt(2));
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
ASSERT_EQ(2, x.sizeAt(0));
|
|
|
|
ASSERT_EQ(3, x.sizeAt(1));
|
|
|
|
ASSERT_EQ(4, x.sizeAt(2));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMMulMultiDim) {
|
|
|
|
const int bS=2;
|
|
|
|
const int K=3;
|
|
|
|
const int N=4;
|
|
|
|
|
|
|
|
auto input = NDArrayFactory::create<double>('c', {bS, K, N});
|
|
|
|
auto weights = NDArrayFactory::create<double>('c', {3*K, K});
|
|
|
|
auto expected = NDArrayFactory::create<double>('c', {bS, 3*K, N}, { 38, 44, 50, 56, 83, 98, 113, 128, 128, 152, 176, 200, 173, 206, 239, 272, 218, 260, 302, 344, 263, 314, 365, 416, 308, 368, 428, 488, 353, 422, 491, 560, 398, 476, 554, 632, 110, 116, 122, 128, 263, 278, 293, 308, 416, 440, 464, 488, 569, 602, 635, 668, 722, 764, 806, 848, 875, 926, 977, 1028, 1028, 1088, 1148, 1208, 1181, 1250, 1319, 1388, 1334, 1412, 1490, 1568});
|
|
|
|
|
|
|
|
input.linspace(1);
|
|
|
|
weights.linspace(1);
|
|
|
|
|
|
|
|
auto result = MmulHelper::mmul(&weights, &input, nullptr, 1., 0.);
|
|
|
|
// result must have such shape [bS x 3K x N]
|
|
|
|
|
|
|
|
ASSERT_TRUE(result->isSameShape(&expected));
|
|
|
|
|
|
|
|
//result->printShapeInfo("result shape");
|
2019-11-19 14:39:36 +01:00
|
|
|
// result->printBuffer("result buffer");
|
2019-06-06 14:21:15 +02:00
|
|
|
ASSERT_TRUE(result->equalsTo(&expected));
|
|
|
|
delete result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
TEST_F(NDArrayTest, AdditionOperator1) {
|
|
|
|
|
|
|
|
auto input1 = NDArrayFactory::create<float>('c', {2,2});
|
|
|
|
auto input2 = NDArrayFactory::create<float>('c', {2,2});
|
|
|
|
auto expected = NDArrayFactory::create<float>('c', {2,2});
|
|
|
|
|
|
|
|
input1.assign(1.5);
|
|
|
|
input2.assign(2.);
|
|
|
|
expected.assign(3.5);
|
|
|
|
|
|
|
|
input2 = input1 + input2;
|
|
|
|
|
|
|
|
ASSERT_TRUE(input2.equalsTo(&expected));
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMatmMul_Again_1) {
|
|
|
|
auto a = NDArrayFactory::create<float>('c', {3, 4, 1});
|
|
|
|
auto b = NDArrayFactory::create<float>('c', {3, 1, 5});
|
|
|
|
|
|
|
|
a.linspace(1);
|
|
|
|
b.linspace(1);
|
|
|
|
|
|
|
|
float _expB[] = {1.f, 2.f, 3.f, 4.f, 5.f, 2.f, 4.f, 6.f, 8.f, 10.f, 3.f, 6.f, 9.f, 12.f, 15.f, 4.f, 8.f, 12.f, 16.f, 20.f, 30.f, 35.f, 40.f, 45.f, 50.f, 36.f, 42.f, 48.f, 54.f, 60.f, 42.f, 49.f, 56.f, 63.f, 70.f, 48.f, 56.f, 64.f, 72.f, 80.f, 99.f, 108.f, 117.f, 126.f, 135.f, 110.f, 120.f, 130.f, 140.f, 150.f, 121.f, 132.f, 143.f, 154.f, 165.f, 132.f, 144.f, 156.f, 168.f, 180.f};
|
|
|
|
Nd4jLong _expS[] = {3, 3, 4, 5, 20, 5, 1, 8192, 1, 99};
|
|
|
|
NDArray c(_expB, _expS, nd4j::LaunchContext ::defaultContext(), false);
|
|
|
|
|
|
|
|
auto c_ = MmulHelper::mmul(&a, &b);
|
|
|
|
|
|
|
|
ASSERT_TRUE(c.isSameShape(c_));
|
|
|
|
ASSERT_TRUE(c.equalsTo(c_));
|
|
|
|
|
|
|
|
delete c_;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, TestMatmMul_Again_2) {
|
|
|
|
auto a = NDArrayFactory::create<double>('c', {2, 5, 4});
|
|
|
|
auto b = NDArrayFactory::create<double>('c', {2, 4, 1});
|
|
|
|
|
|
|
|
a.linspace(1);
|
|
|
|
b.linspace(1);
|
|
|
|
|
|
|
|
double _expB[] = {30.f, 70.f, 110.f, 150.f, 190.f, 590.f, 694.f, 798.f, 902.f, 1006.f};
|
|
|
|
Nd4jLong _expS[] = {3, 2, 5, 1, 5, 1, 1, 16384, 1, 99};
|
|
|
|
NDArray c(_expB, _expS);
|
|
|
|
|
|
|
|
auto c_ = MmulHelper::mmul(&a, &b);
|
|
|
|
|
|
|
|
ASSERT_TRUE(c.isSameShape(c_));
|
|
|
|
|
|
|
|
ASSERT_TRUE(c.equalsTo(c_));
|
|
|
|
|
|
|
|
delete c_;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Plus_Test_1)
|
|
|
|
{
|
|
|
|
double expBuff[] = {2., 3, 3., 4., 4., 5, 5., 6., 6., 7, 7., 8.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
//x.printShapeInfo("x shape");
|
|
|
|
//y.printShapeInfo("y shape");
|
|
|
|
//expected.printShapeInfo("e shape");
|
|
|
|
//expected.printIndexedBuffer("e");
|
|
|
|
|
|
|
|
x.linspace(1);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x + y;
|
|
|
|
|
|
|
|
//result.printIndexedBuffer("result");
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Plus_Test_2)
|
|
|
|
{
|
|
|
|
double expBuff[] = {2., 3, 3., 4., 4., 5, 5., 6., 6., 7, 7., 8.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(1);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x + y;
|
|
|
|
// result.printIndexedBuffer();
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Plus_Test_3)
|
|
|
|
{
|
|
|
|
double expBuff[] = {2., 3, 3., 4., 4., 5, 5., 6., 6., 7, 7., 8.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(1);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x + y;
|
|
|
|
// result.printIndexedBuffer();
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Plus_Test_4)
|
|
|
|
{
|
|
|
|
double expBuff[] = {11.,12., 12.,13., 13.,14., 14.,15., 13.,14., 14.,15., 15.,16., 16.,17., 15.,16., 16.,17., 17.,18., 18.,19.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {4, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x + y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_1)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9. ,10., 10.,11., 11.,12., 12.,13., 17.,18., 18.,19., 19.,20., 20.,21., 25.,26., 26.,27., 27.,28., 28.,29.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {4, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_2)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9., 8., 7., 6., 6., 5., 4., 3., 11.,10., 9., 8., 8., 7., 6., 5., 13.,12.,11.,10., 10., 9., 8., 7.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_3)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9., 8., 7., 6., 6., 5., 4., 3., 11.,10., 9., 8., 8., 7., 6., 5., 13.,12.,11.,10., 10., 9., 8., 7.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_4)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9.,10., 8., 9., 11.,12.,10.,11., 13.,14.,12.,13.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_5)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9. ,8 ,10., 9., 11.,10, 12.,11., 13.,12, 14.,13.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Minus_Test_6)
|
|
|
|
{
|
|
|
|
double expBuff[] = {9., 8, 10., 9, 11.,10, 12.,11., 13.,12, 14.,13, 15.,14, 16.,15., 17.,16, 18.,17, 19.,18, 20.,19.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x - y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_1)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10., 11., 24., 26., 42., 45., 64., 68., 18., 19., 40., 42., 66., 69., 96.,100., 26., 27., 56., 58., 90., 93., 128.,132.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {4, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_2)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,20., 30., 40., 55.,66., 77., 88., 12.,24., 36., 48., 65.,78., 91.,104., 14.,28., 42., 56., 75.,90.,105.,120.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_3)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,20., 30., 40.,55.,66., 77., 88., 12.,24., 36., 48.,65.,78., 91.,104., 14.,28., 42., 56.,75.,90.,105.,120.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_4)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,11.,20.,22., 12.,13.,24.,26., 14.,15.,28.,30.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_5)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,20.,11.,22., 12.,24.,13.,26., 14.,28.,15.,30.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Multiply_Test_6)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10,11.,12.,13.,28.,30.,32.,34.,54.,57.,60.,63.};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {3, 1, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 1});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x * y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_1)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10. ,11. , 6. , 6.5 , 4.6666, 5. , 4. , 4.25 , 18. ,19. , 10. ,10.5 , 7.3333, 7.6666, 6. , 6.25 , 26. ,27. , 14. ,14.5 , 10. ,10.3333, 8. , 8.25};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {4, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result,1e-4));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_2)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10. ,5. ,3.333333,2.5 , 2.2,1.83333,1.571428,1.375, 12. ,6. ,4. ,3. , 2.6,2.16666,1.857142,1.625, 14. ,7. ,4.666666,3.5 , 3. ,2.5 ,2.142857,1.875};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_3)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10. ,5. ,3.333333,2.5 , 2.2,1.833333,1.571428,1.375, 12. ,6. ,4. ,3. , 2.6,2.166666,1.857142,1.625, 14. ,7. ,4.666666,3.5 , 3. ,2.5 ,2.142857,1.875};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_4)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,11., 5., 5.5, 12.,13., 6., 6.5, 14.,15., 7., 7.5};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 1, 2});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {2, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_5)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10.,5., 11.,5.5, 12.,6., 13.,6.5, 14.,7., 15.,7.5};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 2});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 2, 2});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_6)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10. , 5.5 , 4. , 3.25 ,14. , 7.5 , 5.333333, 4.25 ,18. , 9.5 , 6.666666, 5.25};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 4, 1});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 1});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Operator_Divide_Test_7)
|
|
|
|
{
|
|
|
|
double expBuff[] = {10., 5. ,3.333333,2.5 ,11., 5.5,3.666666,2.75,12., 6. ,4. ,3. ,13., 6.5,4.333333,3.25, 14., 7. ,4.666666,3.5 ,15., 7.5,5. ,3.75,16., 8. ,5.333333,4. ,17., 8.5,5.666666,4.25, 18., 9. ,6. ,4.5 ,19., 9.5,6.333333,4.75,20.,10. ,6.666666,5. ,21.,10.5,7. ,5.25};
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {3, 4, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 1, 4});
|
|
|
|
auto expected = NDArrayFactory::create<double>(expBuff, 'c', {3, 4, 4});
|
|
|
|
|
|
|
|
x.linspace(10);
|
|
|
|
y.linspace(1);
|
|
|
|
|
|
|
|
auto result = x / y;
|
|
|
|
|
|
|
|
ASSERT_TRUE(expected.isSameShape(&result));
|
|
|
|
ASSERT_TRUE(expected.equalsTo(&result));
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef __CUDABLAS__
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_Lambda_1) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {1, 5}, {1, 2, 3, 4, 5});
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 5}, {4, 5, 6, 7, 8});
|
|
|
|
|
|
|
|
auto lambda = LAMBDA_F(_val) {
|
|
|
|
return _val + 3.0f;
|
|
|
|
};
|
|
|
|
|
|
|
|
x.applyLambda<float>(lambda);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&x));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_Lambda_2) {
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {1, 5}, {1, 2, 1, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {1, 5}, {1, 2, 1, 2, 1});
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 5}, {3, 5, 3, 5, 3});
|
|
|
|
|
|
|
|
auto lambda = LAMBDA_FF(_x, _y) {
|
|
|
|
return _x + _y + 1.0f;
|
|
|
|
};
|
|
|
|
|
|
|
|
x.applyPairwiseLambda<float>(&y, lambda);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&x));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_Lambda_3) {
|
|
|
|
auto x = NDArrayFactory::create<double>('c', {1, 5}, {1, 2, 1, 2, 1});
|
|
|
|
auto y = NDArrayFactory::create<double>('c', {1, 5}, {1, 2, 1, 2, 1});
|
|
|
|
auto exp = NDArrayFactory::create<double>('c', {1, 5}, {4, 8, 4, 8, 4});
|
|
|
|
|
|
|
|
auto lambda = LAMBDA_DD(_x, _y) {
|
|
|
|
return (_x + _y) * 2;
|
|
|
|
};
|
|
|
|
|
|
|
|
x.applyPairwiseLambda<double>(&y, lambda);
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.equalsTo(&x));
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_swapUnsafe_1) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2}, {1, 2, 3, 4});
|
|
|
|
auto y = NDArrayFactory::create<float>('c', {1, 4}, {5, 6, 7, 8});
|
|
|
|
auto expX = NDArrayFactory::create<float>('c', {2, 2}, {5, 6, 7, 8});
|
|
|
|
auto expY = NDArrayFactory::create<float>('c', {1, 4}, {1, 2, 3, 4});
|
|
|
|
|
|
|
|
x.swapUnsafe(y);
|
|
|
|
|
|
|
|
ASSERT_TRUE(expX.equalsTo(&x));
|
|
|
|
ASSERT_TRUE(expY.equalsTo(&y));
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_1) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 3}, {1, 2, 3, 4, 5, 6});
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {2, 1}, {1, 5});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_2) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 3});
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {2, 1}, {1, 5});
|
|
|
|
x.linspace(1);
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_3) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 2}, {1, 4});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_4) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {1, 2}, {1, 4});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_5) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 2}, {1, 8});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_6) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {1, 2}, {1, 8});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_7) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {2, 1}, {1, 8});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_8) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 3});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 2}, {1, 5});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_9) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {2, 1}, {1, 4});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_10) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {2, 2});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {2, 1}, {1, 4});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_11) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {3, 3});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {3, 1}, {1, 5, 9});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_12) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {3, 3});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 3}, {1, 5, 9});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_13) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {3, 3, 4});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {3, 1}, {1,18,35});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_14) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {3, 3, 4});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 3}, {1,18,35});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_15) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {3, 3, 4});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {1, 3}, {1,18,35});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_16) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {1, 5});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {1, 1}, {1});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('c');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_17) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('c', {5, 1});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('c', {1, 1}, {1});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, Test_diagonal_18) {
|
|
|
|
|
|
|
|
auto x = NDArrayFactory::create<float>('f', {1, 1});
|
|
|
|
x.linspace(1);
|
|
|
|
auto exp = NDArrayFactory::create<float>('f', {1, 1}, {1});
|
|
|
|
|
|
|
|
auto diag = x.diagonal('r');
|
|
|
|
|
|
|
|
ASSERT_TRUE(exp.isSameShape(diag));
|
|
|
|
ASSERT_TRUE(exp.equalsTo(diag));
|
|
|
|
|
|
|
|
delete diag;
|
|
|
|
}
|
|
|
|
|
|
|
|
//////////////////////////////////////////////////////////////////////
|
|
|
|
TEST_F(NDArrayTest, assign_test1) {
|
|
|
|
|
|
|
|
NDArray x('c', {2, 3}, {1,2,3,4,5,6});
|
|
|
|
NDArray y('c', {2, 3}, {10,20,30,40,50,60});
|
|
|
|
y.reshapei('c',{3, 2});
|
|
|
|
|
|
|
|
x.assign(y);
|
|
|
|
x.reshapei('c',{3, 2});
|
|
|
|
ASSERT_TRUE(x.equalsTo(y));
|
|
|
|
}
|