2019-06-06 14:21:15 +02:00
|
|
|
/*******************************************************************************
|
|
|
|
* Copyright (c) 2015-2018 Skymind, Inc.
|
|
|
|
*
|
|
|
|
* This program and the accompanying materials are made available under the
|
|
|
|
* terms of the Apache License, Version 2.0 which is available at
|
|
|
|
* https://www.apache.org/licenses/LICENSE-2.0.
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
|
|
* License for the specific language governing permissions and limitations
|
|
|
|
* under the License.
|
|
|
|
*
|
|
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
|
|
******************************************************************************/
|
|
|
|
|
|
|
|
//
|
|
|
|
// Created by agibsonccc on 2/21/16.
|
|
|
|
//
|
|
|
|
|
|
|
|
#ifndef NATIVEOPERATIONS_NATIVEOPS_H
|
|
|
|
#define NATIVEOPERATIONS_NATIVEOPS_H
|
|
|
|
|
|
|
|
/*
|
|
|
|
#ifndef thread_local
|
|
|
|
# if __STDC_VERSION__ >= 201112 && !defined __STDC_NO_THREADS__
|
|
|
|
# define thread_local _Thread_local
|
|
|
|
# elif defined _WIN32 && ( \
|
|
|
|
defined _MSC_VER || \
|
|
|
|
defined __ICL || \
|
|
|
|
defined __DMC__ || \
|
|
|
|
defined __BORLANDC__ )
|
|
|
|
# define thread_local __declspec(thread)
|
|
|
|
// note that ICC (linux) and Clang are covered by __GNUC__
|
|
|
|
# elif defined __GNUC__ || \
|
|
|
|
defined __SUNPRO_C || \
|
|
|
|
defined __xlC__
|
|
|
|
# define thread_local __thread
|
|
|
|
# else
|
|
|
|
# error "Cannot define thread_local"
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
*/
|
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
#include <system/pointercast.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
#include <types/float16.h>
|
2020-03-02 10:49:41 +01:00
|
|
|
#include <cnpy/cnpy.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
//DO NOT REMOVE: THIS IS AN EDITOR SEMANTICS THING FOR CLION
|
|
|
|
//IT DEFINES THE EXPORT MACRO FOR THE EDITOR AND THEN
|
|
|
|
//RE ADDS THE DEFINITION VIA dll.h
|
|
|
|
#ifdef _WIN32
|
|
|
|
#define ND4J_EXPORT __declspec(dllexport)
|
|
|
|
#else
|
|
|
|
#define ND4J_EXPORT
|
|
|
|
#endif
|
2020-03-02 10:49:41 +01:00
|
|
|
#include <system/dll.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
int tad_threshold = 1;
|
|
|
|
int element_threshold = 32;
|
|
|
|
|
|
|
|
bool debug = false;
|
|
|
|
bool verbose = false;
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <array/ShapeList.h>
|
|
|
|
#include <array/ConstantDescriptor.h>
|
2019-07-05 14:15:09 +02:00
|
|
|
#include <helpers/ConstantShapeHelper.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
#include <array/ConstantDataBuffer.h>
|
2020-01-04 11:27:50 +01:00
|
|
|
#include <array/InteropDataBuffer.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
#include <helpers/ConstantHelper.h>
|
|
|
|
#include <array/TadPack.h>
|
|
|
|
#include <graph/VariablesSet.h>
|
|
|
|
#include <graph/GraphState.h>
|
|
|
|
#include <graph/execution/LogicExecutor.h>
|
|
|
|
#include <graph/ResultWrapper.h>
|
2020-03-02 10:49:41 +01:00
|
|
|
#include <helpers/DebugInfo.h>
|
2020-01-24 08:11:09 +01:00
|
|
|
#include <memory/MemoryCounter.h>
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::InteropDataBuffer OpaqueDataBuffer;
|
2020-01-04 11:27:50 +01:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
extern "C" {
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-08-26 18:57:51 +02:00
|
|
|
/**
|
|
|
|
* This function returns last error code stored,
|
|
|
|
* @return non-zero if something bad happened
|
|
|
|
*/
|
|
|
|
ND4J_EXPORT int lastErrorCode();
|
|
|
|
|
|
|
|
/**
|
|
|
|
* This function returns last error message, if last error code > 0
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
ND4J_EXPORT const char* lastErrorMessage();
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param p
|
|
|
|
* @param len
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void tryPointer(Nd4jPointer extra, Nd4jPointer p, int len);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param num
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void setElementThreshold(int num);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param num
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void setTADThreshold(int num);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execIndexReduceScalar(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfoBuffer
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execIndexReduce(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param y
|
|
|
|
* @param yShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execBroadcast(
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *extraPointers,
|
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execBroadcastBool(
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *extraPointers,
|
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-11-21 13:43:03 +01:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param dx
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param y
|
|
|
|
* @param yShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param n
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execPairwiseTransform(
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *extraPointers,
|
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execPairwiseTransformBool(
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *extraPointers,
|
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceFloat(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceSame(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-06-06 14:21:15 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceBool(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-06-06 14:21:15 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceLong(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-06-06 14:21:15 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceFloat2(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceSame2(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceBool2(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduceLong2(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParamsVals
|
|
|
|
* @param y
|
|
|
|
* @param yShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduce3(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParamsVals,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParamsVals
|
|
|
|
* @param y
|
|
|
|
* @param yShapeInfo
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduce3Scalar(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParamsVals,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo);
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParamsVals
|
|
|
|
* @param y
|
|
|
|
* @param yShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfoBuffer
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduce3Tad(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParamsVals,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape,
|
|
|
|
Nd4jLong const* tadOnlyShapeInfo, Nd4jLong const* tadOffsets,
|
|
|
|
Nd4jLong const* yTadOnlyShapeInfo, Nd4jLong const* yTadOffsets);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execReduce3All(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParamsVals,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeInfo, Nd4jLong const* dYShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape,
|
|
|
|
Nd4jLong const* xTadShapeInfo, Nd4jLong const* xOffsets,
|
|
|
|
Nd4jLong const* yTadShapeInfo, Nd4jLong const* yOffsets);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
* @param scalar
|
|
|
|
* @param extraParams
|
|
|
|
* @param n
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execScalar(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbScalar, Nd4jLong const* hSscalarShapeInfo, Nd4jLong const* dSscalarShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execScalarBool(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbScalar, Nd4jLong const* hSscalarShapeInfo, Nd4jLong const* dSscalarShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execSummaryStatsScalar(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool biasCorrected);
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execSummaryStats(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool biasCorrected);
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfoBuffer
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execSummaryStatsTad(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool biasCorrected,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jLong const* tadShapeInfo, Nd4jLong const* tadOffsets);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param opNum
|
|
|
|
* @param dx
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param result
|
|
|
|
* @param resultShapeInfo
|
|
|
|
* @param extraParams
|
|
|
|
* @param n
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execTransformFloat(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execTransformSame(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execTransformBool(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execTransformAny(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execTransformStrict(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
2019-06-06 14:21:15 +02:00
|
|
|
void *extraParams);
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param opNum
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param z
|
|
|
|
* @param zShapeInfo
|
|
|
|
* @param scalars
|
|
|
|
* @param extraParams
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execScalarTad(Nd4jPointer *extraPointers,
|
2019-06-06 14:21:15 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbScalars, Nd4jLong const* hScalarShapeInfo, Nd4jLong const* dScalarShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape,
|
|
|
|
Nd4jLong const* tadShapeInfo, Nd4jLong const* tadOffsets,
|
|
|
|
Nd4jLong const* tadShapeInfoZ, Nd4jLong const* tadOffsetsZ);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execScalarBoolTad(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeInfo, Nd4jLong const* dXShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeInfo, Nd4jLong const* dZShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbScalars, Nd4jLong const* hScalarShapeInfo, Nd4jLong const* dScalarShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraParams,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbDimension, Nd4jLong const* hDimensionShape, Nd4jLong const* dDimensionShape,
|
|
|
|
Nd4jLong const* tadShapeInfo, Nd4jLong const* tadOffsets,
|
|
|
|
Nd4jLong const* tadShapeInfoZ, Nd4jLong const* tadOffsetsZ);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void specialConcat (
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *extraPointers,
|
|
|
|
int dimension,
|
|
|
|
int numArrays,
|
|
|
|
Nd4jPointer *data,
|
|
|
|
Nd4jPointer *inputShapeInfo,
|
|
|
|
void *result,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jLong const* resultShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *tadPointers,
|
|
|
|
Nd4jPointer *offsetPointers);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* This method implementation exists only for cuda.
|
|
|
|
* The other backends should have dummy method for JNI compatibility reasons.
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void initializeDevicesAndFunctions();
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void initializeFunctions(Nd4jPointer *functions);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* This method acquires memory chunk of requested size on host side
|
|
|
|
*
|
|
|
|
* @param pointer pointer that'll be used for allocation
|
|
|
|
* @param memorySize memory size, in bytes
|
|
|
|
* @param flags optional parameter
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer mallocHost(Nd4jLong memorySize, int flags);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* This method acquires memory chunk of requested size on specified device
|
|
|
|
*
|
|
|
|
* @param pointer pointer that'll be used for allocation
|
|
|
|
* @param memorySize memory size, in bytes
|
|
|
|
* @param ptrToDeviceId pointer to deviceId. For cuda that's just and int, for OpenCL that's pointer to device_id, etc
|
|
|
|
* @param flags optional parameter
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer mallocDevice(Nd4jLong memorySize, int deviceId, int flags);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* This method releases previously allocated host memory space
|
|
|
|
*
|
|
|
|
* @param pointer pointer that'll be freed
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int freeHost(Nd4jPointer pointer);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* This method releases previously allocated memory space on device
|
|
|
|
*
|
|
|
|
* @param pointer pointer that'll be freed
|
|
|
|
* @param ptrToDeviceId pointer to deviceId.
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int freeDevice(Nd4jPointer pointer, int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int ompGetMaxThreads();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int ompGetNumThreads();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param threads
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void setOmpNumThreads(int threads);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param threads
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void setOmpMinThreads(int threads);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-09-04 13:41:08 +02:00
|
|
|
ND4J_EXPORT bool isBlasVersionMatches(int major, int minor, int build);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
2019-07-22 13:34:08 +02:00
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer createContext();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer createStream();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer createEvent();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
2019-07-22 13:34:08 +02:00
|
|
|
*
|
|
|
|
* @param event
|
|
|
|
* @param stream
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int registerEvent(Nd4jPointer event, Nd4jPointer stream);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param event
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int destroyEvent(Nd4jPointer event);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param ptrToDeviceId
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int setDevice(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int getDevice();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param stream
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int streamSynchronize(Nd4jPointer stream);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param event
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int eventSynchronize(Nd4jPointer event);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
2019-07-22 13:34:08 +02:00
|
|
|
* @param ptrToDeviceId
|
2019-06-06 14:21:15 +02:00
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getDeviceFreeMemory(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
* Returns amount of free memory for current device
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getDeviceFreeMemoryDefault();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
2019-07-22 13:34:08 +02:00
|
|
|
*
|
|
|
|
* @param ptrToDeviceId
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getDeviceTotalMemory(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param ptrToDeviceId
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int getDeviceMajor(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
* This method returns amount of cached memory
|
|
|
|
* @param deviceId
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getCachedMemory(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param ptrToDeviceId
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int getDeviceMinor(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
2019-07-22 13:34:08 +02:00
|
|
|
* @param ptrToDeviceId
|
2019-06-06 14:21:15 +02:00
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT const char * getDeviceName(int deviceId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
2019-07-22 13:34:08 +02:00
|
|
|
* @param dst
|
|
|
|
* @param src
|
|
|
|
* @param size
|
|
|
|
* @param flags
|
|
|
|
* @param reserved
|
2019-06-06 14:21:15 +02:00
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int memcpySync(Nd4jPointer dst,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer src,
|
|
|
|
Nd4jLong size,
|
|
|
|
int flags,
|
|
|
|
Nd4jPointer reserved);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
2019-07-22 13:34:08 +02:00
|
|
|
* @param dst
|
|
|
|
* @param src
|
|
|
|
* @param size
|
|
|
|
* @param flags
|
|
|
|
* @param reserved
|
2019-06-06 14:21:15 +02:00
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int memcpyAsync(Nd4jPointer dst,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer src,
|
|
|
|
Nd4jLong size,
|
|
|
|
int flags,
|
|
|
|
Nd4jPointer reserved);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
/**
|
2019-07-22 13:34:08 +02:00
|
|
|
*
|
|
|
|
* @param dst
|
|
|
|
* @param value
|
|
|
|
* @param size
|
|
|
|
* @param flags
|
|
|
|
* @param reserved
|
2019-06-06 14:21:15 +02:00
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int memsetSync(Nd4jPointer dst,
|
2019-07-22 13:34:08 +02:00
|
|
|
int value,
|
|
|
|
Nd4jLong size,
|
|
|
|
int flags,
|
|
|
|
Nd4jPointer reserved);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param dst
|
|
|
|
* @param value
|
|
|
|
* @param size
|
|
|
|
* @param flags
|
|
|
|
* @param reserved
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int memsetAsync(Nd4jPointer dst,
|
2019-07-22 13:34:08 +02:00
|
|
|
int value,
|
|
|
|
Nd4jLong size,
|
|
|
|
int flags,
|
|
|
|
Nd4jPointer reserved);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param dst
|
|
|
|
* @param src
|
|
|
|
* @param size
|
|
|
|
* @param flags
|
|
|
|
* @param reserved
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int memcpyConstantAsync(Nd4jLong dst,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer src,
|
|
|
|
Nd4jLong size,
|
|
|
|
int flags,
|
|
|
|
Nd4jPointer reserved);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer getConstantSpace();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int getAvailableDevices();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param reallyEnable
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void enableDebugMode(bool reallyEnable);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param reallyEnable
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void enableVerboseMode(bool reallyEnable);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param gridSize
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void setGridLimit(int gridSize);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::TadPack OpaqueTadPack;
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param dimension
|
|
|
|
* @param dimensionLength
|
|
|
|
* @param targetBuffer
|
|
|
|
* @param offsetsBuffer
|
|
|
|
*/
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT OpaqueTadPack* tadOnlyShapeInfo(Nd4jLong const*xShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int *dimension,
|
|
|
|
int dimensionLength);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT Nd4jLong const* getPrimaryShapeInfo(OpaqueTadPack* pack);
|
|
|
|
ND4J_EXPORT Nd4jLong const* getPrimaryOffsets(OpaqueTadPack* pack);
|
|
|
|
ND4J_EXPORT Nd4jLong const* getSpecialShapeInfo(OpaqueTadPack* pack);
|
|
|
|
ND4J_EXPORT Nd4jLong const* getSpecialOffsets(OpaqueTadPack* pack);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getNumberOfTads(OpaqueTadPack* pack);
|
|
|
|
ND4J_EXPORT int getShapeInfoLength(OpaqueTadPack* pack);
|
|
|
|
|
|
|
|
ND4J_EXPORT void deleteTadPack(OpaqueTadPack* ptr);
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/*
|
|
|
|
* PullRow special op
|
|
|
|
*/
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param z
|
|
|
|
* @param zShapeInfo
|
|
|
|
* @param n
|
|
|
|
* @param indexes
|
|
|
|
* @param tadShapeInfo
|
|
|
|
* @param tadOffsets
|
|
|
|
* @param zTadShapeInfo
|
|
|
|
* @param zTadOffsets
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void pullRows(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* xShapeInfo, Nd4jLong const* dxShapeInfo,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* zShapeInfo, Nd4jLong const* dzShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jLong n,
|
|
|
|
Nd4jLong *indexes,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jLong const* tadShapeInfo,
|
|
|
|
Nd4jLong const* tadOffsets,
|
|
|
|
Nd4jLong const* zTadShapeInfo,
|
|
|
|
Nd4jLong const* zTadOffsets);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extras
|
|
|
|
* @param dx
|
|
|
|
* @param dz
|
|
|
|
* @param n
|
|
|
|
* @param length
|
|
|
|
* @param propagate
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void average(Nd4jPointer *extras,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jPointer *x, Nd4jLong const* xShapeInfo,
|
|
|
|
Nd4jPointer *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *z, Nd4jLong const* zShapeInfo,
|
|
|
|
void *dz, Nd4jLong const* dzShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int n,
|
|
|
|
Nd4jLong length,
|
|
|
|
bool propagate);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void accumulate(Nd4jPointer *extras,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jPointer *x, Nd4jLong const* xShapeInfo,
|
|
|
|
Nd4jPointer *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *z, Nd4jLong const* zShapeInfo,
|
|
|
|
void *dz, Nd4jLong const* dzShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int n,
|
|
|
|
Nd4jLong length);
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* P2P enabler
|
|
|
|
*/
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param enable
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void enableP2P(bool enable);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void checkP2P();
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT bool isP2PAvailable();
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Shuffle methods
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extras
|
|
|
|
* @param dx
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param dz
|
|
|
|
* @param zShapeInfo
|
|
|
|
* @param N
|
|
|
|
* @param shuffleMap
|
|
|
|
* @param tadShapeInfo
|
|
|
|
* @param tadOffsets
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void shuffle(Nd4jPointer *extras,
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jPointer *x, Nd4jPointer *xShapeInfo,
|
|
|
|
Nd4jPointer *dx, Nd4jPointer *dxShapeInfo,
|
|
|
|
Nd4jPointer *z, Nd4jPointer *zShapeInfo,
|
|
|
|
Nd4jPointer *dz, Nd4jPointer *dzShapeInfo,
|
|
|
|
int N,
|
|
|
|
int *shuffleMap,
|
|
|
|
Nd4jPointer *tadShapeInfo,
|
|
|
|
Nd4jPointer *tadOffsets);
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Type Conversions
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extras
|
|
|
|
* @param srcType
|
|
|
|
* @param x
|
|
|
|
* @param N
|
|
|
|
* @param dstType
|
|
|
|
* @param z
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void convertTypes(Nd4jPointer *extras, int srcType, Nd4jPointer x, Nd4jLong N, int dstType, Nd4jPointer z);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT bool isExperimentalEnabled();
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Aggregate
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param opNum
|
|
|
|
* @param arguments
|
|
|
|
* @param numArguments
|
|
|
|
* @param shapeArguments
|
|
|
|
* @param numShapeArguments
|
|
|
|
* @param indexArguments
|
|
|
|
* @param numIndexArguments
|
|
|
|
* @param intArrays
|
|
|
|
* @param numIntArrays
|
|
|
|
* @param realArguments
|
|
|
|
* @param numRealArguments
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execAggregate(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
|
|
|
void **arguments,
|
|
|
|
int numArguments,
|
|
|
|
Nd4jLong **shapeArguments,
|
|
|
|
int numShapeArguments,
|
|
|
|
int *indexArguments,
|
|
|
|
int numIndexArguments,
|
|
|
|
int **intArrays,
|
|
|
|
int numIntArrays,
|
|
|
|
void *realArguments,
|
|
|
|
int numRealArguments,
|
2020-03-02 10:49:41 +01:00
|
|
|
sd::DataType dtype);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void batchExecutor(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int numAggregates,
|
|
|
|
int opNum,
|
|
|
|
int maxArgs,
|
|
|
|
int maxShapes,
|
|
|
|
int maxIntArrays,
|
|
|
|
int maxIntArraySize,
|
|
|
|
int maxIdx,
|
|
|
|
int maxReals,
|
|
|
|
void *ptrToArguments,
|
2020-03-02 10:49:41 +01:00
|
|
|
sd::DataType dtype);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execAggregateBatch(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int numAggregates,
|
|
|
|
int opNum,
|
|
|
|
int maxArgs,
|
|
|
|
int maxShapes,
|
|
|
|
int maxIntArrays,
|
|
|
|
int maxIntArraySize,
|
|
|
|
int maxIdx,
|
|
|
|
int maxReals,
|
|
|
|
void *ptrToArguments,
|
2020-03-02 10:49:41 +01:00
|
|
|
sd::DataType dtype);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Random operations
|
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param opNum
|
|
|
|
* @param state
|
|
|
|
* @param z
|
|
|
|
* @param zShapeBuffer
|
|
|
|
* @param extraArguments
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execRandom(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
|
|
|
Nd4jPointer state,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeBuffer, Nd4jLong const* dZShapeBuffer,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraArguments);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param opNum
|
|
|
|
* @param state
|
|
|
|
* @param x
|
|
|
|
* @param xShapeBuffer
|
|
|
|
* @param y
|
|
|
|
* @param yShapeBuffer
|
|
|
|
* @param z
|
|
|
|
* @param zShapeBuffer
|
|
|
|
* @param extraArguments
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execRandom3(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
|
|
|
Nd4jPointer state,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeBuffer, Nd4jLong const* dXShapeBuffer,
|
|
|
|
OpaqueDataBuffer *dbY, Nd4jLong const* hYShapeBuffer, Nd4jLong const* dYShapeBuffer,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeBuffer, Nd4jLong const* dZShapeBuffer,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraArguments);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param opNum
|
|
|
|
* @param state
|
|
|
|
* @param x
|
|
|
|
* @param xShapeBuffer
|
|
|
|
* @param z
|
|
|
|
* @param zShapeBuffer
|
|
|
|
* @param extraArguments
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void execRandom2(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
int opNum,
|
|
|
|
Nd4jPointer state,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* hXShapeBuffer, Nd4jLong const* dXShapeBuffer,
|
|
|
|
OpaqueDataBuffer *dbZ, Nd4jLong const* hZShapeBuffer, Nd4jLong const* dZShapeBuffer,
|
2019-07-22 13:34:08 +02:00
|
|
|
void *extraArguments);
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param seed
|
|
|
|
* @param bufferSize
|
|
|
|
* @param ptrToBuffer
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer initRandom(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
long seed,
|
|
|
|
long bufferSize,
|
|
|
|
Nd4jPointer ptrToBuffer);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param seed
|
|
|
|
* @param ptrRandom
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void refreshBuffer(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
long seed,
|
|
|
|
Nd4jPointer ptrRandom);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param extraPointers
|
|
|
|
* @param seed
|
|
|
|
* @param ptrRandom
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void reSeedBuffer(Nd4jPointer *extraPointers,
|
2019-07-22 13:34:08 +02:00
|
|
|
long seed,
|
|
|
|
Nd4jPointer ptrRandom);
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param ptrRandom
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void destroyRandom(Nd4jPointer ptrRandom);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param data
|
|
|
|
* @param shapeBuffer
|
|
|
|
* @param wordSize
|
|
|
|
* @param headerSize
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
|
|
|
|
template <typename T>
|
2020-05-09 07:06:14 +02:00
|
|
|
static Nd4jPointer _numpyHeaderForNd4j(Nd4jPointer data,const Nd4jPointer shapeBuffer,Nd4jLong wordSize,Nd4jLong* headerSize) {
|
|
|
|
Nd4jLong const* shapeBufferCast = reinterpret_cast<const Nd4jLong *>(shapeBuffer);
|
2019-07-22 13:34:08 +02:00
|
|
|
int rank = shape::rank(shapeBufferCast);
|
2020-05-09 07:06:14 +02:00
|
|
|
const Nd4jLong* shape = shape::shapeOf(shapeBufferCast);
|
|
|
|
unsigned int* npShape = new unsigned int[rank];
|
2019-07-22 13:34:08 +02:00
|
|
|
for(int i = 0; i < rank; i++) {
|
|
|
|
npShape[i] = shape[i];
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jLong length = shape::prodLong(shape,rank);
|
|
|
|
auto npHeader = cnpy::createNpyHeader<T>(data,npShape,rank,wordSize);
|
|
|
|
char *ret = new char[npHeader.size() + 1];
|
|
|
|
int count = 0;
|
|
|
|
for(int i = 0; i < npHeader.size(); i++) {
|
|
|
|
ret[count] = npHeader[i];
|
|
|
|
count++;
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
ret[count] = '\0';
|
|
|
|
count++;
|
|
|
|
|
|
|
|
*headerSize = count;
|
|
|
|
return reinterpret_cast<Nd4jPointer>(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
extern "C" {
|
|
|
|
|
2020-05-09 07:06:14 +02:00
|
|
|
static Nd4jPointer numpyHeaderForNd4j(Nd4jPointer data,Nd4jPointer shapeBuffer,Nd4jLong wordSize,Nd4jLong* headerSize) {
|
2019-07-22 13:34:08 +02:00
|
|
|
auto shapeBufferCast = reinterpret_cast<Nd4jLong *>(shapeBuffer);
|
2020-03-02 10:49:41 +01:00
|
|
|
auto type = sd::ArrayOptions::dataType(shapeBufferCast);
|
2019-07-22 13:34:08 +02:00
|
|
|
BUILD_SINGLE_SELECTOR(type, return _numpyHeaderForNd4j, (data, shapeBuffer, wordSize, headerSize), LIBND4J_TYPES);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Load numpy from a header
|
|
|
|
* based on the cnpy parse from header method.
|
|
|
|
* @param data the header data to parse
|
|
|
|
* @return a pointer to a numpy cnpy:NpyArray struct
|
|
|
|
*/
|
|
|
|
static Nd4jPointer loadNpyFromHeader(Nd4jPointer data) {
|
|
|
|
char *header = reinterpret_cast<char *>(data);
|
|
|
|
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromHeader(header);
|
|
|
|
cnpy::NpyArray *ret = new cnpy::NpyArray();
|
|
|
|
int totalLengthOfShape = 1;
|
|
|
|
for(int i = 0; i < arr.shape.size(); i++) {
|
|
|
|
totalLengthOfShape *= arr.shape[i];
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
ret->data = arr.data;
|
|
|
|
ret->wordSize = arr.wordSize;
|
|
|
|
ret->shape = arr.shape;
|
|
|
|
return reinterpret_cast<Nd4jPointer>(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Create a numpy array from an nd4j
|
|
|
|
* array
|
|
|
|
* @param data a pointer to the data
|
|
|
|
* @param shapeBuffer the shapebuffer for the nd4j array
|
|
|
|
* @param wordSize the word size (4 for float, 8 for doubles)
|
|
|
|
* @return a pointer to a numpy array
|
|
|
|
*/
|
|
|
|
|
|
|
|
template <typename T>
|
|
|
|
static Nd4jPointer _numpyFromNd4j(Nd4jPointer data,Nd4jPointer shapeBuffer,Nd4jLong wordSize) {
|
|
|
|
Nd4jLong *shapeBufferCast = reinterpret_cast<Nd4jLong *>(shapeBuffer);
|
|
|
|
int rank = shape::rank(shapeBufferCast);
|
|
|
|
Nd4jLong *shape = shape::shapeOf(shapeBufferCast);
|
|
|
|
unsigned int *npShape = new unsigned int[rank];
|
|
|
|
for(int i = 0; i < rank; i++) {
|
|
|
|
npShape[i] = shape[i];
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
Nd4jLong length = shape::prodLong(shape,rank);
|
|
|
|
auto npHeader = cnpy::createNpyHeader<T>(data,npShape,rank,wordSize);
|
|
|
|
char *dataChar = reinterpret_cast<char *>(data);
|
|
|
|
char *npHeaderData = npHeader.data();
|
|
|
|
char *ret = new char[(wordSize * length) + npHeader.size()];
|
|
|
|
char *cursorStart = ret;
|
|
|
|
std::memcpy(reinterpret_cast<void *>(ret), reinterpret_cast<void *>(npHeaderData), npHeader.size() * sizeof(Nd4jLong));
|
|
|
|
//move to next
|
|
|
|
cursorStart += npHeader.size();
|
|
|
|
std::memcpy(reinterpret_cast<void *>(ret), reinterpret_cast<void *>(dataChar), length * wordSize * sizeof(Nd4jLong));
|
|
|
|
Nd4jPointer rettPointer = reinterpret_cast<Nd4jPointer>(ret);
|
|
|
|
return rettPointer;
|
|
|
|
}
|
|
|
|
|
|
|
|
extern "C" {
|
|
|
|
|
|
|
|
static Nd4jPointer numpyFromNd4j(Nd4jPointer data,Nd4jPointer shapeBuffer,Nd4jLong wordSize) {
|
|
|
|
auto shapeBufferCast = reinterpret_cast<Nd4jLong *>(shapeBuffer);
|
2020-03-02 10:49:41 +01:00
|
|
|
auto type = sd::ArrayOptions::dataType(shapeBufferCast);
|
2019-07-22 13:34:08 +02:00
|
|
|
BUILD_SINGLE_SELECTOR(type, return _numpyFromNd4j, (data, shapeBuffer, wordSize), LIBND4J_TYPES);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param npyArray
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer shapeBufferForNumpy(Nd4jPointer npyArray);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Get the shape buffer from a
|
|
|
|
* numpy array.
|
|
|
|
* **Warning** this allocates memory
|
|
|
|
* @param npyArray
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static Nd4jPointer shapeBufferForNumpyHeader(Nd4jPointer npyArray) {
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromHeader(reinterpret_cast<char *>(npyArray));
|
|
|
|
auto shape = new unsigned int[arr.shape.size()];
|
|
|
|
for(unsigned int i = 0; i < arr.shape.size(); i++) {
|
|
|
|
shape[i] = arr.shape[i];
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
auto shapeBuffer = shape::shapeBufferOfNpy(arr.shape.size(), shape, arr.fortranOrder);
|
|
|
|
delete[] shape;
|
|
|
|
return reinterpret_cast<Nd4jPointer>(shapeBuffer);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param npyArray
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static Nd4jPointer dataPointForNumpyHeader(Nd4jPointer npyArray) {
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromHeader(reinterpret_cast<char *>(npyArray));
|
|
|
|
unsigned char *dataToPrint = reinterpret_cast<unsigned char *>(arr.data);
|
|
|
|
return dataToPrint;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param npyArray
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static Nd4jPointer dataPointForNumpyStruct(Nd4jPointer npyArrayStruct) {
|
|
|
|
cnpy::NpyArray *arrPointer = reinterpret_cast<cnpy::NpyArray *>(npyArrayStruct);
|
|
|
|
unsigned char *dataToPrint = reinterpret_cast<unsigned char *>(arrPointer->data);
|
|
|
|
return reinterpret_cast<Nd4jPointer>(dataToPrint);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param npyArray
|
|
|
|
* @param fromFile
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static Nd4jPointer dataPointForNumpy(Nd4jPointer npyArray) {
|
|
|
|
char *npyArrayBuffer = reinterpret_cast< char *>(npyArray);
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromPointer(npyArrayBuffer);
|
|
|
|
return dataPointForNumpyStruct(reinterpret_cast<Nd4jPointer>(&arr));
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Load a numpy array from a file
|
|
|
|
* and return it as an Nd4jPointer
|
|
|
|
* @param path
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static Nd4jPointer numpyFromFile(std::string path) {
|
|
|
|
char *numpyBuffer = cnpy::loadFile(path.data());
|
|
|
|
return reinterpret_cast<Nd4jPointer >(numpyBuffer);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
////// NPZ //////
|
|
|
|
|
|
|
|
static void* mapFromNpzFile(std::string path){
|
|
|
|
cnpy::npz_t* mapPtr = new cnpy::npz_t();
|
|
|
|
cnpy::npz_t map = cnpy::npzLoad(path);
|
|
|
|
mapPtr->insert(map.begin(), map.end());
|
|
|
|
return reinterpret_cast<void*>(mapPtr);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int getNumNpyArraysInMap(void *map){
|
|
|
|
cnpy::npz_t* arrays = reinterpret_cast<cnpy::npz_t*>(map);
|
|
|
|
int n = arrays->size();
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char* getNpyArrayNameFromMap(void *map, int index){
|
|
|
|
cnpy::npz_t* arrays = reinterpret_cast<cnpy::npz_t*>(map);
|
|
|
|
cnpy::npz_t::iterator it = arrays->begin();
|
|
|
|
cnpy::npz_t::iterator end = arrays->end();
|
|
|
|
int cnt = 0;
|
|
|
|
for(; it != end; ++it, ++cnt){
|
|
|
|
if (cnt == index){
|
|
|
|
// FIXME: @fariz, this is a leak!
|
2019-11-13 15:04:59 +01:00
|
|
|
#ifdef _MSC_VER
|
|
|
|
return const_cast<const char *>(_strdup(it->first.c_str()));
|
|
|
|
#else
|
2019-07-22 13:34:08 +02:00
|
|
|
return const_cast<const char *>(strdup(it->first.c_str()));
|
2019-11-13 15:04:59 +01:00
|
|
|
#endif
|
2019-07-22 13:34:08 +02:00
|
|
|
}
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
2019-07-22 13:34:08 +02:00
|
|
|
throw std::runtime_error("No array at index.");
|
|
|
|
}
|
|
|
|
|
|
|
|
static void* getNpyArrayFromMap(void *map, int index){
|
|
|
|
cnpy::npz_t* arrays = reinterpret_cast<cnpy::npz_t*>(map);
|
|
|
|
cnpy::npz_t::iterator it = arrays->begin();
|
|
|
|
cnpy::npz_t::iterator end = arrays->end();
|
|
|
|
cnpy::NpyArray *arr = new cnpy::NpyArray();
|
|
|
|
int cnt = 0;
|
|
|
|
for(; it != end; ++it, ++cnt){
|
|
|
|
if (cnt == index){
|
|
|
|
*arr = it->second;
|
|
|
|
return arr;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
throw std::runtime_error("No array at index.");
|
|
|
|
}
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int dataTypeFromNpyHeader(void *header);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
static void* getNpyArrayData(void *npArray){
|
|
|
|
cnpy::NpyArray* npyArray2 = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
return reinterpret_cast<void*>(npyArray2->data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int getNpyArrayRank(void *npArray){
|
|
|
|
cnpy::NpyArray* arr = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
int rank = arr->shape.size();
|
|
|
|
return rank;
|
|
|
|
}
|
|
|
|
|
|
|
|
static Nd4jLong* getNpyArrayShape(void *npArray){
|
|
|
|
cnpy::NpyArray* arr = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
int ndim = arr->shape.size();
|
|
|
|
Nd4jLong* shape = new Nd4jLong[ndim];
|
|
|
|
for (int i=0; i<ndim; i++){
|
|
|
|
shape[i] = arr->shape.at(i);
|
2019-06-06 14:21:15 +02:00
|
|
|
}
|
2019-07-22 13:34:08 +02:00
|
|
|
return shape;
|
|
|
|
}
|
|
|
|
|
|
|
|
static char getNpyArrayOrder(void *npArray){
|
|
|
|
cnpy::NpyArray* arr = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
return (arr->fortranOrder)?'f':'c';
|
|
|
|
}
|
|
|
|
|
|
|
|
static int getNpyArrayElemSize(void *npArray){
|
|
|
|
cnpy::NpyArray* arr = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
return arr->wordSize;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void deleteNPArrayStruct(void *npArray){
|
|
|
|
cnpy::NpyArray* arr = reinterpret_cast<cnpy::NpyArray*>(npArray);
|
|
|
|
delete arr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void deleteNPArrayMap(void *map){
|
|
|
|
cnpy::npz_t* arrays = reinterpret_cast<cnpy::npz_t*>(map);
|
|
|
|
delete arrays;
|
|
|
|
}
|
|
|
|
//////
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Get the element size for a numpy array
|
|
|
|
* @param npyArray the numpy array's address
|
|
|
|
* to get the length for
|
|
|
|
* @return
|
|
|
|
*/
|
|
|
|
static int elementSizeForNpyArray(Nd4jPointer npyArray) {
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromPointer(reinterpret_cast<char *>(npyArray));
|
|
|
|
cnpy::NpyArray *arrPointer = &arr;
|
|
|
|
int size = arrPointer->wordSize;
|
|
|
|
// arrPointer->destruct();
|
|
|
|
return size;
|
|
|
|
}
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Get the element size for a numpy array
|
|
|
|
* @param npyArray the numpy array's address
|
|
|
|
* to get the length for
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-22 13:34:08 +02:00
|
|
|
static int elementSizeForNpyArrayHeader(Nd4jPointer npyArray) {
|
|
|
|
cnpy::NpyArray arr = cnpy::loadNpyFromHeader(reinterpret_cast<char *>(npyArray));
|
|
|
|
cnpy::NpyArray *arrPointer = &arr;
|
|
|
|
int size = arrPointer->wordSize;
|
|
|
|
return size;
|
|
|
|
}
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
static void releaseNumpy(Nd4jPointer npyArray) {
|
|
|
|
free(reinterpret_cast<void *>(npyArray));
|
|
|
|
}
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
* Return the length of a shape buffer
|
|
|
|
* based on the pointer
|
|
|
|
* @param buffer the buffer pointer to check
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int lengthForShapeBufferPointer(Nd4jPointer buffer);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
/**
|
|
|
|
* The pointer to get the address for
|
|
|
|
*
|
|
|
|
* @param address the address to get the pointer
|
|
|
|
* @return the pointer for the given address
|
|
|
|
*/
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer pointerForAddress(Nd4jLong address);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* This method takes single N-dimensional tensor, and copies its TADs to target arrays
|
|
|
|
*
|
|
|
|
* @param x
|
|
|
|
* @param xShapeInfo
|
|
|
|
* @param targets
|
|
|
|
* @param zShapeInfo
|
|
|
|
* @return
|
|
|
|
*/
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void tear(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
OpaqueDataBuffer *dbX, Nd4jLong const* xShapeInfo, Nd4jLong const* dxShapeInfo,
|
|
|
|
Nd4jPointer *targets, Nd4jLong const* zShapeInfo,
|
|
|
|
Nd4jLong const* tadShapeInfo,
|
|
|
|
Nd4jLong const* tadOffsets);
|
2019-07-22 13:34:08 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sort(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool descending);
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortByKey(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *y, Nd4jLong const* yShapeInfo,
|
|
|
|
void *dy, Nd4jLong const* dyShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool descending);
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortByValue(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *y, Nd4jLong const* yShapeInfo,
|
|
|
|
void *dy, Nd4jLong const* dyShapeInfo,
|
Merge master to upstream (#7945)
* Shugeo strided slice zeros (#14)
* Modified strided_slice op to properly work with empty-like shapes.
* Fixed test for reduce_mean with empty-like input.
* [WIP] Last merge (#15)
* correct logsoftmax looss (#2)
* Small SameDiff listener fix (#4)
* Various fixes (#6)
* #7839 Fix for asXMatrix and tests
* #7866 EmbeddingSequenceLayer dtype fix + test
* #7856 SameDiff save/load stream methods
* #7859 RegressionEvaluation rank 4 fix + tests + axis configuration
* EvaluationBinary 3d/4d
* More evaluation 3d/4d tests
* #7847 Evaluation empty checks
* Small test ifx
* #7848 Fix median edge case
* Improve DL4J samediff layer tests
* [WIP] FastText wrapper implemented (#8)
* FastText implemented
* Some fixes
* Fix shapes for wordsNearest
* Validation of input vectors
* Fixes
* Fixed test
* Thread tagged
* Some tweaks
* setContextClassLoader for DeallocatorServiceThread
* Numpy format tests (#1)
* Various fixes (#11)
* #7852 SameDiff gather fix
* #7892 SameDiff placeholder to constant conversion
* #7890 validate input rank for MLN/CG init methods
* Fix broken permute shape calculation
* Permute and gather fixes
* Tests
* #7850 LogSumExp fix + test
* Handful of test fixes
* Empty arrays with non-scalar shapes (#10)
* minor rearrangements for lambdas
* empty tensors with non-scalar shapes
* numpy empty tensors with non-scalar shapes
* few more empty tweaks
* Small fixes
* conv3d signature update
* micro fix in batchnorm mkldnn
* Import fixes
* Fix
* MKL-DNN update
* Small fill fix
* fill with empty input + test
* Fixes
* Small error improvement
* Fix
* one special test
* couple of fixes for lstm
* Rewrite TFGraphMapper.getNDArrayFromTensor to be maintainable and less error prone
* Fixes
* FP16
* Unsigned
* BFloat16
* Fill op - empty tweaks
* - couple of fixes for empty arrays construction
- stack updated
* strided slice fix
* one transform test
* provide method for reducing shapeInfo in case of input array is empty
* Fixed reduceAlongDimensions to use empty input properly.
* couple of broadcast tests
* couple of tests broadcast tests + tweak to make them pass
* add check of non-empty to methods producing sub-arrays
* Fixed reshapeC with zeros in shape.
* complete empty check in reduce_... legacy ops
* Concat and cumsum/prod
* Tweak to empty shape inference on import
* add empty check to the rest of reduce legacy ops
* one more test
* correct typo in evalReduceShapeInfoEmpty
* Added tests for reduce_* ops to tests with zero shapes.
* few more tests for empty reductions
* Fixed strided_slice op with empty case and tests.
* one more empty reduction test
* Fixed strided_slice test.
* add empty check to NDArray::reshapei
* infOrMax
* empty min/max with infinity tests
* made unstack working correctly with empty arrays
* few IndexReduce tests + tweaks for empty shapes
* add test for empty concat
* few tests fixed
* Validation fix for reductions on empty shapes
* Reverse fix
* Reduction shape calc fixes
* SameDiff.generateOutputVariable: don't use shape function to determine number of outputs
* Range fix
* - NDArray constructor updated for scalars/empty arrays
- few tests fixed
* More fixes
* Empty creator fixes
* concat fix
* concat fix
* TF import tests: allow 'both all NaN' and 'both all inf' to pass
* Slice, zero fraction, and reshape fixes
* transpose, gather
* Zero fraction
* scalar cast fix
* Empty reduction axis support
* few more tests fixed
* Fixed input checks conforming with TF for concat op and tests.
* few tests fixed
* matmul scalar shape fix
* Fixed checkout for data type and scalarity with concat to allow non-empty scalars with vector concats.
* broadcast bool fix
* few more tests
* few more tests
* correct evalReduceShapeInfoEmpty
* argmax/argmin + tests
* one more empty edge case + one more test
* argmax/argmin/realdiv_bp tweaks
* empty reshape test + fix
* Helper fixes
* Small fixes
* Gather test fix
* Gather test fix
* Small fixes
* reduce scalar zero values
* scalar mean workaround
* Remove debug code
* along dim mean workaround
* one more test
* - equalsTo() tweak for empty arrays
- one more test
* broadcast tweaks
* [WIP] Fixing outstanding issues for NLP (#9)
* Avoid using not-inited objects
* Test fixed.
* Redundant method avoided for models like FastText
* KMeans++ implementation
* KMeans++ implementation
* Disable parallel execution
* KMeans++
* Tests
* Dev branch merge (#16)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Fix some issues on master (#17)
* Fix DataVec test issue
* Fix issue with dl4j SameDiff output layer
* Dtype fix for lambda layers
* #7912 BertIterator dtype fix (use float32 not global default)
* [WIP] Next set of CUDA stuff (#7)
New CUDA implementations and improvements
* bad file
* Dev branch master merge (#23)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* SameDiff ops, TF import and fixes (#24)
* CheckNumerics tests + fixes + misc fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fake quant
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* FakeQuantWithMinMaxArgs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* CheckNumerics fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix libnd4j ALL_INTS and ALL_FLOATS declaration (uint and bfloat types)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Javadoc
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Exception tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix for out of scope stack allocated var use
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignores
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Ignore for known failing test (already logged issue)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Merge upstream to fork (#25)
* Add thousand-separator commas to TotalParams (#7915)
* Add thousand-separator commas to TotalParams
The number of parameters can be quite large, and it would help the reading of the summary printout to have the TotalParams column & values at the bottom have thousand-separator-commas in them.
* Add thousand-separator commas to MultiLayerNetwork
Corresponding change to MultiLayerNetwork
Signed-off-by: Jxtps Jxtps <jxtps435@gmail.com>
* Update contributing and issue/PR templates (#7934)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Fix link to AdaDelta paper (#7942)
Fix link to AdaDelta paper hosted on matthewzeiler.com
Signed-off-by: Jxtps
* Fixes, and ignores for known/logged failing issues (#7943)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* SameDiff + DL4J/SameDiff: Multiple fixes (#28)
* #7919 HDF5 attribute buffer length fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7909 Arbiter constructor exception ux improvements
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7925 RNN output layer length checks
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Add listener for validating inputs are not incorrectly modified
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7939 Integrate NonInplaceValidationListener into tests
* #7844 DL4J SameDiff fixes for variable minibatch size
* DL4J SameDiff fixes - ensure gradient for input placeholder is available
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tweaks to ExternalErrorsFunction - use placeholders, make more robust
* Another fix
* More fixes
* More SameDiff/DL4J fixes
* Scope out scalar array creation in BaseScalarOp
* Remove debug code
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Final dev branch merge (#29)
* SameDiff: convertDataType and gradient check util improvements (#12)
* GradCheck util improvements
* StopGradient constructor + test
* SameDiff: Add datatype conversion
* Javadoc and add DataType.isNumerical()
* Small fix
* Fix SameDiff TF import test cases intermediate naming (workaround for bad default)
* TFGraphTestAllHelper: check intermediates in execution order
* Add missing debug listener
* [WIP] lstmBlock fix + other changes (#13)
- fixes lstmBlock issue
- changes NDArray method reshape(), permute(), transpose() by making them return instance instead of pointer
- CheckNumerics op
- fixes for ReduceBool IsInfOrNan & IsFinite
* Small test fix
* CheckNumerics op wrapper
* Compatibility of deserialization (#18)
Signed-off-by: Alexander Stoyakin <alexander.stoyakin@gmail.com>
* SameDiff: add activation gradient checking support for debugging (#19)
* SameDiff gradient checker: first pass on activation gradient checks
* Fixes + tests for activation gradient checking
* Javadoc
* [WIP] Some nd4j data type corrections (#20)
* Adjust data type
* Set correct Data type.
* Size of proper data type.
* fix averaged cpu load (#22)
* [WIP] Multiple dataset iterators (#27)
* Splitting dataset into arbitrary number
* Fixes
* Multiple split of iterator
* Test
* Test
* Some fixes
* signature change
* one more tweak
Signed-off-by: raver119 <raver119@gmail.com>
* one more test for sequential use of DataSetIteratorSplitter
Signed-off-by: raver119 <raver119@gmail.com>
* Fixes
* Fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* one more test for Alexander
Signed-off-by: raver119 <raver119@gmail.com>
* minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Some fixes
* Some fixes
* couple of assertions tweaked
Signed-off-by: raver119 <raver119@gmail.com>
* MDS splitter test :/
Signed-off-by: raver119 <raver119@gmail.com>
* Minor refactoring
* Multi dataset
* Some fixes
* More tests
* Small number of test fixes/improvements (failures on CI) (#31)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] More CUDA stuff (#26)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* LRN BP CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* less memory
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed bug with crop_and_resize op helper.
* get rid of unnecessary index-calculation dunction
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed sort with nth_element cuda-based helper.
* Refactored nth_element.
* Refactored nth_element op and tests.
* Modified usage of dim array with sortTad routine.
* Refactored main routine of helper for non_max_image_suppression op.
* non_max_image_suppression op helper with cuda kernel implementation. Initial revision.
* fix vol2col cuda kernel
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* topK concept
Signed-off-by: raver119 <raver119@gmail.com>
* unsorted topK with scanWitdh of 1
Signed-off-by: raver119 <raver119@gmail.com>
* correct vol2col tests
* sorted/unsorted topK
Signed-off-by: raver119 <raver119@gmail.com>
* implementation and fixing col2im/col2vol
* Corrected usage flags with input/output with reverse op.
* dup is const now
Signed-off-by: raver119 <raver119@gmail.com>
* percentile op
Signed-off-by: raver119 <raver119@gmail.com>
* group tests for mapool2d
Signed-off-by: Yurii <yurii@skymind.io>
* special test for george
Signed-off-by: raver119 <raver119@gmail.com>
* less threads for sortTad
Signed-off-by: raver119 <raver119@gmail.com>
* provide conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* remove auther in sort tad kernel code
Signed-off-by: Yurii <yurii@skymind.io>
* provide depthwise_conv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* - max_pooling_with_argmax
- null check for special use
Signed-off-by: raver119 <raver119@gmail.com>
* dts cuda
Signed-off-by: raver119 <raver119@gmail.com>
* provide sconv2d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* std cuda
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op to conform TF implementation.
* Improved suppression helper.
* provide pooling3d for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* more of minor lstm rearrangements
Signed-off-by: raver119 <raver119@gmail.com>
* (bi)dynamic_rnn
Signed-off-by: raver119 <raver119@gmail.com>
* templates init order
Signed-off-by: raver119 <raver119@gmail.com>
* Refactored non_max_suppression op.
* Added cuda kernel for non_max_suppression.
* CPU sort by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value
Signed-off-by: raver119 <raver119@gmail.com>
* CPU sort TAD by key/value tests
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminate compiler error with cuda implementation.
* - repaired gradCheck in cuda
- provide conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* missed signature
Signed-off-by: raver119 <raver119@gmail.com>
* provide depthwise_conv2d_bp for cuda
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of lup helper with cuda kernel. Initial commit.
* further work on backprops for convolutions
Signed-off-by: Yurii <yurii@skymind.io>
* CUDA linear sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA tad sort by key/val
Signed-off-by: raver119 <raver119@gmail.com>
* start providing of backprop for pooling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* Added atomicAdd for bool datatype.
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic partition scalar CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* important comment
Signed-off-by: raver119 <raver119@gmail.com>
* fix pooling2d/3d backprop helpers
Signed-off-by: Yurii <yurii@skymind.io>
* Added non-linear test with dynamic_partition.
* Improved test for dynamic_partition.
* dynamic_partition TAD concept
Signed-off-by: raver119 <raver119@gmail.com>
* - dynamic_partition TAD CUDA impl
- dynamic_partition TAD CPU fix
Signed-off-by: raver119 <raver119@gmail.com>
* - rewrite cpu code for usampling2d/3d
- write cuda code for usampling2d/3d
Signed-off-by: Yurii <yurii@skymind.io>
* dynamic_stitch CUDA vector case
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case concept
Signed-off-by: raver119 <raver119@gmail.com>
* dynamic_stitch CUDA TAD case impl
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for dynamic_stitch 3D-4D cases.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed type check for dynamic stitch.
* min/max bp
Signed-off-by: raver119 <raver119@gmail.com>
* rewrite code for upsampling2d/3d cpu
Signed-off-by: Yurii <yurii@skymind.io>
* reduce min/max/norm_max bp
Signed-off-by: raver119 <raver119@gmail.com>
* lup implementation. Additional enhancements.
* provide code for upsamling2d/3d backprop
Signed-off-by: Yurii <yurii@skymind.io>
* weightedCrossEntropyWithLogits
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed template math atomicMul for 64bit ints.
* Refactored dynamic_partition_bp op.
* inverseBroadcast fix
Signed-off-by: raver119 <raver119@gmail.com>
* DynamicPartitionBP test datatype fixed.
* - nd4j_atomicMul Windows fix
- cpu/NDArrayLambda.hpp excluded from CUDA
Signed-off-by: raver119 <raver119@gmail.com>
2019-06-27 17:37:04 +02:00
|
|
|
bool descending);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortTad(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int *dimension,
|
|
|
|
int dimensionLength,
|
2020-05-09 07:06:14 +02:00
|
|
|
Nd4jLong const* tadShapeInfo,
|
|
|
|
Nd4jLong const* tadOffsets,
|
2019-07-22 13:34:08 +02:00
|
|
|
bool descending);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortTadByKey(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *y, Nd4jLong const* yShapeInfo,
|
|
|
|
void *dy, Nd4jLong const* dyShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int *dimension,
|
|
|
|
int dimensionLength,
|
|
|
|
bool descending);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortTadByValue(Nd4jPointer *extraPointers,
|
2020-05-09 07:06:14 +02:00
|
|
|
void *x, Nd4jLong const* xShapeInfo,
|
|
|
|
void *dx, Nd4jLong const* dxShapeInfo,
|
|
|
|
void *y, Nd4jLong const* yShapeInfo,
|
|
|
|
void *dy, Nd4jLong const* dyShapeInfo,
|
2019-07-22 13:34:08 +02:00
|
|
|
int *dimension,
|
|
|
|
int dimensionLength,
|
|
|
|
bool descending);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
// special sort impl for sorting out COO indices and values
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void sortCooIndices(Nd4jPointer *extraPointers, Nd4jLong *indices, void *values, Nd4jLong length, int rank);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jLong* mmapFile(Nd4jPointer *extraPointers, const char *fileName, Nd4jLong length);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void munmapFile(Nd4jPointer *extraPointers, Nd4jLong* ptrMap, Nd4jLong length);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::graph::ResultWrapper OpaqueResultWrapper;
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
// flatbuffers execution
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT OpaqueResultWrapper* executeFlatGraph(Nd4jPointer *extraPointers, Nd4jPointer flatBufferPointer);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getResultWrapperSize(OpaqueResultWrapper* ptr);
|
|
|
|
ND4J_EXPORT Nd4jPointer getResultWrapperPointer(OpaqueResultWrapper* ptr);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT const char* getAllCustomOps();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT const char* getAllOperations();
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
// customOp executioner
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int execCustomOp(Nd4jPointer* extraPointers, Nd4jLong hash, Nd4jPointer* inputBuffers, Nd4jPointer* inputShapes, int numInputs, Nd4jPointer* outputBuffers, Nd4jPointer* outputShapes, int numOutputs, double* tArgs, int numTArgs, Nd4jLong *iArgs, int numIArgs, bool* bArgs, int numBArgs, bool isInplace);
|
|
|
|
ND4J_EXPORT int execCustomOp2(Nd4jPointer* extraPointers, Nd4jLong hash, Nd4jPointer opContext);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::ShapeList OpaqueShapeList;
|
2019-07-24 14:14:54 +02:00
|
|
|
|
|
|
|
ND4J_EXPORT OpaqueShapeList* calculateOutputShapes(Nd4jPointer* extraPointers, Nd4jLong hash, Nd4jPointer* inputShapes, int numInputShapes, double* tArgs, int numTArgs, Nd4jLong *iArgs, int numIArgs);
|
2020-01-30 16:46:12 +01:00
|
|
|
ND4J_EXPORT OpaqueShapeList* calculateOutputShapes2(Nd4jPointer* extraPointers, Nd4jLong hash, Nd4jPointer* inputBuffers, Nd4jPointer* inputShapes, int numInputShapes, double* tArgs, int numTArgs, Nd4jLong *iArgs, int numIArgs, bool *bArgs, int numBArgs, int *dArgs, int numDArgs);
|
2019-07-24 14:14:54 +02:00
|
|
|
|
|
|
|
ND4J_EXPORT Nd4jLong getShapeListSize(OpaqueShapeList* list);
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT Nd4jLong const* getShape(OpaqueShapeList* list, Nd4jLong i);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void deleteShapeList(Nd4jPointer shapeList);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int registerGraph(Nd4jPointer *extraPointers, Nd4jLong graphId, Nd4jPointer flatBufferPointer);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::graph::VariablesSet OpaqueVariablesSet;
|
|
|
|
typedef sd::graph::Variable OpaqueVariable;
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2019-07-26 09:22:44 +02:00
|
|
|
ND4J_EXPORT OpaqueVariablesSet *executeStoredGraph(Nd4jPointer *extraPointers, Nd4jLong graphId, Nd4jPointer *inputBuffers, Nd4jPointer *inputShapes, int* inputIndices, int numInputs);
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2019-07-26 09:22:44 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getVariablesSetSize(OpaqueVariablesSet* set);
|
|
|
|
ND4J_EXPORT Nd4jStatus getVariablesSetStatus(OpaqueVariablesSet* set);
|
|
|
|
ND4J_EXPORT OpaqueVariable* getVariable(OpaqueVariablesSet* set, Nd4jLong i);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT int getVariableId(OpaqueVariable* variable);
|
|
|
|
ND4J_EXPORT int getVariableIndex(OpaqueVariable* variable);
|
|
|
|
ND4J_EXPORT const char* getVariableName(OpaqueVariable* variable);
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT Nd4jLong const* getVariableShape(OpaqueVariable* variable);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT void* getVariableBuffer(OpaqueVariable* variable);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT int unregisterGraph(Nd4jPointer *extraPointers, Nd4jLong graphId);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void deleteCharArray(Nd4jPointer pointer);
|
|
|
|
ND4J_EXPORT void deleteIntArray(Nd4jPointer pointer);
|
|
|
|
ND4J_EXPORT void deleteLongArray(Nd4jPointer pointer);
|
|
|
|
ND4J_EXPORT void deletePointerArray(Nd4jPointer pointer);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-26 09:22:44 +02:00
|
|
|
ND4J_EXPORT void deleteVariablesSet(OpaqueVariablesSet* pointer);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
// GraphState creation
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer getGraphState(Nd4jLong id);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void deleteGraphState(Nd4jPointer state);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void deleteResultWrapper(Nd4jPointer ptr);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT int estimateThreshold(Nd4jPointer *extraPointers, Nd4jPointer x, Nd4jLong const* xShapeInfo, int N, float threshold);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
// this method executes op that requires scope to be present: if/while/cond/whatever
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jStatus execCustomOpWithScope(Nd4jPointer *extraPointers, Nd4jPointer state, Nd4jLong opHash, Nd4jLong *scopes, int numScopes, Nd4jPointer *inputBuffers, Nd4jPointer *inputShapes, int numInputs, Nd4jPointer *outputBuffers, Nd4jPointer *outputShapes, int numOutputs);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
//void fillUtf8String(Nd4jPointer *extraPointers, const char **string, int numStrings, Nd4jPointer buffer);
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer createUtf8String(Nd4jPointer *extraPointers, const char *string, int length);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT Nd4jLong getUtf8StringLength(Nd4jPointer *extraPointers, Nd4jPointer ptr);
|
|
|
|
ND4J_EXPORT char* getUtf8StringBuffer(Nd4jPointer *extraPointers, Nd4jPointer ptr);
|
|
|
|
ND4J_EXPORT void deleteUtf8String(Nd4jPointer *extraPointers, Nd4jPointer ptr);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void scatterUpdate(Nd4jPointer *extraPointers, int opCode, int numOfSubArrs,
|
2020-05-09 07:06:14 +02:00
|
|
|
void* hX, Nd4jLong const* hXShapeInfo, Nd4jLong const* hXOffsets,
|
|
|
|
void* dX, Nd4jLong const* dXShapeInfo, Nd4jLong const* dXOffsets,
|
|
|
|
void* hY, Nd4jLong const* hYShapeInfo, Nd4jLong const* hYOffsets,
|
|
|
|
void* dY, Nd4jLong const* dYShapeInfo, Nd4jLong const* dYOffsets,
|
|
|
|
void* hIindexes, Nd4jLong const* hIndicesShapeInfo, void* dIindexes, Nd4jLong const* dIndicesShapeInfo);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT void inspectArray(Nd4jPointer *extraPointers, Nd4jPointer buffer, Nd4jLong *shapeInfo, Nd4jPointer specialBuffer, Nd4jLong *specialShapeInfo, Nd4jPointer debugInfo);
|
2019-07-12 07:21:15 +02:00
|
|
|
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::ConstantDataBuffer OpaqueConstantDataBuffer;
|
2020-06-06 14:26:55 +02:00
|
|
|
typedef sd::ConstantShapeBuffer OpaqueConstantShapeBuffer;
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2020-06-06 14:26:55 +02:00
|
|
|
ND4J_EXPORT OpaqueConstantShapeBuffer* shapeBuffer(int rank, Nd4jLong *shape, Nd4jLong *strides, sd::DataType dtype, char order, Nd4jLong ews, bool empty);
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2020-05-09 07:06:14 +02:00
|
|
|
ND4J_EXPORT OpaqueConstantDataBuffer* constantBufferLong(sd::DataType dtype, Nd4jLong const* data, int length);
|
2020-03-02 10:49:41 +01:00
|
|
|
ND4J_EXPORT OpaqueConstantDataBuffer* constantBufferDouble(sd::DataType dtype, double *data, int length);
|
|
|
|
ND4J_EXPORT OpaqueConstantDataBuffer* constantBuffer(sd::DataType dtype, sd::ConstantDescriptor *descriptor);
|
2019-07-24 14:14:54 +02:00
|
|
|
|
|
|
|
ND4J_EXPORT Nd4jPointer getConstantDataBufferPrimary(OpaqueConstantDataBuffer* dbf);
|
|
|
|
ND4J_EXPORT Nd4jPointer getConstantDataBufferSpecial(OpaqueConstantDataBuffer* dbf);
|
|
|
|
ND4J_EXPORT Nd4jLong getConstantDataBufferLength(OpaqueConstantDataBuffer* dbf);
|
|
|
|
|
2020-06-06 14:26:55 +02:00
|
|
|
ND4J_EXPORT Nd4jPointer getConstantShapeBufferPrimary(OpaqueConstantShapeBuffer* dbf);
|
|
|
|
ND4J_EXPORT Nd4jPointer getConstantShapeBufferSpecial(OpaqueConstantShapeBuffer* dbf);
|
|
|
|
|
|
|
|
ND4J_EXPORT void deleteConstantShapeBuffer(OpaqueConstantShapeBuffer* ptr);
|
|
|
|
ND4J_EXPORT void deleteConstantDataBuffer(OpaqueConstantDataBuffer* ptr);
|
2019-07-24 14:14:54 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::graph::Context OpaqueContext;
|
|
|
|
typedef sd::graph::RandomGenerator OpaqueRandomGenerator;
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT OpaqueContext* createGraphContext(int nodeId);
|
|
|
|
ND4J_EXPORT OpaqueRandomGenerator* getGraphContextRandomGenerator(OpaqueContext* ptr);
|
2019-11-14 12:35:02 +01:00
|
|
|
ND4J_EXPORT void ctxAllowHelpers(OpaqueContext* ptr, bool reallyAllow);
|
2020-01-04 07:06:44 +01:00
|
|
|
ND4J_EXPORT void ctxShapeFunctionOverride(OpaqueContext* ptr, bool reallyOverride);
|
2020-01-27 08:00:07 +01:00
|
|
|
ND4J_EXPORT void ctxSetExecutionMode(OpaqueContext* ptr, int execMode);
|
2020-02-05 05:27:24 +01:00
|
|
|
ND4J_EXPORT void ctxPurge(OpaqueContext* ptr);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT void markGraphContextInplace(OpaqueContext* ptr, bool reallyInplace);
|
|
|
|
ND4J_EXPORT void setGraphContextCudaContext(OpaqueContext* ptr, void *stream, void *reductionPointer, void *allocationPointer);
|
|
|
|
ND4J_EXPORT void setGraphContextInputArray(OpaqueContext* ptr, int index, void *buffer, void *shapeInfo, void *specialBuffer, void *specialShapeInfo);
|
|
|
|
ND4J_EXPORT void setGraphContextOutputArray(OpaqueContext* ptr, int index, void *buffer, void *shapeInfo, void *specialBuffer, void *specialShapeInfo);
|
2020-01-04 11:27:50 +01:00
|
|
|
ND4J_EXPORT void setGraphContextInputBuffer(OpaqueContext* ptr, int index, OpaqueDataBuffer *buffer, void *shapeInfo, void *specialShapeInfo);
|
|
|
|
ND4J_EXPORT void setGraphContextOutputBuffer(OpaqueContext* ptr, int index, OpaqueDataBuffer *buffer, void *shapeInfo, void *specialShapeInfo);
|
2020-01-30 08:07:24 +01:00
|
|
|
ND4J_EXPORT void setGraphContextDArguments(OpaqueContext* ptr, int *arguments, int numberOfArguments);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT void setGraphContextTArguments(OpaqueContext* ptr, double *arguments, int numberOfArguments);
|
|
|
|
ND4J_EXPORT void setGraphContextIArguments(OpaqueContext* ptr, Nd4jLong *arguments, int numberOfArguments);
|
|
|
|
ND4J_EXPORT void setGraphContextBArguments(OpaqueContext* ptr, bool *arguments, int numberOfArguments);
|
|
|
|
ND4J_EXPORT void deleteGraphContext(OpaqueContext* ptr);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT OpaqueRandomGenerator* createRandomGenerator(Nd4jLong rootSeed = 0, Nd4jLong nodeSeed = 0);
|
|
|
|
ND4J_EXPORT Nd4jLong getRandomGeneratorRootState(OpaqueRandomGenerator* ptr);
|
|
|
|
ND4J_EXPORT Nd4jLong getRandomGeneratorNodeState(OpaqueRandomGenerator* ptr);
|
|
|
|
ND4J_EXPORT void setRandomGeneratorStates(OpaqueRandomGenerator* ptr, Nd4jLong rootSeed = 0, Nd4jLong nodeSeed = 0);
|
2020-05-30 20:13:33 +02:00
|
|
|
ND4J_EXPORT float getRandomGeneratorRelativeFloat(OpaqueRandomGenerator* ptr, Nd4jLong index);
|
|
|
|
ND4J_EXPORT double getRandomGeneratorRelativeDouble(OpaqueRandomGenerator* ptr, Nd4jLong index);
|
2019-07-24 14:14:54 +02:00
|
|
|
ND4J_EXPORT int getRandomGeneratorRelativeInt(OpaqueRandomGenerator* ptr, Nd4jLong index);
|
|
|
|
ND4J_EXPORT Nd4jLong getRandomGeneratorRelativeLong(OpaqueRandomGenerator* ptr, Nd4jLong index);
|
|
|
|
ND4J_EXPORT void deleteRandomGenerator(OpaqueRandomGenerator* ptr);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2019-07-23 07:36:48 +02:00
|
|
|
ND4J_EXPORT const char* runLightBenchmarkSuit(bool printOut);
|
|
|
|
ND4J_EXPORT const char* runFullBenchmarkSuit(bool printOut);
|
2019-06-06 14:21:15 +02:00
|
|
|
|
2020-03-02 10:49:41 +01:00
|
|
|
typedef sd::LaunchContext OpaqueLaunchContext;
|
[WIP] multi-device support (#80)
* fix pad javadoc and @see links. (#72)
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* [WIP] More fixes (#73)
* special tests for ConstantTadHelper/ConstantShapeHelper
Signed-off-by: raver119 <raver119@gmail.com>
* release methods for data buffers
Signed-off-by: raver119 <raver119@gmail.com>
* delete temporary buffer Java side
Signed-off-by: raver119 <raver119@gmail.com>
* delete temporary buffer Java side
Signed-off-by: raver119 <raver119@gmail.com>
* delete temporary TadPack C++/Java side (#74)
Signed-off-by: raver119 <raver119@gmail.com>
* Zoo model TF import test updates (#75)
* argLine fix, update compression_gru comment
* updated comment for xception
* undid but commented argLine change
* updated xlnet comment
* copyright headers
* - new NDArray methods like()/ulike() (#77)
- fix for depthwise_conv2d_bp + special test
Signed-off-by: raver119 <raver119@gmail.com>
* upsampling2d fix CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* DL4J trace logging (#79)
* MLN/CG trace logging for debugging
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Tiny tweak
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* strided_slice_bp shape fn leak fix
Signed-off-by: raver119 <raver119@gmail.com>
* SameDiff fixes and naming (#78)
* remove SDVariable inplace methods
* import methods
* npe fix in OpVal
* removed SameDiff inplace ops from tests
* Naming updates, moved to centralized methods in SameDiff, should use op_#:# for everything
* quick fixes
* javadoc
* SDVariable eval with placeholders
* use regex match
* better matching
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* fix javadoc. (#76)
* fix javadoc.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* replace most @see with @link s.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* 4 additional tests
Signed-off-by: raver119 <raver119@gmail.com>
* launch context reorganization
Signed-off-by: raver119 <raver119@gmail.com>
* LaunchContext reorganization
Signed-off-by: raver119 <raver119@gmail.com>
* per-device LaunchContext
Signed-off-by: raver119 <raver119@gmail.com>
* Various DL4J/ND4J fixes (#81)
* #7954 Force refresh of UI when switching tabs on overview page
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8017 Concurrent modification exception (synchronize) fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8033 Don't initialize updater in middle of writing memory crash dump
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8208 Fix shape checks for ND4J int[] creator methods
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #6385 #7992 Keras import naming fixes + cleanup
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8016 Upsampling3D - add NDHWC format support
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* ContextBuffers as separate entity
Signed-off-by: raver119 <raver119@gmail.com>
* Refactor NativeOps.h to export C functions
* Actually export functions from NativeOps.h
* Adapt the Java wrappers in ND4J generated with JavaCPP
* Create C wrappers for some of the C++ classes currently used by ND4J
* ContextBuffers as separate entity
Signed-off-by: raver119 <raver119@gmail.com>
* remove duplicate code in createBufferDetached. (#83)
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* Keras model import - updater lr fix (#84)
* Keras model import - updater lr fix
Signed-off-by: eraly <susan.eraly@gmail.com>
* Keras model import - updater lr fix, cleanup
Signed-off-by: eraly <susan.eraly@gmail.com>
* ContextBuffers as separate entity
Signed-off-by: raver119 <raver119@gmail.com>
* ContextBuffers as separate entity
Signed-off-by: raver119 <raver119@gmail.com>
* Fix functions of OpaqueVariablesSet
* thread-local buffers/affinity
Signed-off-by: raver119 <raver119@gmail.com>
* thread safety for LaunchContext
Signed-off-by: raver119 <raver119@gmail.com>
* more of thread safety
Signed-off-by: raver119 <raver119@gmail.com>
* one more multi threaded test
Signed-off-by: raver119 <raver119@gmail.com>
* SameDiff Convolution Config validation, better output methods (#82)
* Conv Config validation & tests
Signed-off-by: Ryan Nett <rnett@skymind.io>
* stackOutputs utility method
Signed-off-by: Ryan Nett <rnett@skymind.io>
* use constructor for validation, support negative kernel sizes (infered from weights)
Signed-off-by: Ryan Nett <rnett@skymind.io>
* better output methods
Signed-off-by: Ryan Nett <rnett@skymind.io>
* move output to be with fit and evaluate
Signed-off-by: Ryan Nett <rnett@skymind.io>
* fixes
Signed-off-by: Ryan Nett <rnett@skymind.io>
* more fixes
Signed-off-by: Ryan Nett <rnett@skymind.io>
* refactor duplicate code from pad methods. (#86)
* refactor duplicate code from pad methods.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* replace switch with if.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* Various ND4J/DL4J fixes and improvements (#87)
* Reshape and reallocate - small fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Reshape and reallocate - small fixes
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #6488 ElementWiseVertex broadcast support
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Constructors and broadcast supported it Transforms.max/min
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8054 ElementWiseVertex now supports broadcast inputs
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8057 Nd4j.create overload dtype fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7551 ND4J Shape validation fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* [WIP] Numpy boolean import (#91)
* numpy bool type
Signed-off-by: raver119 <raver119@gmail.com>
* numpy bool java side
Signed-off-by: raver119 <raver119@gmail.com>
* remove create method with unused parameter. (#89)
* remove create method with unused parameter.
* removed more unused methods.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* removing more unused code.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* last removal of unused code.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* remove createSparse methods. (#92)
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* Various ND4J/DL4J fixes (#90)
* Deprecate Old*Op instances
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8063 #8054 Broadcast exceptions + cleanup inplace ops
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Small fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* Remove bad test condition
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #7993 Fix shape function issue in crop_and_resize op
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* DL4J SameDiff lambda layer fix
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8029 Fix for pnorm backprop math
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* #8038 Fix Op profiler NaN/Inf triggering + add tests (#93)
Signed-off-by: AlexDBlack <blacka101@gmail.com>
* createUninitializedDetached refactoring. (#94)
* wip
* update interface, add null implementations.
* Breaking one test in a weird way.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* createUninitializedDetached refactored.
Signed-off-by: Robert Altena <Rob@Ra-ai.com>
* cuda build fix for issues introduced by recent refactoring
Signed-off-by: raver119 <raver119@gmail.com>
* [WIP] More of CUDA (#95)
* initial commit
Signed-off-by: raver119 <raver119@gmail.com>
* Implementation of hashcode cuda helper. Working edition.
* Fixed parallel test input arangements.
* Fixed tests for hashcode op.
* Fixed shape calculation for image:crop_and_resize op and test.
* NativeOps tests. Initial test suite.
* Added tests for indexReduce methods.
* Added test on execBroadcast with NDArray as dimensions.
* Added test on execBroadcastBool with NDArray as dimensions.
* Added tests on execPairwiseTransform and execPairwiseTransofrmBool.
* Added tests for execReduce with scalar results.
* Added reduce tests for non-empty dims array.
* Added tests for reduce3.
* Added tests for execScalar.
* Added tests for execSummaryStats.
* - provide cpu/cuda code for batch_to_space
- testing it
Signed-off-by: Yurii <yurii@skymind.io>
* - remove old test for batch_to_space (had wrong format and numbers were not checked)
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed complilation errors with test.
* Added test for execTransformFloat.
* Added test for execTransformSame.
* Added test for execTransformBool.
* Added test for execTransformStrict.
* Added tests for execScalar/execScalarBool with TADs.
* Added test for flatten.
* - provide cpu/cuda code for space_to_Batch operaion
Signed-off-by: Yurii <yurii@skymind.io>
* Added test for concat.
* comment unnecessary stuff in s_t_b
Signed-off-by: Yurii <yurii@skymind.io>
* Added test for specialConcat.
* Added tests for memcpy/set routines.
* Fixed pullRow cuda test.
* Added pullRow test.
* Added average test.
* - correct typo in NDArray::applyPairwiseTransform(nd4j::pairwise::BoolOps op...)
Signed-off-by: Yurii <yurii@skymind.io>
* - debugging and fixing cuda tests in JavaInteropTests file
Signed-off-by: Yurii <yurii@skymind.io>
* - correct some tests
Signed-off-by: Yurii <yurii@skymind.io>
* Added test for shuffle.
* Fixed ops declarations.
* Restored omp and added shuffle test.
* Added convertTypes test.
* Added tests for execRandom. Eliminated usage of RandomBuffer with NativeOps.
* Added sort tests.
* Added tests for execCustomOp.
* - further debuging and fixing tests terminated with crash
Signed-off-by: Yurii <yurii@skymind.io>
* Added tests for calculateOutputShapes.
* Addded Benchmarks test.
* Commented benchmark tests.
* change assertion
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for apply_sgd op. Added cpu helper for that op.
* Implement cuda helper for aplly_sgd op. Fixed tests for NativeOps.
* Added test for assign broadcastable.
* Added tests for assign_bp op.
* Added tests for axpy op.
* - assign/execScalar/execTransformAny signature change
- minor test fix
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed axpy op.
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* - fix tests for nativeOps::concat
Signed-off-by: Yurii <yurii@skymind.io>
* sequential transform/scalar
Signed-off-by: raver119 <raver119@gmail.com>
* allow nested parallelism
Signed-off-by: raver119 <raver119@gmail.com>
* assign_bp leak fix
Signed-off-by: raver119 <raver119@gmail.com>
* block setRNG fix
Signed-off-by: raver119 <raver119@gmail.com>
* enable parallelism by default
Signed-off-by: raver119 <raver119@gmail.com>
* enable nested parallelism by default
Signed-off-by: raver119 <raver119@gmail.com>
* Added cuda implementation for row_count helper.
* Added implementation for tnse gains op helper.
* - take into account possible situations when input arrays are empty in reduce_ cuda stuff
Signed-off-by: Yurii <yurii@skymind.io>
* Implemented tsne/edge_forces op cuda-based helper. Parallelized cpu-based helper for edge_forces.
* Added kernel for tsne/symmetrized op heleper.
* Implementation of tsne/symmetrized op cuda helper. Working edition.
* Eliminated waste printfs.
* Added test for broadcastgradientargs op.
* host-only fallback for empty reduce float
Signed-off-by: raver119 <raver119@gmail.com>
* - some tests fixes
Signed-off-by: Yurii <yurii@skymind.io>
* - correct the rest of reduce_ stuff
Signed-off-by: Yurii <yurii@skymind.io>
* - further correction of reduce_ stuff
Signed-off-by: Yurii <yurii@skymind.io>
* Added test for Cbow op. Also added cuda implementation for cbow helpers.
* - improve code of stack operation for scalar case
Signed-off-by: Yurii <yurii@skymind.io>
* - provide cuda kernel for gatherND operation
Signed-off-by: Yurii <yurii@skymind.io>
* Implementation of cbow helpers with cuda kernels.
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* minor tests tweaks
Signed-off-by: raver119 <raver119@gmail.com>
* - further correction of cuda stuff
Signed-off-by: Yurii <yurii@skymind.io>
* Implementatation of cbow op helper with cuda kernels. Working edition.
* Skip random testing for cudablas case.
* lstmBlockCell context fix
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for ELU and ELU_BP ops.
* Added tests for eq_scalar, gt_scalar, gte_scalar and lte_scalar ops.
* Added tests for neq_scalar.
* Added test for noop.
* - further work on clipbynorm_bp
Signed-off-by: Yurii <yurii@skymind.io>
* - get rid of concat op call, use instead direct concat helper call
Signed-off-by: Yurii <yurii@skymind.io>
* lstmBlockCell context fix
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for lrelu and lrelu_bp.
* Added tests for selu and selu_bp.
* Fixed lrelu derivative helpers.
* - some corrections in lstm
Signed-off-by: Yurii <yurii@skymind.io>
* operator * result shape fix
Signed-off-by: raver119 <raver119@gmail.com>
* - correct typo in lstmCell
Signed-off-by: Yurii <yurii@skymind.io>
* few tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* CUDA inverse broadcast bool fix
Signed-off-by: raver119 <raver119@gmail.com>
* disable MMAP test for CUDA
Signed-off-by: raver119 <raver119@gmail.com>
* BooleanOp syncToDevice
Signed-off-by: raver119 <raver119@gmail.com>
* meh
Signed-off-by: raver119 <raver119@gmail.com>
* additional data types for im2col/col2im
Signed-off-by: raver119 <raver119@gmail.com>
* Added test for firas_sparse op.
* one more RandomBuffer test excluded
Signed-off-by: raver119 <raver119@gmail.com>
* Added tests for flatten op.
* Added test for Floor op.
* bunch of tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* mmulDot tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* more tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* Implemented floordiv_bp op and tests.
* Fixed scalar case with cuda implementation for bds.
* - work on cuda kernel for clip_by_norm backprop op is completed
Signed-off-by: Yurii <yurii@skymind.io>
* Eliminate cbow crach.
* more tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* more tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* Eliminated abortion with batched nlp test.
* more tests fixed
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed shared flag initializing.
* disabled bunch of cpu workspaces tests
Signed-off-by: raver119 <raver119@gmail.com>
* scalar operators fix: missing registerSpecialUse call
Signed-off-by: raver119 <raver119@gmail.com>
* Fixed logdet for cuda and tests.
* - correct clipBynorm_bp
Signed-off-by: Yurii <yurii@skymind.io>
* Fixed crop_and_resize shape datatype.
* - correct some mmul tests
Signed-off-by: Yurii <yurii@skymind.io>
* build fix
Signed-off-by: raver119 <raver119@gmail.com>
* exclude two methods for JNI
Signed-off-by: raver119 <raver119@gmail.com>
* exclude two methods for JNI
Signed-off-by: raver119 <raver119@gmail.com>
* exclude two methods for JNI (#97)
Signed-off-by: raver119 <raver119@gmail.com>
* temporary stack fix
Signed-off-by: raver119 <raver119@gmail.com>
* round robin affinity test
Signed-off-by: raver119 <raver119@gmail.com>
* get rid of legacy CudaContext methods
Signed-off-by: raver119 <raver119@gmail.com>
* get rid of legacy ContextPool classes/methods
Signed-off-by: raver119 <raver119@gmail.com>
* one legacy test removed
Signed-off-by: raver119 <raver119@gmail.com>
* few more fields rearranged
Signed-off-by: raver119 <raver119@gmail.com>
* OpaqueLaunchContext
Signed-off-by: raver119 <raver119@gmail.com>
* OpaqueLaunchContext++
Signed-off-by: raver119 <raver119@gmail.com>
* more of OpaqueLaunchContext methods
Signed-off-by: raver119 <raver119@gmail.com>
* LaunchContext -> CudaContext
Signed-off-by: raver119 <raver119@gmail.com>
* AffinityManger changes
Signed-off-by: raver119 <raver119@gmail.com>
* AffinityManger changes
Signed-off-by: raver119 <raver119@gmail.com>
* cusolver handles
Signed-off-by: raver119 <raver119@gmail.com>
* typo
Signed-off-by: raver119 <raver119@gmail.com>
* cusolver method
Signed-off-by: raver119 <raver119@gmail.com>
* cusolver handle propagated
Signed-off-by: raver119 <raver119@gmail.com>
* blas/solver handles
Signed-off-by: raver119 <raver119@gmail.com>
* one more test
Signed-off-by: raver119 <raver119@gmail.com>
* legacy concat implementations replaced with new CustomOp
Signed-off-by: raver119 <raver119@gmail.com>
* one more test
Signed-off-by: raver119 <raver119@gmail.com>
* concat now uses way more blocks
Signed-off-by: raver119 <raver119@gmail.com>
* print
Signed-off-by: raver119 <raver119@gmail.com>
* no more triple template mmul
Signed-off-by: raver119 <raver119@gmail.com>
* bunch of kernels have dtypes reconsidered
Signed-off-by: raver119 <raver119@gmail.com>
* bunch of kernels have dtypes reconsidered
Signed-off-by: raver119 <raver119@gmail.com>
* bitonic sort reorganized
Signed-off-by: raver119 <raver119@gmail.com>
* bunch of cpu stuff removed from cuda scope
Signed-off-by: raver119 <raver119@gmail.com>
* bunch of cpu stuff removed from cuda scope
Signed-off-by: raver119 <raver119@gmail.com>
* type conversions moved to generic impl
Signed-off-by: raver119 <raver119@gmail.com>
* cpu data types pass
Signed-off-by: raver119 <raver119@gmail.com>
* non_max_suppression
Signed-off-by: raver119 <raver119@gmail.com>
* sortByValue fix
Signed-off-by: raver119 <raver119@gmail.com>
* ignore all mixed datatype tests for mmul
Signed-off-by: raver119 <raver119@gmail.com>
* special handling of OpProfiler exceptions
Signed-off-by: raver119 <raver119@gmail.com>
* - one failing concat test in cpp
- Nd4j.tile now uses op internally
Signed-off-by: raver119 <raver119@gmail.com>
* get back dtype exception for legacy arrays deserialization
Signed-off-by: raver119 <raver119@gmail.com>
2019-08-14 15:52:34 +02:00
|
|
|
|
|
|
|
ND4J_EXPORT OpaqueLaunchContext* defaultLaunchContext();
|
|
|
|
ND4J_EXPORT Nd4jPointer lcScalarPointer(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcReductionPointer(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcAllocationPointer(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcExecutionStream(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcCopyStream(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcBlasHandle(OpaqueLaunchContext* lc);
|
|
|
|
ND4J_EXPORT Nd4jPointer lcSolverHandle(OpaqueLaunchContext* lc);
|
|
|
|
|
2020-01-04 11:27:50 +01:00
|
|
|
ND4J_EXPORT OpaqueDataBuffer* allocateDataBuffer(Nd4jLong elements, int dataType, bool allocateBoth);
|
2020-04-28 19:38:16 +02:00
|
|
|
ND4J_EXPORT OpaqueDataBuffer* dbAllocateDataBuffer(Nd4jLong elements, int dataType, bool allocateBoth);
|
|
|
|
ND4J_EXPORT OpaqueDataBuffer* dbCreateExternalDataBuffer(Nd4jLong elements, int dataType, Nd4jPointer primary, Nd4jPointer special);
|
2020-01-04 11:27:50 +01:00
|
|
|
ND4J_EXPORT OpaqueDataBuffer* dbCreateView(OpaqueDataBuffer *dataBuffer, Nd4jLong length, Nd4jLong offset);
|
|
|
|
ND4J_EXPORT Nd4jPointer dbPrimaryBuffer(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT Nd4jPointer dbSpecialBuffer(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbExpandBuffer(OpaqueDataBuffer *dataBuffer, Nd4jLong elements);
|
|
|
|
ND4J_EXPORT void dbAllocatePrimaryBuffer(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbAllocateSpecialBuffer(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbSetPrimaryBuffer(OpaqueDataBuffer *dataBuffer, Nd4jPointer primaryBuffer, Nd4jLong numBytes);
|
|
|
|
ND4J_EXPORT void dbSetSpecialBuffer(OpaqueDataBuffer *dataBuffer, Nd4jPointer specialBuffer, Nd4jLong numBytes);
|
|
|
|
ND4J_EXPORT void dbSyncToSpecial(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbSyncToPrimary(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT int dbLocality(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT int dbDeviceId(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbSetDeviceId(OpaqueDataBuffer *dataBuffer, int deviceId);
|
|
|
|
ND4J_EXPORT void dbTickHostRead(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbTickHostWrite(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbTickDeviceRead(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbTickDeviceWrite(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbClose(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void deleteDataBuffer(OpaqueDataBuffer *dataBuffer);
|
|
|
|
ND4J_EXPORT void dbExpand(OpaqueDataBuffer *dataBuffer, Nd4jLong elements);
|
|
|
|
|
2019-09-11 20:50:28 +02:00
|
|
|
|
|
|
|
ND4J_EXPORT int binaryLevel();
|
|
|
|
ND4J_EXPORT int optimalLevel();
|
|
|
|
|
|
|
|
ND4J_EXPORT bool isMinimalRequirementsMet();
|
|
|
|
ND4J_EXPORT bool isOptimalRequirementsMet();
|
|
|
|
|
2019-07-22 13:34:08 +02:00
|
|
|
}
|
2019-06-06 14:21:15 +02:00
|
|
|
|
|
|
|
#endif //NATIVEOPERATIONS_NATIVEOPS_H
|