100 lines
3.5 KiB
C++
100 lines
3.5 KiB
C++
/*******************************************************************************
|
|
* Copyright (c) 2015-2018 Skymind, Inc.
|
|
*
|
|
* This program and the accompanying materials are made available under the
|
|
* terms of the Apache License, Version 2.0 which is available at
|
|
* https://www.apache.org/licenses/LICENSE-2.0.
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
* License for the specific language governing permissions and limitations
|
|
* under the License.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
******************************************************************************/
|
|
|
|
//
|
|
// @author raver119@gmail.com
|
|
//
|
|
|
|
#ifndef PROJECT_SPECIALS_CUDA_H
|
|
#define PROJECT_SPECIALS_CUDA_H
|
|
|
|
#include <helpers/shape.h>
|
|
#include <helpers/DebugHelper.h>
|
|
|
|
#ifdef __CUDACC__
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__host__ void bitonicSortStepGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int j, int k, int length, bool descending);
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__host__ void bitonicArbitraryStepGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int window, int length, int reverse, bool descending);
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__host__ void oesTadGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int *dimension, int dimensionLength, Nd4jLong *tadShapeInfo, Nd4jLong *tadOffsets, bool descending);
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__global__ void printCudaGlobal(void* pointer, const int len) {
|
|
|
|
for(int i = 0; i < len; ++i)
|
|
printf("%f, ", (double)reinterpret_cast<T*>(pointer)[i] );
|
|
printf("\n");
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__device__ void printCudaDevice(void* pointer, const int len, const int tid = 0) {
|
|
|
|
if(blockIdx.x * blockDim.x + threadIdx.x != tid) return;
|
|
for(int i = 0; i < len; ++i)
|
|
printf("%f, ", (double)reinterpret_cast<T*>(pointer)[i] );
|
|
printf("\n");
|
|
}
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
template<typename T>
|
|
__host__ void printCudaHost(void* pointer, const int len, cudaStream_t& stream) {
|
|
|
|
void* ptr = malloc(sizeof(T)*len);
|
|
|
|
cudaMemcpyAsync(ptr, pointer, sizeof(T)*len, cudaMemcpyDeviceToHost, stream);
|
|
cudaError_t cudaResult = cudaStreamSynchronize(stream);
|
|
if(cudaResult != 0)
|
|
throw std::runtime_error("printCudaHost:: cudaStreamSynchronize failed!");
|
|
|
|
for(int i = 0; i < len; ++i)
|
|
printf("%f, ", (double)reinterpret_cast<T*>(ptr)[i]);
|
|
printf("\n");
|
|
|
|
free(ptr);
|
|
}
|
|
|
|
|
|
////////////////////////////////////////////////////////////////////////
|
|
__device__ inline int getDevicePosition(Nd4jLong *xShapeInfo, int index, Nd4jLong length) {
|
|
|
|
int xEWS = shape::elementWiseStride(xShapeInfo);
|
|
char order = shape::order(xShapeInfo);
|
|
|
|
if (xEWS == 1 && order == 'c') {
|
|
return index;
|
|
}
|
|
else if (xEWS > 1 && order == 'c') {
|
|
return index * xEWS;
|
|
}
|
|
else {
|
|
return shape::getIndexOffset(index, xShapeInfo, length);
|
|
}
|
|
}
|
|
|
|
|
|
#endif
|
|
|
|
#endif //PROJECT_SPECIALS_CUDA_H
|