/******************************************************************************* * Copyright (c) 2015-2018 Skymind, Inc. * * This program and the accompanying materials are made available under the * terms of the Apache License, Version 2.0 which is available at * https://www.apache.org/licenses/LICENSE-2.0. * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the * License for the specific language governing permissions and limitations * under the License. * * SPDX-License-Identifier: Apache-2.0 ******************************************************************************/ // // @author raver119@gmail.com // #ifndef PROJECT_SPECIALS_CUDA_H #define PROJECT_SPECIALS_CUDA_H #include #include #ifdef __CUDACC__ //////////////////////////////////////////////////////////////////////// template __host__ void bitonicSortStepGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int j, int k, int length, bool descending); //////////////////////////////////////////////////////////////////////// template __host__ void bitonicArbitraryStepGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int window, int length, int reverse, bool descending); //////////////////////////////////////////////////////////////////////// template __host__ void oesTadGeneric(dim3 &launchDims, cudaStream_t *stream, void *vx, Nd4jLong *xShapeInfo, int *dimension, int dimensionLength, Nd4jLong *tadShapeInfo, Nd4jLong *tadOffsets, bool descending); //////////////////////////////////////////////////////////////////////// template __global__ void printCudaGlobal(void* pointer, const int len) { for(int i = 0; i < len; ++i) printf("%f, ", (double)reinterpret_cast(pointer)[i] ); printf("\n"); } //////////////////////////////////////////////////////////////////////// template __device__ void printCudaDevice(void* pointer, const int len, const int tid = 0) { if(blockIdx.x * blockDim.x + threadIdx.x != tid) return; for(int i = 0; i < len; ++i) printf("%f, ", (double)reinterpret_cast(pointer)[i] ); printf("\n"); } //////////////////////////////////////////////////////////////////////// template __host__ void printCudaHost(void* pointer, const int len, cudaStream_t& stream) { void* ptr = malloc(sizeof(T)*len); cudaMemcpyAsync(ptr, pointer, sizeof(T)*len, cudaMemcpyDeviceToHost, stream); cudaError_t cudaResult = cudaStreamSynchronize(stream); if(cudaResult != 0) throw std::runtime_error("printCudaHost:: cudaStreamSynchronize failed!"); for(int i = 0; i < len; ++i) printf("%f, ", (double)reinterpret_cast(ptr)[i]); printf("\n"); free(ptr); } //////////////////////////////////////////////////////////////////////// __device__ inline int getDevicePosition(Nd4jLong *xShapeInfo, int index, Nd4jLong length) { int xEWS = shape::elementWiseStride(xShapeInfo); char order = shape::order(xShapeInfo); if (xEWS == 1 && order == 'c') { return index; } else if (xEWS > 1 && order == 'c') { return index * xEWS; } else { return shape::getIndexOffset(index, xShapeInfo, length); } } #endif #endif //PROJECT_SPECIALS_CUDA_H