cavis/libnd4j/include/ops/declarable/generic
raver119 7783012f39
cuDNN integration (#150)
* initial commit

Signed-off-by: raver119 <raver119@gmail.com>

* one file

Signed-off-by: raver119 <raver119@gmail.com>

* few more includes

Signed-off-by: raver119 <raver119@gmail.com>

* m?

Signed-off-by: raver119 <raver119@gmail.com>

* const

Signed-off-by: raver119 <raver119@gmail.com>

* cudnn linkage in tests

Signed-off-by: raver119 <raver119@gmail.com>

* culibos

Signed-off-by: raver119 <raver119@gmail.com>

* static reminder

Signed-off-by: raver119 <raver119@gmail.com>

* platform engine tag

Signed-off-by: raver119 <raver119@gmail.com>

* HAVE_CUDNN moved to config.h.in

Signed-off-by: raver119 <raver119@gmail.com>

* include

Signed-off-by: raver119 <raver119@gmail.com>

* include

Signed-off-by: raver119 <raver119@gmail.com>

* skip cudnn handle creation if there's not cudnn

Signed-off-by: raver119 <raver119@gmail.com>

* meh

Signed-off-by: raver119 <raver119@gmail.com>

* target device in context

Signed-off-by: raver119 <raver119@gmail.com>

* platform engines

Signed-off-by: raver119 <raver119@gmail.com>

* platform engines

Signed-off-by: raver119 <raver119@gmail.com>

* allow multiple -h args

Signed-off-by: raver119 <raver119@gmail.com>

* allow multiple -h args

Signed-off-by: raver119 <raver119@gmail.com>

* move mkldnn out of CPU block

Signed-off-by: raver119 <raver119@gmail.com>

* link to mkldnn on cuda

Signed-off-by: raver119 <raver119@gmail.com>

* less prints

Signed-off-by: raver119 <raver119@gmail.com>

* minor tweaks

Signed-off-by: raver119 <raver119@gmail.com>

* next step

Signed-off-by: raver119 <raver119@gmail.com>

* conv2d NCHW draft

Signed-off-by: raver119 <raver119@gmail.com>

* conv2d biasAdd

Signed-off-by: raver119 <raver119@gmail.com>

* test for MKL/CUDNN combined use

Signed-off-by: raver119 <raver119@gmail.com>

* - provide additional code for conv2d ff based on cudnn api, not tested yet

Signed-off-by: Yurii <iuriish@yahoo.com>

* - further work on conv2d helper based on using cudnn api

Signed-off-by: Yurii <iuriish@yahoo.com>

* - fixing several cuda bugs which appeared after cudnn lib had been started to use

Signed-off-by: Yurii <iuriish@yahoo.com>

* - implementation of conv2d backprop op based on cudnn api

Signed-off-by: Yurii <iuriish@yahoo.com>

* - implementaion of conv3d and conv3d_bp ops based on cudnn api

Signed-off-by: Yurii <iuriish@yahoo.com>

* - bugs fixing in conv3d/conv3d_bp ops (cudnn in use)

Signed-off-by: Yurii <iuriish@yahoo.com>

* - implementation of depthwiseConv2d (ff/bp) op based on cudnn api

Signed-off-by: Yurii <iuriish@yahoo.com>

* - implementation of batchnorm ff op based on cudnn api

Signed-off-by: Yurii <iuriish@yahoo.com>

* - disable cudnn batchnorm temporary

Signed-off-by: Yurii <iuriish@yahoo.com>

* - add minor change in cmake

Signed-off-by: Yurii <iuriish@yahoo.com>

* engine for depthwise mkldnn

Signed-off-by: raver119 <raver119@gmail.com>

* couple of includes

Signed-off-by: raver119 <raver119@gmail.com>

* - provide permutation to cudnn batchnorm ff when format is NHWC

Signed-off-by: Yurii <iuriish@yahoo.com>

* lgamma fix

Signed-off-by: raver119 <raver119@gmail.com>

* - eliminate memory leak in two tests

Signed-off-by: Yurii <iuriish@yahoo.com>

Co-authored-by: Yurii Shyrma <iuriish@yahoo.com>
2020-01-20 21:32:46 +03:00
..
activations Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
bitwise Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
blas Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
boolean Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
broadcastable Oleh powderev (#171) 2020-01-20 12:59:12 +03:00
compat String changes (#3) 2020-01-04 13:27:50 +03:00
datatypes String changes (#3) 2020-01-04 13:27:50 +03:00
flow Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
grad Eclipse Migration Initial Commit 2019-06-06 15:21:15 +03:00
helpers Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
images RgbToYuv & YuvToRgb skip empty arrays 2019-12-24 18:45:54 +03:00
kernels Merge pull request #7 from KonduitAI/asto_nd4s_10172019 2019-10-23 12:11:25 +03:00
list Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
loss Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
nlp [WIP] Error handling (#169) 2019-08-26 19:57:51 +03:00
nn cuDNN integration (#150) 2020-01-20 21:32:46 +03:00
parity_ops Shugeo lgamma (#170) 2020-01-20 12:29:36 +03:00
random Oleh multinomial (#163) 2020-01-06 22:35:05 +03:00
recurrent Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
shape negative handling for empty arrays (#146) 2019-12-24 13:23:25 +03:00
strings String changes (#3) 2020-01-04 13:27:50 +03:00
tests Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
thrid_party Shyrma temp (#131) 2019-12-20 22:35:39 +03:00
transforms minor build fix (#139) 2019-12-21 08:07:13 +03:00
tsne [WIP] More of CUDA (#95) 2019-08-05 11:27:05 +10:00
util String changes (#3) 2020-01-04 13:27:50 +03:00
CustomOperations.cpp Eclipse Migration Initial Commit 2019-06-06 15:21:15 +03:00
README.md Eclipse Migration Initial Commit 2019-06-06 15:21:15 +03:00
parity_ops.cpp Eclipse Migration Initial Commit 2019-06-06 15:21:15 +03:00

README.md

Generic ops

Put any cross platform operations here. Anything platform specific should be put in: in ops/helpers.