Open Ristovski opened 7 years ago
you can make it work by yourself, just replace cblas code from mkl
void gemm(int TA, int TB, int M, int N, int K, float ALPHA, float A, int lda, float B, int ldb, float BETA, float *C, int ldc) {
cblas_sgemm(CblasRowMajor, TA?CblasTrans:CblasNoTrans, TB?CblasTrans:CblasNoTrans, M, N, K, ALPHA, A, lda, B, ldb, BETA, C, ldc);
gemm_cpu( TA, TB, M, N, K, ALPHA,A,lda, B, ldb,BETA,C,ldc);
}
Since we already have CUDA. Are there any plans to integrate Intel MKL/MKL-DNN support? Dlib does this, and it greatly speeds up the neural networks. I am aware this would not be a trivial task, but good to have nonetheless.