On Linux:
sudo mkdir /usr/local/mklml
sudo wget https://github.com/01org/mkl-dnn/releases/download/v0.11/mklml_lnx_2018.0.1.20171007.tgz
sudo tar -xzf mklml_lnx_2018.0.1.20171007.tgz -C /usr/local/mklml
On Windows:
Create a directory on your machine to hold MKLML, e.g. mkdir c:\local\mklml
Download the file [mklml_win_2018.0.1.20171007.zip](https://github.com/01org/mkl-dnn/releases/download/v0.11/mklml_win_2018.0.1.20171007.zip).
Unzip it into your MKLML path, creating a versioned sub directory within.
Set the environment variable `MKLML_PATH` to the versioned sub directory, e.g. setx MKLML_PATH c:\local\mklml\mklml_win_2018.0.1.20171007
This change also enables CPU convolution forward/backward using MKL, which leads to ~4x speedup in AlexNet training.
minor fixes
fixing Native utils path, loader, and the Native Load Utils
remove path and classpath setting in tests
cleaning up native utils
Unified CNTKNativeUtils class. Changed the code generation to use CNTKNativeUtils directly instead of the intermediary CNTK.init().
adding fixes to post bild
Added NATIVE_LOAD_MANIFEST for the cases when only specific high-level libraries should be loaded
linux side
Add gpu support
linux gpu .so copying
C#; and various fixes
Details:
fix wrong link
use /Zo option instead of undocumented /d2Zi+; remove /d2Zi+ for debug build;
move EvalMultithreads.cpp into Tests/EndToEndTests/EvalClientsTests/CNTKLibraryCPPEvalExamplesTests as preparation for the new CPP examples
add new CNTKLibraryCPPEvalExamples
use stdout
add first sample in CPPEvalExamples
add parallel evaluation examples
complete cpp examples
adapt linux build
add CPP tests to UWP
enable more tests on UWP
add running new tests; update baseline
add new cpp exampels to CNTKLibraryExamples.sln; use stdout instead of stderr in legacy eval dll; add missing cs examples
add file types for UWP
flush stdout to make cygwin happy
update baseline file for CPPUWPEval
update baseline for CNTKLibraryCSEvalExamplesTest
* Refactor index data structures and rewrite indexers (with most changes
in the text index builder).
* Add best effort caching: the cache is written out asynchronously in a
separate thread, on restart the index builder tries to restore the
index from cache (as long as the cache is not older than the input
file) and goes back no normal indexing if that fails (i.e., the cache
is corrupt).
* Refactor and simplify MemoryBuffer (renamed to BufferedFileReader).
* Use KMP patter-matching to simply sample counting with non-empty main
stream (num samples in sequence = number of lines that contain main
stream name).
* Refactor and simplify MLFIndexBuilder (it now also uses
BufferedFileReader)
* Use 512KB chunks when loading index from cache for faster reading.
* Add a number of unit tests for the indexing both with and without
caching.
adding some doc
use $$
add comments
automatically loadLibrary in java
semicolon
removing .iml
add more dependencies
move java static block code to cntk_java.i
remove also class files.
Adding .java files to jar
changing test location
typo
ignore DeviceDescriptorVector size constructor
DRY in Main.java
use List interface instead of ArrayList implementation
moving test location, expanding tests to gpu, fixing comments
move linux java tests
updating baseline.txt