62ab074d2f | ||
---|---|---|
Dependencies/CNTKCustomMKL | ||
Documentation | ||
Examples | ||
Scripts | ||
Source | ||
Tests | ||
Tools | ||
bindings/python | ||
contrib | ||
.clang-format | ||
.gitattributes | ||
.gitignore | ||
.gitmodules | ||
CNTK.Cpp.props | ||
CNTK.sln | ||
CONTRIBUTING.md | ||
CppCntk.vssettings | ||
LICENSE.md | ||
Makefile | ||
README.md | ||
configure |
README.md
Latest news
2016-11-03. V 2.0 Beta 2 Release
Highlights of this Release:
- Feature tuning and bug fixing based on the feedback on Beta 1
- Changes in the Examples and Tutorials based on the same feedback
- New Tutorial on Reinforcement Learning
See more in the Release Notes
Get the Release from the CNTK Releases page
2016-10-25. New CNTK Name, new Web Site and V 2.0 Beta 1 Release
CNTK becomes The Microsoft Cognitive Toolkit. See more at our new Web Site.
With the today's Release we start delivering CNTK V2 - a major upgrade of Microsoft Cognitive Toolkit.
Expect a set of Beta Releases in the Coming Weeks.
Highlights of this Release:
- CNTK can now be used as a library with brand new C++ and Python APIs
- New Python Examples and Tutorials
- Support of Protocol Buffers serialization
- Support of Fast R-CNN algorithm
- New automated installation procedures
- Improvements in CNTK Evaluation library including support of CNTK APIs
See more in the Release Notes. You will find there links to the materials about the new features.
Get the Release from the CNTK Releases page
2016-10-03. V 1.7.2 Binary release
This is a Hot Fix Release. It affects all users of Model Evaluation Library
If you are NOT using Model Evaluation Library you may skip this release.
If you ARE using Model Evaluation Library we strongly recommend installing version 1.7.2 instead of any previous version you might be using.
See Release Notes for details.
2016-09-28. V 1.7.1 Binary release
Highlights of this Release:
- Two Breaking Changes related to Layers library default initialization and
fsAdagrad
gradient-normalization scheme - Improvements in BrainScript
- Enabling of Deterministic Algorithm enforcement
- Improvements in Model Evaluation including the support of Evaluation for Azure Applications
- Different Performance improvements
- Multiple bug fixes
See more in the Release Notes (including the full list of bugs fixed)
Get the Release from the CNTK Releases page
2016-08-31. V 1.7 Binary release
Highlights of this Release:
- Improvements in BrainScript (new library of predefined common layer types; common random-initialization types; GRUs)
- Support of NVIDIA cuDNN 5.1 and cuDNN RNN
- Improvements in readers and deserializers
- Additions to Evaluator library (Eval Client Sample, strong name for EvalWrapper)
- New unit tests incl. Linux support
- Python API Preview (since V.1.5)
- Multiple bug fixes
See more in the Release Notes
Get the Release from the CNTK Releases page
See all news.
What is CNTK
CNTK (http://www.cntk.ai/), the Computational Network Toolkit by Microsoft Research, is a unified deep-learning toolkit that describes neural networks as a series of computational steps via a directed graph. In this directed graph, leaf nodes represent input values or network parameters, while other nodes represent matrix operations upon their inputs. CNTK allows to easily realize and combine popular model types such as feed-forward DNNs, convolutional nets (CNNs), and recurrent networks (RNNs/LSTMs). It implements stochastic gradient descent (SGD, error backpropagation) learning with automatic differentiation and parallelization across multiple GPUs and servers. CNTK has been available under an open-source license since April 2015. It is our hope that the community will take advantage of CNTK to share ideas more quickly through the exchange of open source working code.
Wiki: Go to the CNTK Wiki for all information on CNTK including setup, examples, etc.
License: See LICENSE.md in the root of this repository for the full license information.
Tutorial: Microsoft Computational Network Toolkit (CNTK) @ NIPS 2015 Workshops
Blogs:
- Microsoft Computational Network Toolkit offers most efficient distributed deep learning computational performance
- Microsoft researchers win ImageNet computer vision challenge (December 2015)
Performance
The figure below compares processing speed (frames processed per second) of CNTK to that of four other well-known toolkits. The configuration uses a fully connected 4-layer neural network (see our benchmark scripts) and an effective mini batch size (8192). All results were obtained on the same hardware with the respective latest public software versions as of Dec 3, 2015.
Citation
If you used this toolkit or part of it to do your research, please cite the work as:
Amit Agarwal, Eldar Akchurin, Chris Basoglu, Guoguo Chen, Scott Cyphers, Jasha Droppo, Adam Eversole, Brian Guenter, Mark Hillebrand, T. Ryan Hoens, Xuedong Huang, Zhiheng Huang, Vladimir Ivanov, Alexey Kamenev, Philipp Kranen, Oleksii Kuchaiev, Wolfgang Manousek, Avner May, Bhaskar Mitra, Olivier Nano, Gaizka Navarro, Alexey Orlov, Hari Parthasarathi, Baolin Peng, Marko Radmilac, Alexey Reznichenko, Frank Seide, Michael L. Seltzer, Malcolm Slaney, Andreas Stolcke, Huaming Wang, Yongqiang Wang, Kaisheng Yao, Dong Yu, Yu Zhang, Geoffrey Zweig (in alphabetical order), "An Introduction to Computational Networks and the Computational Network Toolkit", Microsoft Technical Report MSR-TR-2014-112, 2014.
Disclaimer
CNTK is in active use at Microsoft and constantly evolving. There will be bugs.
Microsoft Open Source Code of Conduct
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.