-
Notifications
You must be signed in to change notification settings - Fork 4.3k
CNTK FAQ
At this page we present the answers to some of the most common questions we get about CNTK and related Subjects. For questions and answers about problems you are seeing setting up or running CNTK, see the Common Pitfalls page.
CNTK, the Microsoft Cognitive Toolkit, is a framework for deep learning. A Computational Network defines the function to be learned as a directed graph where each leaf node consists of an input value or parameter, and each non-leaf node represents a matrix or tensor operation upon its children. The beauty of CNTK is that once a computational network has been described, all the computation required to learn the network parameters is taken care of automatically. There is no need to derive gradients analytically or to code the interactions between variables for backpropagation.
We first created CNTK for ourselves. CNTK was developed for the fastest training on the biggest data sets. Many of Microsoft's critical services run on models trained with CNTK. The results were so positive, we wanted to share our toolkit with the world.
For mission critical AI research, we believe efficiency and performance are important criteria. CNTK was designed for peak performance for not only CPUs but also single-GPU, multi-GPU, and multi-machine-multi-GPU scenarios. Additionally, Microsoft’s 1-bit compression technique or Block momentum technique dramatically reduced communication costs -- enabling highly scalable parallel training on a large number of GPUs spanning multiple machines.
In addition to a wide variety of built-in computation nodes, CNTK provides a plug-in architecture allowing users to define their own computation nodes. So if your workload requires special customization, CNTK makes that easy to do. Readers are also fully customizable allowing support for arbitrary input formats.
Today CNTK supports the following algorithms:
- Feed Forward
- CNN
- RNN
- LSTM
- Sequence-to-Sequence.
CNTK is developed by Microsoft's Technology and Research division. Additionally, CNTK gets major contributions from nearly all of Microsoft production teams.
The development of CNTK has been underway since late 2014.
No. CNTK is used in production for the Speech Recognition as well as for Image and Text training.
Using CNTK is easy and straightforward. Here are some ways to get started.
Unlike training, eval (as distributed through the Nuget package) is currently CPU only (e.g. on Windows it is built using the Release_CpuOnly target). If you need full GPU support, use the evaluation libraries built with the release target (e.g. Release on Windows) from the main distribution.
Doing so prevents the same samples from always appearing in a mini-batch together. This leads to improvements in the validation accuracy.
Yes. See the description at Understanding and Extending Readers and look for the section describing how to "compose several data deserializers"
When looking at dropout, does chosen hidden unit omitted for the entire minibatch as updates takes place only after minibatchSize?
Typically a different set of hidden units are set to 0 for different samples in the same minibatch. For recurrent neural networks, some people constrain to drop same set of hidden units across time for the same sequence.
Dropout documentation mentions about hidden units, does dropout apply to units in Convolutional Layer if there are multiple of them?
Convolution layer is also a hidden layer if it’s not the last output layer.
In CNTK, you need to explicitly indicate to use dropout. For example, if your original model has h2=W1*h1
and you want to apply dropout to h1
you need to change it to h2=W1*Dropout(h1)
. Currently CNTK only allows to specify one dropout rate for all dropout nodes used in the same model. However, it allows you to select different dropout rate across epochs.
BTW, we are not big fans of using different values for different dropout or model initial values. This would significantly increase the number of hyper parameters to be determined. If this is used in product, this means the model is lack of engineering stability.
Yes, you can compute anything based on weight and add it to your core criterion and use the combined criterion as your training objective.
Yes, you are able to read models trained by BrainScript.