Обратите внимание, что наш сайт не размещает какие-либо файлы из списка. Вы не можете скачать
эти файлы или скачать torrent-файл.
|
10 - 1 - Why it helps to combine models [13 min].mp4 |
15.12Мб |
10 - 2 - Mixtures of Experts [13 min].mp4 |
14.98Мб |
10 - 3 - The idea of full Bayesian learning [7 min].mp4 |
8.39Мб |
10 - 4 - Making full Bayesian learning practical [7 min].mp4 |
8.13Мб |
10 - 5 - Dropout [9 min].mp4 |
9.69Мб |
1 - 1 - Why do we need machine learning [13 min].mp4 |
15.05Мб |
1 - 2 - What are neural networks [8 min].mp4 |
9.76Мб |
1 - 3 - Some simple models of neurons [8 min].mp4 |
9.26Мб |
1 - 4 - A simple example of learning [6 min].mp4 |
6.57Мб |
1 - 5 - Three types of learning [8 min].mp4 |
8.96Мб |
2 - 1 - Types of neural network architectures [7 min].mp4 |
8.78Мб |
2 - 2 - Perceptrons The first generation of neural networks [8 min].mp4 |
9.39Мб |
2 - 3 - A geometrical view of perceptrons [6 min].mp4 |
7.32Мб |
2 - 4 - Why the learning works [5 min].mp4 |
5.90Мб |
2 - 5 - What perceptrons cant do [15 min].mp4 |
16.57Мб |
3 - 1 - Learning the weights of a linear neuron [12 min].mp4 |
13.52Мб |
3 - 2 - The error surface for a linear neuron [5 min].mp4 |
5.89Мб |
3 - 3 - Learning the weights of a logistic output neuron [4 min].mp4 |
4.37Мб |
3 - 4 - The backpropagation algorithm [12 min].mp4 |
13.35Мб |
3 - 5 - Using the derivatives computed by backpropagation [10 min].mp4 |
11.15Мб |
4 - 1 - Learning to predict the next word [13 min].mp4 |
14.28Мб |
4 - 2 - A brief diversion into cognitive science [4 min].mp4 |
5.31Мб |
4 - 3 - Another diversion The softmax output function [7 min].mp4 |
8.03Мб |
4 - 4 - Neuro-probabilistic language models [8 min].mp4 |
8.93Мб |
4 - 5 - Ways to deal with the large number of possible outputs [15 min].mp4 |
14.26Мб |
5 - 1 - Why object recognition is difficult [5 min].mp4 |
5.37Мб |
5 - 2 - Achieving viewpoint invariance [6 min].mp4 |
6.89Мб |
5 - 3 - Convolutional nets for digit recognition [16 min].mp4 |
18.46Мб |
5 - 4 - Convolutional nets for object recognition [17min].mp4 |
23.03Мб |
6 - 1 - Overview of mini-batch gradient descent.mp4 |
9.60Мб |
6 - 2 - A bag of tricks for mini-batch gradient descent.mp4 |
14.90Мб |
6 - 3 - The momentum method.mp4 |
9.74Мб |
6 - 4 - Adaptive learning rates for each connection.mp4 |
6.63Мб |
6 - 5 - Rmsprop Divide the gradient by a running average of its recent magnitude.mp4 |
15.12Мб |
7 - 1 - Modeling sequences A brief overview.mp4 |
20.13Мб |
7 - 2 - Training RNNs with back propagation.mp4 |
7.33Мб |
7 - 3 - A toy example of training an RNN.mp4 |
7.24Мб |
7 - 4 - Why it is difficult to train an RNN.mp4 |
8.89Мб |
7 - 5 - Long-term Short-term-memory.mp4 |
10.23Мб |
8 - 1 - A brief overview of Hessian Free optimization.mp4 |
16.24Мб |
8 - 2 - Modeling character strings with multiplicative connections [14 mins].mp4 |
16.56Мб |
8 - 3 - Learning to predict the next character using HF [12 mins].mp4 |
13.92Мб |
8 - 4 - Echo State Networks [9 min].mp4 |
11.28Мб |
9 - 1 - Overview of ways to improve generalization [12 min].mp4 |
13.57Мб |
9 - 2 - Limiting the size of the weights [6 min].mp4 |
7.36Мб |
9 - 3 - Using noise as a regularizer [7 min].mp4 |
8.48Мб |
9 - 4 - Introduction to the full Bayesian approach [12 min].mp4 |
12.00Мб |
9 - 5 - The Bayesian interpretation of weight decay [11 min].mp4 |
12.27Мб |
9 - 6 - MacKays quick and dirty method of setting weight costs [4 min].mp4 |
4.37Мб |