|
Please note that this page does not hosts or makes available any of the listed filenames. You
cannot download any of those files from here.
|
| [TGx]Downloaded from torrentgalaxy.to .txt |
585B |
| 0 |
14B |
| 01__resources.html |
5.46KB |
| 01_actor-critic-with-softmax-policies.en.srt |
5.99KB |
| 01_actor-critic-with-softmax-policies.en.txt |
3.71KB |
| 01_actor-critic-with-softmax-policies.mp4 |
16.53MB |
| 01_agent-architecture-meeting-with-martha-overview-of-design-choices.en.srt |
10.80KB |
| 01_agent-architecture-meeting-with-martha-overview-of-design-choices.en.txt |
5.78KB |
| 01_agent-architecture-meeting-with-martha-overview-of-design-choices.mp4 |
15.62MB |
| 01_average-reward-a-new-way-of-formulating-control-problems.en.srt |
15.17KB |
| 01_average-reward-a-new-way-of-formulating-control-problems.en.txt |
9.43KB |
| 01_average-reward-a-new-way-of-formulating-control-problems.mp4 |
19.08MB |
| 01_bellman-equation-derivation.en.srt |
9.64KB |
| 01_bellman-equation-derivation.en.txt |
5.14KB |
| 01_bellman-equation-derivation.mp4 |
17.03MB |
| 01_congratulations.en.srt |
6.34KB |
| 01_congratulations.en.srt |
3.34KB |
| 01_congratulations.en.txt |
3.39KB |
| 01_congratulations.en.txt |
2.09KB |
| 01_congratulations.mp4 |
11.18MB |
| 01_congratulations.mp4 |
4.36MB |
| 01_congratulations-course-4-preview.en.srt |
4.22KB |
| 01_congratulations-course-4-preview.en.txt |
2.27KB |
| 01_congratulations-course-4-preview.mp4 |
22.11MB |
| 01_continuing-tasks.en.srt |
7.64KB |
| 01_continuing-tasks.en.txt |
3.99KB |
| 01_continuing-tasks.mp4 |
12.67MB |
| 01_course-3-introduction.en.srt |
8.91KB |
| 01_course-3-introduction.en.txt |
4.67KB |
| 01_course-3-introduction.mp4 |
16.33MB |
| 01_course-4-introduction.en.srt |
4.24KB |
| 01_course-4-introduction.en.txt |
2.29KB |
| 01_course-4-introduction.mp4 |
22.11MB |
| 01_course-introduction.en.srt |
4.00KB |
| 01_course-introduction.en.txt |
2.12KB |
| 01_course-introduction.mp4 |
11.27MB |
| 01_dynamic-programming_quiz.html |
157.49KB |
| 01_epsilon-soft-policies.en.srt |
7.55KB |
| 01_epsilon-soft-policies.en.txt |
4.77KB |
| 01_epsilon-soft-policies.mp4 |
12.69MB |
| 01_estimating-the-policy-gradient.en.srt |
7.48KB |
| 01_estimating-the-policy-gradient.en.txt |
4.63KB |
| 01_estimating-the-policy-gradient.mp4 |
13.63MB |
| 01_expected-sarsa.en.srt |
4.52KB |
| 01_expected-sarsa.en.txt |
2.80KB |
| 01_expected-sarsa.mp4 |
6.26MB |
| 01_exploration-under-function-approximation.en.srt |
6.53KB |
| 01_exploration-under-function-approximation.en.txt |
3.48KB |
| 01_exploration-under-function-approximation.mp4 |
11.05MB |
| 01_flexibility-of-the-policy-iteration-framework.en.srt |
7.08KB |
| 01_flexibility-of-the-policy-iteration-framework.en.txt |
3.80KB |
| 01_flexibility-of-the-policy-iteration-framework.mp4 |
12.44MB |
| 01_gradient-descent-for-training-neural-networks.en.srt |
13.98KB |
| 01_gradient-descent-for-training-neural-networks.en.txt |
7.45KB |
| 01_gradient-descent-for-training-neural-networks.mp4 |
15.53MB |
| 01_initial-project-meeting-with-martha-formalizing-the-problem.en.srt |
6.76KB |
| 01_initial-project-meeting-with-martha-formalizing-the-problem.en.txt |
4.22KB |
| 01_initial-project-meeting-with-martha-formalizing-the-problem.mp4 |
13.25MB |
| 01_learning-action-values.en.srt |
6.98KB |
| 01_learning-action-values.en.txt |
3.80KB |
| 01_learning-action-values.mp4 |
14.22MB |
| 01_lets-review-comparing-td-and-monte-carlo.en.srt |
8.10KB |
| 01_lets-review-comparing-td-and-monte-carlo.en.txt |
4.28KB |
| 01_lets-review-comparing-td-and-monte-carlo.mp4 |
9.81MB |
| 01_lets-review-expected-sarsa.en.srt |
4.52KB |
| 01_lets-review-expected-sarsa.en.txt |
2.80KB |
| 01_lets-review-expected-sarsa.mp4 |
6.26MB |
| 01_lets-review-markov-decision-processes.en.srt |
9.62KB |
| 01_lets-review-markov-decision-processes.en.txt |
5.18KB |
| 01_lets-review-markov-decision-processes.mp4 |
12.36MB |
| 01_lets-review-non-linear-approximation-with-neural-networks.en.srt |
6.10KB |
| 01_lets-review-non-linear-approximation-with-neural-networks.en.txt |
3.85KB |
| 01_lets-review-non-linear-approximation-with-neural-networks.mp4 |
9.59MB |
| 01_lets-review-optimization-strategies-for-nns.en.srt |
8.41KB |
| 01_lets-review-optimization-strategies-for-nns.en.txt |
4.52KB |
| 01_lets-review-optimization-strategies-for-nns.mp4 |
14.28MB |
| 01_mdps_quiz.html |
11.79KB |
| 01_meeting-with-adam-getting-the-agent-details-right.en.srt |
7.09KB |
| 01_meeting-with-adam-getting-the-agent-details-right.en.txt |
4.41KB |
| 01_meeting-with-adam-getting-the-agent-details-right.mp4 |
12.60MB |
| 01_meeting-with-adam-parameter-studies-in-rl.en.srt |
8.08KB |
| 01_meeting-with-adam-parameter-studies-in-rl.en.txt |
5.07KB |
| 01_meeting-with-adam-parameter-studies-in-rl.mp4 |
11.49MB |
| 01_meeting-with-martha-discussing-your-results.en.srt |
3.90KB |
| 01_meeting-with-martha-discussing-your-results.en.txt |
2.42KB |
| 01_meeting-with-martha-discussing-your-results.mp4 |
10.95MB |
| 01_meeting-with-niko-choosing-the-learning-algorithm.en.srt |
4.56KB |
| 01_meeting-with-niko-choosing-the-learning-algorithm.en.txt |
2.84KB |
| 01_meeting-with-niko-choosing-the-learning-algorithm.mp4 |
7.88MB |
| 01_module-1-learning-objectives_instructions.html |
2.80KB |
| 01_module-1-learning-objectives_instructions.html |
3.97KB |
| 01_module-1-learning-objectives_instructions.html |
3.02KB |
| 01_module-2-learning-objectives_instructions.html |
2.39KB |
| 01_module-2-learning-objectives_instructions.html |
3.09KB |
| 01_module-2-learning-objectives_instructions.html |
1.73KB |
| 01_module-3-learning-objectives_instructions.html |
3.20KB |
| 01_module-3-learning-objectives_instructions.html |
2.21KB |
| 01_module-3-learning-objectives_instructions.html |
2.84KB |
| 01_module-4-learning-objectives_instructions.html |
3.00KB |
| 01_module-4-learning-objectives_instructions.html |
3.53KB |
| 01_module-4-learning-objectives_instructions.html |
2.87KB |
| 01_optimal-policies.en.srt |
12.16KB |
| 01_optimal-policies.en.txt |
6.42KB |
| 01_optimal-policies.mp4 |
18.46MB |
| 01_policy-improvement.en.srt |
6.52KB |
| 01_policy-improvement.en.txt |
3.51KB |
| 01_policy-improvement.mp4 |
9.99MB |
| 01_practice-value-functions-and-bellman-equations_quiz.html |
7.98KB |
| 01_random-tabular-q-planning.en.srt |
5.38KB |
| 01_random-tabular-q-planning.en.txt |
2.93KB |
| 01_random-tabular-q-planning.mp4 |
7.83MB |
| 01_semi-gradient-td-for-policy-evaluation.en.srt |
4.57KB |
| 01_semi-gradient-td-for-policy-evaluation.en.txt |
2.86KB |
| 01_semi-gradient-td-for-policy-evaluation.mp4 |
15.35MB |
| 01_sequential-decision-making_quiz.html |
210.30KB |
| 01_specialization-introduction.en.srt |
4.92KB |
| 01_specialization-introduction.en.txt |
2.63KB |
| 01_specialization-introduction.mp4 |
18.26MB |
| 01_the-advantages-of-temporal-difference-learning.en.srt |
8.16KB |
| 01_the-advantages-of-temporal-difference-learning.en.txt |
4.30KB |
| 01_the-advantages-of-temporal-difference-learning.mp4 |
9.10MB |
| 01_the-dyna-architecture.en.srt |
6.93KB |
| 01_the-dyna-architecture.en.txt |
4.28KB |
| 01_the-dyna-architecture.mp4 |
9.59MB |
| 01_the-goal-of-reinforcement-learning.en.srt |
4.90KB |
| 01_the-goal-of-reinforcement-learning.en.txt |
2.62KB |
| 01_the-goal-of-reinforcement-learning.mp4 |
8.02MB |
| 01_the-linear-td-update.en.srt |
6.28KB |
| 01_the-linear-td-update.en.txt |
3.30KB |
| 01_the-linear-td-update.mp4 |
9.90MB |
| 01_the-objective-for-learning-policies.en.srt |
8.91KB |
| 01_the-objective-for-learning-policies.en.txt |
4.75KB |
| 01_the-objective-for-learning-policies.mp4 |
13.35MB |
| 01_the-value-error-objective.en.srt |
6.24KB |
| 01_the-value-error-objective.en.txt |
3.43KB |
| 01_the-value-error-objective.mp4 |
10.86MB |
| 01_using-monte-carlo-for-action-values.en.srt |
4.73KB |
| 01_using-monte-carlo-for-action-values.en.txt |
2.51KB |
| 01_using-monte-carlo-for-action-values.mp4 |
6.47MB |
| 01_what-if-the-model-is-inaccurate.en.srt |
7.00KB |
| 01_what-if-the-model-is-inaccurate.en.txt |
3.76KB |
| 01_what-if-the-model-is-inaccurate.mp4 |
7.69MB |
| 01_what-is-a-neural-network.en.srt |
5.50KB |
| 01_what-is-a-neural-network.en.txt |
2.96KB |
| 01_what-is-a-neural-network.mp4 |
7.03MB |
| 01_what-is-q-learning.en.srt |
4.95KB |
| 01_what-is-q-learning.en.txt |
2.60KB |
| 01_what-is-q-learning.mp4 |
7.84MB |
| 01_what-is-the-trade-off.en.srt |
12.17KB |
| 01_what-is-the-trade-off.en.txt |
6.57KB |
| 01_what-is-the-trade-off.mp4 |
21.58MB |
| 01_why-does-off-policy-learning-matter.en.srt |
5.93KB |
| 01_why-does-off-policy-learning-matter.en.txt |
3.82KB |
| 01_why-does-off-policy-learning-matter.mp4 |
14.39MB |
| 02_actor-critic-algorithm.en.srt |
9.18KB |
| 02_actor-critic-algorithm.en.txt |
4.91KB |
| 02_actor-critic-algorithm.mp4 |
14.07MB |
| 02 andy-barto-on-what-are-eligibility-traces-and-why-are-they-so-named en srt |
12.50KB |
| 02 andy-barto-on-what-are-eligibility-traces-and-why-are-they-so-named en txt |
6.62KB |
| 02_andy-barto-on-what-are-eligibility-traces-and-why-are-they-so-named.mp4 |
38.51MB |
| 02_bandits-and-exploration-exploitation_instructions.html |
1.13KB |
| 02_comparing-td-and-monte-carlo.en.srt |
8.10KB |
| 02_comparing-td-and-monte-carlo.en.txt |
4.28KB |
| 02_comparing-td-and-monte-carlo.mp4 |
9.81MB |
| 02_comparing-td-and-monte-carlo-with-state-aggregation.en.srt |
6.85KB |
| 02_comparing-td-and-monte-carlo-with-state-aggregation.en.txt |
3.61KB |
| 02_comparing-td-and-monte-carlo-with-state-aggregation.mp4 |
11.54MB |
| 02_course-introduction.en.srt |
10.43KB |
| 02_course-introduction.en.txt |
5.62KB |
| 02_course-introduction.mp4 |
32.39MB |
| 02_course-wrap-up.en.srt |
2.95KB |
| 02_course-wrap-up.en.txt |
1.83KB |
| 02_course-wrap-up.mp4 |
7.76MB |
| 02_demonstration-with-actor-critic.en.srt |
10.86KB |
| 02_demonstration-with-actor-critic.en.txt |
5.87KB |
| 02_demonstration-with-actor-critic.mp4 |
28.82MB |
| 02_drew-bagnell-on-system-id-optimal-control.en.srt |
10.52KB |
| 02_drew-bagnell-on-system-id-optimal-control.en.txt |
6.76KB |
| 02_drew-bagnell-on-system-id-optimal-control.mp4 |
31.29MB |
| 02_efficiency-of-dynamic-programming.en.srt |
7.72KB |
| 02_efficiency-of-dynamic-programming.en.txt |
4.84KB |
| 02_efficiency-of-dynamic-programming.mp4 |
14.03MB |
| 02_estimating-action-values-incrementally.en.srt |
8.05KB |
| 02_estimating-action-values-incrementally.en.txt |
4.30KB |
| 02_estimating-action-values-incrementally.mp4 |
19.40MB |
| 02_examples-of-episodic-and-continuing-tasks.en.srt |
4.66KB |
| 02_examples-of-episodic-and-continuing-tasks.en.txt |
2.52KB |
| 02_examples-of-episodic-and-continuing-tasks.mp4 |
9.14MB |
| 02_expected-sarsa-in-the-cliff-world.en.srt |
3.73KB |
| 02_expected-sarsa-in-the-cliff-world.en.txt |
2.31KB |
| 02_expected-sarsa-in-the-cliff-world.mp4 |
5.69MB |
| 02_graded-assignment-describe-three-mdps_peer_assignment_instructions.html |
2.33KB |
| 02_graded-value-functions-and-bellman-equations_exam.html |
31.06KB |
| 02_importance-sampling.en.srt |
6.58KB |
| 02_importance-sampling.en.txt |
3.47KB |
| 02_importance-sampling.mp4 |
7.41MB |
| 02_in-depth-with-changing-environments.en.srt |
9.22KB |
| 02_in-depth-with-changing-environments.en.txt |
4.92KB |
| 02_in-depth-with-changing-environments.mp4 |
11.94MB |
| 02_introducing-gradient-descent.en.srt |
9.91KB |
| 02_introducing-gradient-descent.en.txt |
6.15KB |
| 02_introducing-gradient-descent.mp4 |
15.10MB |
| 02_joelle-pineau-about-rl-that-matters.en.srt |
13.67KB |
| 02_joelle-pineau-about-rl-that-matters.en.txt |
8.75KB |
| 02_joelle-pineau-about-rl-that-matters.mp4 |
29.50MB |
| 02_lets-review-examples-of-episodic-and-continuing-tasks.en.srt |
4.66KB |
| 02_lets-review-examples-of-episodic-and-continuing-tasks.en.txt |
2.52KB |
| 02_lets-review-examples-of-episodic-and-continuing-tasks.mp4 |
9.14MB |
| 02_lets-review-expected-sarsa-with-function-approximation.en.srt |
3.93KB |
| 02_lets-review-expected-sarsa-with-function-approximation.en.txt |
2.08KB |
| 02_lets-review-expected-sarsa-with-function-approximation.mp4 |
7.63MB |
| 02_lets-review-what-is-q-learning.en.srt |
4.95KB |
| 02_lets-review-what-is-q-learning.en.txt |
2.60KB |
| 02_lets-review-what-is-q-learning.mp4 |
7.84MB |
| 02_meet-your-instructors.en.srt |
13.43KB |
| 02_meet-your-instructors.en.srt |
13.43KB |
| 02_meet-your-instructors.en.srt |
13.43KB |
| 02_meet-your-instructors.en.txt |
8.62KB |
| 02_meet-your-instructors.en.txt |
8.62KB |
| 02_meet-your-instructors.en.txt |
8.62KB |
| 02_meet-your-instructors.mp4 |
43.87MB |
| 02_meet-your-instructors.mp4 |
43.87MB |
| 02_meet-your-instructors.mp4 |
43.87MB |
| 02_michael-littman-the-reward-hypothesis.en.srt |
18.48KB |
| 02_michael-littman-the-reward-hypothesis.en.txt |
11.58KB |
| 02_michael-littman-the-reward-hypothesis.mp4 |
84.01MB |
| 02_non-linear-approximation-with-neural-networks.en.srt |
6.10KB |
| 02_non-linear-approximation-with-neural-networks.en.txt |
3.85KB |
| 02_non-linear-approximation-with-neural-networks.mp4 |
9.59MB |
| 02_optimal-policies-with-dynamic-programming_instructions.html |
1.13KB |
| 02_optimal-value-functions.en.srt |
8.34KB |
| 02_optimal-value-functions.en.txt |
4.51KB |
| 02_optimal-value-functions.mp4 |
10.19MB |
| 02_optimistic-initial-values.en.srt |
8.50KB |
| 02_optimistic-initial-values.en.txt |
5.36KB |
| 02_optimistic-initial-values.mp4 |
13.13MB |
| 02_optimization-strategies-for-nns.en.srt |
8.41KB |
| 02_optimization-strategies-for-nns.en.txt |
4.52KB |
| 02_optimization-strategies-for-nns.mp4 |
14.28MB |
| 02_policy-iteration.en.srt |
13.33KB |
| 02_policy-iteration.en.txt |
7.12KB |
| 02_policy-iteration.mp4 |
17.86MB |
| 02_q-learning-in-the-windy-grid-world.en.srt |
5.78KB |
| 02_q-learning-in-the-windy-grid-world.en.txt |
3.03KB |
| 02_q-learning-in-the-windy-grid-world.mp4 |
7.24MB |
| 02_satinder-singh-on-intrinsic-rewards.en.srt |
20.96KB |
| 02_satinder-singh-on-intrinsic-rewards.en.txt |
10.98KB |
| 02_satinder-singh-on-intrinsic-rewards.mp4 |
26.91MB |
| 02_the-dyna-algorithm.en.srt |
7.81KB |
| 02_the-dyna-algorithm.en.txt |
4.16KB |
| 02_the-dyna-algorithm.mp4 |
11.24MB |
| 02_the-policy-gradient-theorem.en.srt |
9.28KB |
| 02_the-policy-gradient-theorem.en.txt |
4.91KB |
| 02_the-policy-gradient-theorem.mp4 |
9.31MB |
| 02_the-true-objective-for-td.en.srt |
8.23KB |
| 02_the-true-objective-for-td.en.txt |
4.31KB |
| 02_the-true-objective-for-td.mp4 |
13.66MB |
| 02_using-monte-carlo-methods-for-generalized-policy-iteration.en.srt |
4.02KB |
| 02_using-monte-carlo-methods-for-generalized-policy-iteration.en.txt |
2.11KB |
| 02_using-monte-carlo-methods-for-generalized-policy-iteration.mp4 |
5.17MB |
| 02_weekly-reading_instructions.html |
1.16KB |
| 02_weekly-reading_instructions.html |
1.17KB |
| 02_weekly-reading_instructions.html |
1.16KB |
| 02_weekly-reading_instructions.html |
1.17KB |
| 02_weekly-reading_instructions.html |
1.19KB |
| 02_weekly-reading_instructions.html |
1.17KB |
| 02_weekly-reading_instructions.html |
1.17KB |
| 02_weekly-reading_instructions.html |
1.16KB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading-on-policy-control-with-approximation_instructions.html |
1.27KB |
| 02_weekly-reading-on-policy-control-with-approximation_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading-on-policy-prediction-with-approximation_instructions.html |
1.17KB |
| 02_weekly-reading-on-policy-prediction-with-approximation_RLbook2018.pdf |
85.28MB |
| 02 weekly-reading-on-policy-prediction-with-approximation-ii instructions html |
1.24KB |
| 02_weekly-reading-on-policy-prediction-with-approximation-ii_RLbook2018.pdf |
85.28MB |
| 02_weekly-reading-policy-gradient-methods_instructions.html |
1.17KB |
| 02_weekly-reading-policy-gradient-methods_RLbook2018.pdf |
85.28MB |
| 02_why-bellman-equations.en.srt |
6.99KB |
| 02_why-bellman-equations.en.txt |
4.41KB |
| 02_why-bellman-equations.mp4 |
11.87MB |
| 03_andy-barto-and-rich-sutton-more-on-the-history-of-rl.en.srt |
15.92KB |
| 03_andy-barto-and-rich-sutton-more-on-the-history-of-rl.en.txt |
8.07KB |
| 03_andy-barto-and-rich-sutton-more-on-the-history-of-rl.mp4 |
80.21MB |
| 03_coarse-coding.en.srt |
4.86KB |
| 03_coarse-coding.en.txt |
3.01KB |
| 03_coarse-coding.mp4 |
9.59MB |
| 03_david-silver-on-deep-learning-rl-ai.en.srt |
14.71KB |
| 03_david-silver-on-deep-learning-rl-ai.en.txt |
9.52KB |
| 03_david-silver-on-deep-learning-rl-ai.mp4 |
41.41MB |
| 03_deep-neural-networks.en.srt |
5.88KB |
| 03_deep-neural-networks.en.txt |
3.17KB |
| 03_deep-neural-networks.mp4 |
15.33MB |
| 03 doina-precup-building-knowledge-for-ai-agents-with-reinforcement-learning en srt |
11.34KB |
| 03 doina-precup-building-knowledge-for-ai-agents-with-reinforcement-learning en txt |
6.08KB |
| 03 doina-precup-building-knowledge-for-ai-agents-with-reinforcement-learning mp4 |
55.29MB |
| 03_drew-bagnell-self-driving-robotics-and-model-based-rl.en.srt |
10.32KB |
| 03_drew-bagnell-self-driving-robotics-and-model-based-rl.en.txt |
6.66KB |
| 03_drew-bagnell-self-driving-robotics-and-model-based-rl.mp4 |
35.21MB |
| 03_dyna-q-learning-in-a-simple-maze.en.srt |
6.90KB |
| 03_dyna-q-learning-in-a-simple-maze.en.txt |
4.19KB |
| 03_dyna-q-learning-in-a-simple-maze.mp4 |
10.76MB |
| 03_episodic-sarsa-with-function-approximation.en.srt |
6.23KB |
| 03_episodic-sarsa-with-function-approximation.en.txt |
3.85KB |
| 03_episodic-sarsa-with-function-approximation.mp4 |
18.05MB |
| 03_gaussian-policies-for-continuous-actions.en.srt |
12.82KB |
| 03_gaussian-policies-for-continuous-actions.en.txt |
6.94KB |
| 03_gaussian-policies-for-continuous-actions.mp4 |
19.95MB |
| 03_generality-of-expected-sarsa.en.srt |
2.88KB |
| 03_generality-of-expected-sarsa.en.txt |
1.52KB |
| 03_generality-of-expected-sarsa.mp4 |
5.21MB |
| 03_gradient-monte-for-policy-evaluation.en.srt |
9.31KB |
| 03_gradient-monte-for-policy-evaluation.en.txt |
4.90KB |
| 03_gradient-monte-for-policy-evaluation.mp4 |
15.24MB |
| 03_how-is-q-learning-off-policy.en.srt |
7.23KB |
| 03_how-is-q-learning-off-policy.en.txt |
3.99KB |
| 03_how-is-q-learning-off-policy.mp4 |
9.96MB |
| 03_learning-policies-directly.en.srt |
10.17KB |
| 03_learning-policies-directly.en.txt |
5.42KB |
| 03_learning-policies-directly.mp4 |
17.10MB |
| 03 lets-review-average-reward-a-new-way-of-formulating-control-problems en srt |
15.17KB |
| 03 lets-review-average-reward-a-new-way-of-formulating-control-problems en txt |
9.43KB |
| 03_lets-review-average-reward-a-new-way-of-formulating-control-problems.mp4 |
19.08MB |
| 03_lets-review-dyna-q-learning-in-a-simple-maze.en.srt |
6.90KB |
| 03_lets-review-dyna-q-learning-in-a-simple-maze.en.txt |
4.19KB |
| 03_lets-review-dyna-q-learning-in-a-simple-maze.mp4 |
10.76MB |
| 03_markov-decision-processes.en.srt |
9.62KB |
| 03_markov-decision-processes.en.txt |
5.18KB |
| 03_markov-decision-processes.mp4 |
12.36MB |
| 03_meet-your-instructors.en.srt |
15.89KB |
| 03_meet-your-instructors.en.txt |
8.41KB |
| 03_meet-your-instructors.mp4 |
43.87MB |
| 03_moving-to-parameterized-functions.en.srt |
10.44KB |
| 03_moving-to-parameterized-functions.en.txt |
5.59KB |
| 03_moving-to-parameterized-functions.mp4 |
24.38MB |
| 03_off-policy-monte-carlo-prediction.en.srt |
7.80KB |
| 03_off-policy-monte-carlo-prediction.en.txt |
4.15KB |
| 03_off-policy-monte-carlo-prediction.mp4 |
12.52MB |
| 03_policy-evaluation-vs-control.en.srt |
6.66KB |
| 03_policy-evaluation-vs-control.en.txt |
4.18KB |
| 03_policy-evaluation-vs-control.mp4 |
13.32MB |
| 03_read-me-pre-requisites-and-learning-objectives_instructions.html |
3.19KB |
| 03 read-me-pre-requisites-and-learning-objectives Prediction and Control with Function Approximation Learning Objectives pdf |
59.93KB |
| 03_reinforcement-learning-textbook_instructions.html |
2.19KB |
| 03_reinforcement-learning-textbook_instructions.html |
2.19KB |
| 03_reinforcement-learning-textbook_RLbook2018.pdf |
85.28MB |
| 03_reinforcement-learning-textbook_RLbook2018.pdf |
85.28MB |
| 03_sarsa-gpi-with-td.en.srt |
6.11KB |
| 03_sarsa-gpi-with-td.en.txt |
3.23KB |
| 03_sarsa-gpi-with-td.mp4 |
7.38MB |
| 03_sequential-decision-making-with-evaluative-feedback.en.srt |
8.71KB |
| 03_sequential-decision-making-with-evaluative-feedback.en.txt |
4.65KB |
| 03_sequential-decision-making-with-evaluative-feedback.mp4 |
16.27MB |
| 03_solving-the-blackjack-example.en.srt |
6.49KB |
| 03_solving-the-blackjack-example.en.txt |
3.42KB |
| 03_solving-the-blackjack-example.mp4 |
13.91MB |
| 03_specialization-wrap-up.en.srt |
5.41KB |
| 03_specialization-wrap-up.en.txt |
3.39KB |
| 03_specialization-wrap-up.mp4 |
18.62MB |
| 03_specifying-policies.en.srt |
7.52KB |
| 03_specifying-policies.en.txt |
4.01KB |
| 03_specifying-policies.mp4 |
14.99MB |
| 03_susan-murphy-on-rl-in-mobile-health.en.srt |
10.39KB |
| 03_susan-murphy-on-rl-in-mobile-health.en.txt |
6.32KB |
| 03_susan-murphy-on-rl-in-mobile-health.mp4 |
27.63MB |
| 03_upper-confidence-bound-ucb-action-selection.en.srt |
7.54KB |
| 03_upper-confidence-bound-ucb-action-selection.en.txt |
4.03KB |
| 03_upper-confidence-bound-ucb-action-selection.mp4 |
11.77MB |
| 03_using-optimal-value-functions-to-get-optimal-policies.en.srt |
10.83KB |
| 03_using-optimal-value-functions-to-get-optimal-policies.en.txt |
6.69KB |
| 03_using-optimal-value-functions-to-get-optimal-policies.mp4 |
16.73MB |
| 03 warren-powell-approximate-dynamic-programming-for-fleet-management-short en srt |
12.06KB |
| 03 warren-powell-approximate-dynamic-programming-for-fleet-management-short en txt |
7.52KB |
| 03 warren-powell-approximate-dynamic-programming-for-fleet-management-short mp4 |
47.13MB |
| 03_week-1-summary.en.srt |
6.69KB |
| 03_week-1-summary.en.txt |
3.58KB |
| 03_week-1-summary.mp4 |
16.31MB |
| 03_week-2-summary.en.srt |
2.77KB |
| 03_week-2-summary.en.txt |
1.46KB |
| 03_week-2-summary.mp4 |
5.42MB |
| 03_week-3-review.en.srt |
4.65KB |
| 03_week-3-review.en.txt |
2.47KB |
| 03_week-3-review.mp4 |
8.88MB |
| 03_what-is-a-model.en.srt |
7.53KB |
| 03_what-is-a-model.en.txt |
3.98KB |
| 03_what-is-a-model.mp4 |
11.33MB |
| 03_what-is-monte-carlo.en.srt |
10.50KB |
| 03_what-is-monte-carlo.en.txt |
5.65KB |
| 03_what-is-monte-carlo.mp4 |
14.88MB |
| 03_what-is-temporal-difference-td-learning.en.srt |
7.77KB |
| 03_what-is-temporal-difference-td-learning.en.txt |
4.11KB |
| 03_what-is-temporal-difference-td-learning.mp4 |
10.31MB |
| 04_advantages-of-policy-parameterization.en.srt |
7.66KB |
| 04_advantages-of-policy-parameterization.en.txt |
4.72KB |
| 04_advantages-of-policy-parameterization.mp4 |
26.06MB |
| 04_comparing-sample-and-distribution-models.en.srt |
3.87KB |
| 04_comparing-sample-and-distribution-models.en.txt |
2.09KB |
| 04_comparing-sample-and-distribution-models.mp4 |
6.65MB |
| 04_emma-brunskill-batch-reinforcement-learning.en.srt |
24.91KB |
| 04_emma-brunskill-batch-reinforcement-learning.en.txt |
13.16KB |
| 04_emma-brunskill-batch-reinforcement-learning.mp4 |
37.38MB |
| 04_episodic-sarsa-in-mountain-car.en.srt |
8.68KB |
| 04_episodic-sarsa-in-mountain-car.en.txt |
4.65KB |
| 04_episodic-sarsa-in-mountain-car.mp4 |
15.47MB |
| 04_examples-of-mdps.en.srt |
6.85KB |
| 04_examples-of-mdps.en.txt |
3.70KB |
| 04_examples-of-mdps.mp4 |
12.20MB |
| 04_generalization-and-discrimination.en.srt |
8.69KB |
| 04_generalization-and-discrimination.en.txt |
4.63KB |
| 04_generalization-and-discrimination.mp4 |
12.86MB |
| 04_generalization-properties-of-coarse-coding.en.srt |
9.38KB |
| 04_generalization-properties-of-coarse-coding.en.txt |
5.04KB |
| 04_generalization-properties-of-coarse-coding.mp4 |
17.98MB |
| 04_iterative-policy-evaluation.en.srt |
13.68KB |
| 04_iterative-policy-evaluation.en.txt |
7.15KB |
| 04_iterative-policy-evaluation.mp4 |
18.79MB |
| 04 jonathan-langford-contextual-bandits-for-real-world-reinforcement-learning en srt |
14.04KB |
| 04 jonathan-langford-contextual-bandits-for-real-world-reinforcement-learning en txt |
7.34KB |
| 04 jonathan-langford-contextual-bandits-for-real-world-reinforcement-learning mp4 |
11.94MB |
| 04_lets-review-actor-critic-algorithm.en.srt |
9.18KB |
| 04_lets-review-actor-critic-algorithm.en.txt |
4.91KB |
| 04_lets-review-actor-critic-algorithm.mp4 |
14.07MB |
| 04_meeting-with-martha-in-depth-on-experience-replay.en.srt |
7.36KB |
| 04_meeting-with-martha-in-depth-on-experience-replay.en.txt |
4.65KB |
| 04_meeting-with-martha-in-depth-on-experience-replay.mp4 |
21.42MB |
| 04 pre-requisites-and-learning-objectives A Complete Reinforcement Learning System Capstone Learning Objectives pdf |
56.79KB |
| 04_pre-requisites-and-learning-objectives_instructions.html |
3.59KB |
| 04 read-me-pre-requisites-and-learning-objectives Course 2 Sample Based Learning Methods Learning Objectives pdf |
83.14KB |
| 04_read-me-pre-requisites-and-learning-objectives_instructions.html |
2.96KB |
| 04_reinforcement-learning-textbook_instructions.html |
2.19KB |
| 04_reinforcement-learning-textbook_RLbook2018.pdf |
85.28MB |
| 04_rich-sutton-the-importance-of-td-learning.en.srt |
11.24KB |
| 04_rich-sutton-the-importance-of-td-learning.en.txt |
5.88KB |
| 04_rich-sutton-the-importance-of-td-learning.mp4 |
35.65MB |
| 04_sarsa-in-the-windy-grid-world.en.srt |
3.89KB |
| 04_sarsa-in-the-windy-grid-world.en.txt |
2.37KB |
| 04_sarsa-in-the-windy-grid-world.mp4 |
5.85MB |
| 04_state-aggregation-with-monte-carlo.en.srt |
10.23KB |
| 04_state-aggregation-with-monte-carlo.en.txt |
6.24KB |
| 04_state-aggregation-with-monte-carlo.mp4 |
20.26MB |
| 04_using-monte-carlo-for-prediction.en.srt |
10.60KB |
| 04_using-monte-carlo-for-prediction.en.txt |
5.61KB |
| 04_using-monte-carlo-for-prediction.mp4 |
16.17MB |
| 04_value-functions.en.srt |
10.34KB |
| 04_value-functions.en.txt |
5.53KB |
| 04_value-functions.mp4 |
21.10MB |
| 04 warren-powell-approximate-dynamic-programming-for-fleet-management-long en srt |
40.71KB |
| 04 warren-powell-approximate-dynamic-programming-for-fleet-management-long en txt |
21.34KB |
| 04 warren-powell-approximate-dynamic-programming-for-fleet-management-long mp4 |
145.35MB |
| 04_week-2-review.en.srt |
4.07KB |
| 04_week-2-review.en.txt |
2.16KB |
| 04_week-2-review.mp4 |
8.50MB |
| 04_week-2-summary.en.srt |
3.13KB |
| 04_week-2-summary.en.txt |
1.68KB |
| 04_week-2-summary.mp4 |
7.41MB |
| 04_week-3-summary.en.srt |
6.38KB |
| 04_week-3-summary.en.srt |
2.65KB |
| 04_week-3-summary.en.txt |
3.37KB |
| 04_week-3-summary.en.txt |
1.60KB |
| 04_week-3-summary.mp4 |
11.95MB |
| 04_week-3-summary.mp4 |
3.68MB |
| 04_week-4-summary.en.srt |
7.00KB |
| 04_week-4-summary.en.srt |
2.58KB |
| 04_week-4-summary.en.txt |
3.65KB |
| 04_week-4-summary.en.txt |
1.36KB |
| 04_week-4-summary.mp4 |
9.96MB |
| 04_week-4-summary.mp4 |
4.25MB |
| 04_your-specialization-roadmap.en.srt |
6.64KB |
| 04_your-specialization-roadmap.en.txt |
3.46KB |
| 04_your-specialization-roadmap.mp4 |
14.88MB |
| 05_chapter-summary_instructions.html |
1.14KB |
| 05_chapter-summary_instructions.html |
1.23KB |
| 05_chapter-summary_instructions.html |
1.17KB |
| 05_chapter-summary_RLbook2018.pdf |
85.28MB |
| 05_chapter-summary_RLbook2018.pdf |
85.28MB |
| 05_chapter-summary_RLbook2018.pdf |
85.28MB |
| 05_csaba-szepesvari-on-problem-landscape.en.srt |
9.57KB |
| 05_csaba-szepesvari-on-problem-landscape.en.txt |
6.06KB |
| 05_csaba-szepesvari-on-problem-landscape.mp4 |
38.81MB |
| 05_expected-sarsa-with-function-approximation.en.srt |
3.93KB |
| 05_expected-sarsa-with-function-approximation.en.txt |
2.08KB |
| 05_expected-sarsa-with-function-approximation.mp4 |
7.63MB |
| 05_framing-value-estimation-as-supervised-learning.en.srt |
6.26KB |
| 05_framing-value-estimation-as-supervised-learning.en.txt |
3.35KB |
| 05_framing-value-estimation-as-supervised-learning.mp4 |
10.69MB |
| 05 martin-riedmiller-on-the-collect-and-infer-framework-for-data-efficient-rl en srt |
8.17KB |
| 05 martin-riedmiller-on-the-collect-and-infer-framework-for-data-efficient-rl en txt |
5.12KB |
| 05 martin-riedmiller-on-the-collect-and-infer-framework-for-data-efficient-rl mp4 |
23.54MB |
| 05_reinforcement-learning-textbook_instructions.html |
2.19KB |
| 05_reinforcement-learning-textbook_RLbook2018.pdf |
85.28MB |
| 05_rich-sutton-and-andy-barto-a-brief-history-of-rl.en.srt |
10.49KB |
| 05_rich-sutton-and-andy-barto-a-brief-history-of-rl.en.txt |
5.40KB |
| 05_rich-sutton-and-andy-barto-a-brief-history-of-rl.mp4 |
48.75MB |
| 05_tile-coding.en.srt |
5.18KB |
| 05_tile-coding.en.txt |
2.79KB |
| 05_tile-coding.mp4 |
7.57MB |
| 05_week-1-summary.en.srt |
5.60KB |
| 05_week-1-summary.en.srt |
4.33KB |
| 05_week-1-summary.en.txt |
2.68KB |
| 05_week-1-summary.en.txt |
2.95KB |
| 05_week-1-summary.mp4 |
9.59MB |
| 05_week-1-summary.mp4 |
9.48MB |
| 05_week-4-summary.en.srt |
4.48KB |
| 05_week-4-summary.en.txt |
2.37KB |
| 05_week-4-summary.mp4 |
9.61MB |
| 06_andy-and-rich-advice-for-students.en.srt |
5.84KB |
| 06_andy-and-rich-advice-for-students.en.txt |
3.52KB |
| 06_andy-and-rich-advice-for-students.mp4 |
33.39MB |
| 06_chapter-summary_instructions.html |
1.19KB |
| 06_chapter-summary_instructions.html |
1.18KB |
| 06_chapter-summary_instructions.html |
1.19KB |
| 06_chapter-summary_RLbook2018.pdf |
85.28MB |
| 06_chapter-summary_RLbook2018.pdf |
85.28MB |
| 06_chapter-summary_RLbook2018.pdf |
85.28MB |
| 06 read-me-pre-requisites-and-learning-objectives Fundamentals of Reinforcement Learning Learning Objectives pdf |
64.66KB |
| 06_read-me-pre-requisites-and-learning-objectives_instructions.html |
2.63KB |
| 06_text-book-part-1-summary_instructions.html |
1.21KB |
| 06_text-book-part-1-summary_RLbook2018.pdf |
85.28MB |
| 06_using-tile-coding-in-td.en.srt |
8.31KB |
| 06_using-tile-coding-in-td.en.txt |
4.31KB |
| 06_using-tile-coding-in-td.mp4 |
23.07MB |
| 1 |
3B |
| 10 |
734.24KB |
| 100 |
319.14KB |
| 101 |
338.31KB |
| 102 |
410.31KB |
| 103 |
496.29KB |
| 104 |
569.91KB |
| 105 |
659.13KB |
| 106 |
659.13KB |
| 107 |
815.04KB |
| 108 |
55.29KB |
| 109 |
58.51KB |
| 11 |
734.24KB |
| 110 |
65.82KB |
| 111 |
131.76KB |
| 112 |
234.13KB |
| 113 |
466.17KB |
| 114 |
525.13KB |
| 115 |
685.62KB |
| 116 |
751.55KB |
| 117 |
780.48KB |
| 118 |
842.17KB |
| 119 |
977.52KB |
| 12 |
734.24KB |
| 120 |
47.98KB |
| 121 |
142.95KB |
| 122 |
248.88KB |
| 123 |
248.88KB |
| 124 |
315.91KB |
| 125 |
709.15KB |
| 126 |
825.30KB |
| 127 |
8.25KB |
| 128 |
42.04KB |
| 129 |
43.86KB |
| 13 |
734.24KB |
| 130 |
102.24KB |
| 131 |
192.93KB |
| 132 |
192.93KB |
| 133 |
396.76KB |
| 134 |
415.26KB |
| 135 |
416.35KB |
| 136 |
422.79KB |
| 137 |
422.93KB |
| 138 |
422.93KB |
| 139 |
534.71KB |
| 14 |
734.24KB |
| 140 |
703.79KB |
| 141 |
880.61KB |
| 142 |
880.61KB |
| 143 |
926.56KB |
| 144 |
125.93KB |
| 145 |
513.30KB |
| 146 |
1004.89KB |
| 147 |
122.88KB |
| 148 |
166.56KB |
| 149 |
166.56KB |
| 15 |
734.24KB |
| 150 |
177.78KB |
| 151 |
246.55KB |
| 152 |
315.86KB |
| 153 |
379.72KB |
| 154 |
379.72KB |
| 155 |
442.92KB |
| 156 |
602.73KB |
| 157 |
607.42KB |
| 158 |
633.69KB |
| 159 |
779.60KB |
| 16 |
734.24KB |
| 160 |
997.81KB |
| 161 |
363.22KB |
| 162 |
544.45KB |
| 163 |
752.70KB |
| 164 |
752.70KB |
| 165 |
152.05KB |
| 166 |
318.59KB |
| 167 |
592.56KB |
| 168 |
805.56KB |
| 169 |
847.62KB |
| 17 |
734.24KB |
| 170 |
658.17KB |
| 171 |
764.29KB |
| 18 |
734.24KB |
| 19 |
734.24KB |
| 2 |
275B |
| 20 |
734.24KB |
| 21 |
734.24KB |
| 22 |
734.24KB |
| 23 |
734.24KB |
| 24 |
1014.75KB |
| 25 |
813.02KB |
| 26 |
729.88KB |
| 27 |
258.97KB |
| 28 |
895.76KB |
| 29 |
137.59KB |
| 3 |
155.03KB |
| 30 |
137.59KB |
| 31 |
137.59KB |
| 32 |
137.59KB |
| 33 |
601.48KB |
| 34 |
198.01KB |
| 35 |
503.67KB |
| 36 |
629.93KB |
| 37 |
363.32KB |
| 38 |
804.81KB |
| 39 |
625.21KB |
| 4 |
734.24KB |
| 40 |
622.03KB |
| 41 |
730.93KB |
| 42 |
515.01KB |
| 43 |
189.41KB |
| 44 |
376.24KB |
| 45 |
90.43KB |
| 46 |
966.59KB |
| 47 |
636.51KB |
| 48 |
471.78KB |
| 49 |
952.72KB |
| 5 |
734.24KB |
| 50 |
908.67KB |
| 51 |
908.67KB |
| 52 |
425.84KB |
| 53 |
592.96KB |
| 54 |
925.03KB |
| 55 |
762.57KB |
| 56 |
50.22KB |
| 57 |
616.19KB |
| 58 |
945.52KB |
| 59 |
945.52KB |
| 6 |
734.24KB |
| 60 |
216.95KB |
| 61 |
384.96KB |
| 62 |
548.41KB |
| 63 |
760.07KB |
| 64 |
969.13KB |
| 65 |
25.32KB |
| 66 |
138.62KB |
| 67 |
917.08KB |
| 68 |
994.19KB |
| 69 |
275.09KB |
| 7 |
734.24KB |
| 70 |
480.31KB |
| 71 |
686.84KB |
| 72 |
701.69KB |
| 73 |
742.78KB |
| 74 |
844.97KB |
| 75 |
387.41KB |
| 76 |
477.02KB |
| 77 |
541.71KB |
| 78 |
666.02KB |
| 79 |
686.11KB |
| 8 |
734.24KB |
| 80 |
781.47KB |
| 81 |
926.20KB |
| 82 |
12.41KB |
| 83 |
119.06KB |
| 84 |
123.28KB |
| 85 |
622.04KB |
| 86 |
734.31KB |
| 87 |
734.31KB |
| 88 |
802.94KB |
| 89 |
949.45KB |
| 9 |
734.24KB |
| 90 |
949.45KB |
| 91 |
989.12KB |
| 92 |
94.59KB |
| 93 |
351.81KB |
| 94 |
375.79KB |
| 95 |
660.65KB |
| 96 |
698.22KB |
| 97 |
766.59KB |
| 98 |
894.40KB |
| 99 |
144.13KB |
| TutsNode.net.txt |
63B |