Temporal Difference Learning

ID: D3A-TDL | Type: Technique | Ontology: d3f:TemporalDifferenceLearning
Unpublished

Description

Temporal difference (TD) learning refers to a class of model-free reinforcement learning methods which learn by bootstrapping from the current estimate of the value function. These methods sample from the environment, like Monte Carlo methods, and perform updates based on current estimates, like dynamic programming methods

Technical Details

Framework MITRE D3FEND
Ontology URI d3f:TemporalDifferenceLearning
Local Identifier TemporalDifferenceLearning
Publication Status Exists in ontology only

Relationships

Parent Tactics

Child Concepts