policy

Double_deep_q_learning_policy_distillation_for_optimal_execution

TosiMatte · PyTorch

or hover any field below to flag it

Overview

Name
Double_deep_q_learning_policy_distillation_for_optimal_execution
Author
TosiMatte
Framework
PyTorch
License
unknown
Skill type
other
Evidence level
untested
Task description
Optimal exection reinforcement learning framework with a double deep q-learning teacher trained on perfect market information trough policy distillation shares knowledge to a student network build upon a simplier double deep q-learning architecture and imperfect market information

Spaces

Action space
other · 0-dim · 0Hz
Observation space
  • type: other

Links

HuggingFace repo
null
Paper (arXiv)
null

Compatible robots

3+17 mentioned but not in catalog yet

Compatible environments

0

No environments list Double_deep_q_learning_policy_distillation_for_optimal_execution yet.

Datasets that reference this policy

0

No datasets reference Double_deep_q_learning_policy_distillation_for_optimal_execution yet.