Ersin - On the Use of MDPs in Cognitive Radar: An Application to Target Tracking
Date: 7/27/2018
Author: Jianyuan (Jet) Yu
Contact: [email protected]
Affiliate: Wireless, ECE, Virginia Tech
The work replaces the mdp solver in Ersin work git link as dqn(deep q network) solver. The codes are mostly written in .m except for the dqn.py
file, and python files are mostly based on open source RL with Tensorflow by Morvan Zhou.
It is NOT preferable to run codes in this blended way due to the low efficient computation and difficult to adjust the parameters within the dqn solver. Moreover, these codes may bring the new compatible problem when moving to ARC platform.
Refer to READMEconfig.md
Refer to READMEcodes.md
- 10/22 adding DRQN
- 9/6 add in dpg(deep policy gradient) method, which apt for episode model.
- 9/1 dqn get same result as mdp.
- 7/17 the codes running through.
unmatch ofState
andNumBands
, i.e. the NumBands in 5 in the default case, while the State is 1 times 4 vector instead of 1 times 5.theCurrentActionNumber
outrange, probably comes from the first one, I just did roll over.
- fix DoubleDQN PriDQN DuelDQN
- add DRQN