Skip to content

Huy-Quang-Dao/Inverse-RL-Intelligent-Control-DT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

[Julia] Inverse Reinforcement Learning for Expert-Learner Discrete-time System

Introduction

Here is my implementation of the model described in the paper Inverse Reinforcement Q-Learning Through Expert Imitation for Discrete-Time Systems paper.

Experiments:

The algorithm makes the Learner achieve the same control matrix as the Expert, while The state-reward weight converges to a different value than the Expert.

The Expert's control matrix is ​​as follows

$$K_{Expert} = \begin{bmatrix} -0.1688 & -0.2009 & 0.1285 \end{bmatrix}$$

I show the results obtained from my experiments.

$$Q = \begin{bmatrix} 4.62943 & 4.68051 & 1.2945 \\\ 4.68051 & -1.14605 & 1.06186 \\\ 1.2945 & 1.06186 & -1.63109 \end{bmatrix} , K_{Learner} = \begin{bmatrix} -0.172843 & -0.204428 & 0.125967 \end{bmatrix}$$

Results

Convergence of the proposed algorithm
drawing
Output result
drawing

Docker

I will provide DockerFile soon.

Requirements

  • Julia v1.10.3
  • LinearAlgebra
  • Plots
  • Kronecker

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages