Skip to content

Commit

Permalink
Add docs for evaluation
Browse files Browse the repository at this point in the history
  • Loading branch information
yeukfu committed Oct 22, 2020
1 parent 9147627 commit 541a58b
Show file tree
Hide file tree
Showing 2 changed files with 32 additions and 0 deletions.
1 change: 1 addition & 0 deletions docs/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -71,6 +71,7 @@ and how to implement new MDPs and new algorithms.
user/environment_libraries
user/concept_experiment
user/sampling
user/evaluation
.. toctree::
:maxdepth: 2
Expand Down
31 changes: 31 additions & 0 deletions docs/user/evaluation.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,31 @@
# Evaluation

garage provides two useful function to evaluate the performance of an
algorithm, [log_performance](../_autoapi/garage/index.html#garage.log_performance)
and [log_multitask_performance](../_autoapi/garage/index.html#garage.log_multitask_performance).
`log_performance` is used for generous algorithms, while
`log_multitask_performance` is used for multiple tasks algorithms.

The input of the both functions is [EpisodeBatch](../_autoapi/garage/index.html#garage.EpisodeBatch),
which is a batch of episodes.

These functions will evaluate algorithms in from the following aspects:

- `AverageReturn`: The average return (sum of rewards in an episode) of all
episodes.

- `AverageDiscountedReturn`: The average discounted return of all episodes.

- `StdReturn`: The standard deviation of undiscounted returns.

- `MaxReturn`: The maximum undiscounted return.

- `MinReturn`: The minimum undiscounted return.

- `TerminationRate`: Terminated episodes / all episodes.

- `SuccessRate` (if applicable): The rate of success among all episodes.

----

*This page was authored by Ruofu Wang ([@yeukfu](https://github.com/yeukfu)).*

0 comments on commit 541a58b

Please sign in to comment.