rickstaa / stable-learning-control

A framework for training theoretically stable (and robust) Reinforcement Learning control algorithms.
https://rickstaa.dev/stable-learning-control
MIT License
6 stars 1 forks source link

Different output activation function #97

Closed rickstaa closed 3 years ago

rickstaa commented 3 years ago

When trying to solve rickstaa/LAC-TF2-TORCH-translation#38 I found out that the old TensorFlow version uses a relu activation function as the output activation function for the fully connected layers of the Gaussian Actor and Lyapunov Critic. The following output activation functions are used in literature:

Gaussian Actor output tfunction

Lyapunov critic

When inspecting the structures further Minghoas version is equal to Haarnojas version the output layer now however becomes a hidden layer and the Square activation function is used after that. The package already contains a method to change the activation functions from the CLI.

rickstaa commented 3 years ago

All the different output activation functions allow the agent to train. Based on the task and environment they might however still infleunce the performance.