lhnguyen102 / cuTAGI

CUDA implementation of Tractable Approximate Gaussian Inference
MIT License
24 stars 9 forks source link

New formulation for mixture-based activation functions #58

Closed jamesgoulet closed 3 months ago

jamesgoulet commented 3 months ago

Description

This PR contains the new formulation by Lucas Alric for the mixture-based activation function, i.e. mixture- ReLU, Sigmoid and Tanh. The new formulations are simpler and most importantly, they remove the need to have omega_tol to avoid having numerical issues with /0.

Changes Made

Note for Reviewers

You can test the new activation functions through either the test.py or test_lstm.py

The mathematical formulation implemented as well as the comparison with MC sampling are presented in the following file: mRELU_Goulet_2022.pdf

Screenshot 2024-04-03 at 18 26 19