YCCS-Summer-2023-DDNMA / project

1 stars 3 forks source link

Dynamic Deep Network Models

This repository contains code and examples for training Dynamic Deep Network (DDN) models that can adapt to evolving data distributions.

Background

Neural network models are typically trained on static datasets with fixed data distributions. In real-world scenarios, the underlying data distribution often changes over time due to various factors. When these distributional shifts occur, the performance of models trained on the original data can deteriorate.

DDN models offer a solution by enabling models to continuously learn and adapt from new data as it becomes available. This allows the models to track changes in the data distribution in a controlled manner.

Approach

We use State Space Models (SSMs) based on the Extended Kalman Filter (EKF) for training DDN models, rather than standard training algorithms like gradient descent.

With EKF, we obtain not only the model weights, but also the uncertainties over the weights. This allows finely controlling the adaptation rate of different parts of the model to new data by tweaking the prior uncertainties.

Contents

This repo contains:

More Details

See the presentation for more details on the background, motivation, and technical approach for this project.

Authors

This project was developed by:

Under the mentorship of Ramesh Natarajan.