microsoft / deep-language-networks

We view Large Language Models as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a Deep Language Network - DLN
MIT License
87 stars 13 forks source link

added abstract and limitation to the readme #5

Closed matheper closed 1 year ago

MarcCote commented 1 year ago

We could add links to Arxiv, PDF, blog post like they've done here (and maybe the badges too) image

matheper commented 1 year ago

Sure, we can do that once we have the links to the arxiv submission and the project page. The instructions on how to run experiments will be included when we review the scripts.