LMOps is a research initiative on fundamental research and technology for building AI products w/ foundation models, especially on the general technology for enabling AI capabilities w/ LLMs and Generative AI models.
According to the demonstration examples, GPT produces meta gradients for In-Context Learning (ICL) through forward computation. ICL works by applying these meta gradients to the model through attention.
The meta optimization process of ICL shares a dual view with finetuning that explicitly updates the model parameters with back-propagated gradients.
We can translate optimization algorithms (such as SGD with Momentum) to their corresponding Transformer architectures.
We are hiring at all levels (including FTE researchers and interns)! If you are interested in working with us on Foundation Models (aka large-scale pre-trained models) and AGI, NLP, MT, Speech, Document AI and Multimodal AI, please send your resume to fuwei@microsoft.com.
License
This project is licensed under the license found in the LICENSE file in the root directory of this source tree.
For help or issues using the pre-trained models, please submit a GitHub issue.
For other communications, please contact Furu Wei (fuwei@microsoft.com).
LMOps/README.md at main · microsoft/LMOps
LMOps
LMOps is a research initiative on fundamental research and technology for building AI products w/ foundation models, especially on the general technology for enabling AI capabilities w/ LLMs and Generative AI models.
Links
News
Prompt Intelligence
Advanced technologies facilitating prompting language models.
Promptist: reinforcement learning for automatic prompt optimization
Structured Prompting: consume long-sequence prompts in an efficient way
[Paper] Structured Prompting: Scaling In-Context Learning to 1,000 Examples
Example use cases:
X-Prompt: extensible prompts beyond NL for descriptive instructions
LLMA: LLM Accelerators
Accelerate LLM Inference with References
Fundamental Understanding of LLMs
Understanding In-Context Learning
Hiring: aka.ms/GeneralAI
We are hiring at all levels (including FTE researchers and interns)! If you are interested in working with us on Foundation Models (aka large-scale pre-trained models) and AGI, NLP, MT, Speech, Document AI and Multimodal AI, please send your resume to fuwei@microsoft.com.
License
This project is licensed under the license found in the LICENSE file in the root directory of this source tree.
Microsoft Open Source Code of Conduct
Contact Information
For help or issues using the pre-trained models, please submit a GitHub issue. For other communications, please contact Furu Wei (
fuwei@microsoft.com
).Suggested labels