huggingface / transformers

πŸ€— Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
https://huggingface.co/transformers
Apache License 2.0
132.82k stars 26.48k forks source link

Mixture of All Intelligence (MoAI) #29823

Open Dev-Khant opened 6 months ago

Dev-Khant commented 6 months ago

Model description

A new large language and vision model (LLVM) that uses auxiliary visual information and natural language for prediction.

It uses 2 modules: π™ˆπ™€π˜Όπ™„-π˜Ύπ™€π™’π™₯π™§π™šπ™¨π™¨π™€π™§ and π™ˆπ™€π˜Όπ™„-π™ˆπ™žπ™­π™šπ™§. Here π—–π—Όπ—Ίπ—½π—Ώπ—²π˜€π˜€π—Όπ—Ώ condenses the verbalized outputs of the external CV models into auxiliary visual information and π— π—Άπ˜…π—²π—Ώ blends three types of intelligence β€” visual features, auxiliary features from external CV models and language features into a cohesive whole.

MoAI-7B surpasses both open-source and closed-source LLVMs in vision language tasks.

Model repo: https://github.com/ByungKwanLee/MoAI

Open source status

Provide useful links for the implementation

No response

Dev-Khant commented 6 months ago

@NielsRogge Can I work on this to add it to the library?

amyeroberts commented 6 months ago

Sure! Free free to open a PR and let us know when it's ready for review or you need help integrating into the library.

In general, we prioritise reviewing based on PRs opened rather than comments on issues, as we find this prevents issues from becoming stale. You're free to work on something if there's no active linked PRs open.

Dev-Khant commented 6 months ago

Thanks @amyeroberts I'll start working on this as I don't see any open PR regarding this.

Dev-Khant commented 5 months ago

Waiting for this https://github.com/huggingface/transformers/pull/29667 to get merged, because internally model uses InternLM.