microsoft / unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
https://aka.ms/GeneralAI
MIT License
20.2k stars 2.55k forks source link

Kosmos 2.5 #1297

Closed conceptofmind closed 6 months ago

conceptofmind commented 1 year ago

Hello,

Thank you for all your great research.

I was wondering if there were plans to release Kosmos 2.5 similar to how Kosmos 2 was released on Huggingface.

Thank you,

Enrico

pribadihcr commented 1 year ago

+1

logan-markewich commented 1 year ago

+1

FrancescoSaverioZuppichini commented 1 year ago

+1

fernando-neto-ai commented 1 year ago

+1

evolu8 commented 1 year ago

please +1

luvwinnie commented 1 year ago

+1

shenguojun commented 1 year ago

+1

MiaSanLei commented 1 year ago

+1

dydxdt commented 1 year ago

+1

coder4nlp commented 1 year ago

+1

zykzyk1993 commented 1 year ago

+1 please

ydshieh commented 1 year ago

+1 please 🙏

acul3 commented 1 year ago
contrebande-labs commented 1 year ago

Would be nice if there was an official acknowledgement of this.

Natyren commented 1 year ago

Hi @wolfshow, сould you please let people there know if there are plans to release the code and weights for this model?

wolfshow commented 1 year ago

Thank you all for expressing interest in Kosmos-2.5. We plan to release the code and models, though we currently do not have a specific ETA. It's important to note that we are currently facing limitations in GPU resources. Despite this challenge, we are committed to making the models extremely user-friendly. There is still a significant amount of work ahead of us before we can open source the project, and we appreciate your understanding and patience as we navigate these constraints. We look forward to updating you on our progress.

ydshieh commented 1 year ago

Thank you @wolfshow for the information. Super glad your team considers to open source this exciting model!

FrancescoSaverioZuppichini commented 1 year ago

Thank you all for expressing interest in Kosmos-2.5. We plan to release the code and models, though we currently do not have a specific ETA. It's important to note that we are currently facing limitations in GPU resources. Despite this challenge, we are committed to making the models extremely user-friendly. There is still a significant amount of work ahead of us before we can open source the project, and we appreciate your understanding and patience as we navigate these constraints. We look forward to updating you on our progress.

Thanks a lot @wolfshow for breaking the silence :) I have to wonder, the paper did shows result that result must have come from some code and some trained model. Why do you need to use GPUs now? Would it be possible to release what you have done for the paper and let the community take care of the rest? Otherwise, it is very hard for us to validate the results and also to take advantage of the research done. I am sure all the authors would love to see people using their work.

Thank you again

logan-markewich commented 1 year ago

@FrancescoSaverioZuppichini it's microsoft research, so I'm sure they have their own red tape, responsibilities, and procedures to adhere too :)

FrancescoSaverioZuppichini commented 1 year ago

@FrancescoSaverioZuppichini it's microsoft research, so I'm sure they have their own red tape, responsibilities, and procedures to adhere too :)

Sure, They could have said it weeks ago if they cared to share them :) Well looks like this discussion as come to an end.

TL;DR At some point hopefully when the model is still relevant, we'll get some code that for some reason needs GPUs time while the paper showed results that must have been run on GPUs. We are living in the matrix

wolfshow commented 1 year ago

@FrancescoSaverioZuppichini it's microsoft research, so I'm sure they have their own red tape, responsibilities, and procedures to adhere too :)

Sure, They could have said it weeks ago if they cared to share them :) Well looks like this discussion as come to an end.

TL;DR At some point hopefully when the model is still relevant, we'll get some code that for some reason needs GPUs time while the paper showed results that must have been run on GPUs. We are living in the matrix

We need computations to run more experiements which are not included in the paper right now.

evolu8 commented 1 year ago

Thank you @wolfshow . Your efforts are very much appreciated! Good luck with tying off the final pieces.

shubhamagarwal92 commented 12 months ago

Hi @wolfshow! Interesting work on Kosmos 2.5!

Is there any plan to release the datasets as well? Or the corresponding pipeline code for different datasets (Sec 2.3/2.4 in the paper)?

evolu8 commented 11 months ago

@wolfshow - any news?

evolu8 commented 9 months ago

@wolfshow - another month, another nudge :) Please keep us posted.

ydshieh commented 9 months ago

Really hope this could be released - it would have huge impact 🙏 🔥

coder4nlp commented 8 months ago

Please!

FrancescoSaverioZuppichini commented 8 months ago

Hard F

atlury commented 8 months ago

Any ETA for the code and models for this paper? Its been nearly 3-4 months.

radiachkik commented 8 months ago

+1

luohao123 commented 7 months ago

Still plan to release the code? The opensource community really needs such a work!

sairin94 commented 7 months ago

+1

AlexAndrei98 commented 6 months ago

+11 🚀 🚀 🚀

wolfshow commented 6 months ago

Kosmos 2.5 models and code have been publicly available at https://huggingface.co/microsoft/kosmos-2.5 and https://github.com/microsoft/unilm/tree/master/kosmos-2.5. Thanks!

wolfshow commented 5 months ago

Hello,

Thank you for all your great research.

I was wondering if there were plans to release Kosmos 2.5 similar to how Kosmos 2 was released on Huggingface.

Thank you,

Enrico

https://github.com/huggingface/transformers/issues/30877

EwoutH commented 4 months ago

Maybe notable: Kosmos-2.5 now also is published in Safetensors format on HuggingFace.

atlury commented 4 months ago

Ok an example with image input on hugging spaces would be highly helpful.