Closed hxan closed 7 months ago
Yes it's same. But it may differ from a pure llama weights since openflamingo would add cross gates attention layers into llm decoder layers.
Hello, author, would you like to ask, does it mean that I only need to download one of the pre-trained weights of about 30GB, and then change the pre-trained model by downloading different final_weights.pt? If so this will help me save a lot of download time
Hello, authors, among the series of pre-trained weights of huggingface, are the weights of Llama all the same?