Open ZTAP0011 opened 4 months ago
can i know how you merged the model, because i am not able to merge the model in a 24GB VRAM, always run out of memory when i merge model instance i am using: ml g5 12x large
The model was trained with your data, and I've taken a look and I dont see any big difference, its just that after fine tuned it answers more closely to what your data suggested. I also noticed that you have a huge amount of cases where there is a single category, and this might also be playing a role. You do not have the category "Tees" at all in your dataset for example. There might be missing diversity. The output from the fine tuned is not wrong per see neither- it does seem indeed like it was properly fine tuned with your data.
Otherwise it can also depend on hyperparameters and the quantity of data you have. But your case is more related to the data itself.
Also, I noticed that your dataset has a lot of "You are a shopping assistant for the user.\n\n input: viewed products", what is "viewed products" here? 🤔
Im also confused why your expected result is to have the same output as the mistral 7b model, as usually we want to improve it?
can i know how you merged the model, because i am not able to merge the model in a 24GB VRAM, always run out of memory when i merge model instance i am using: ml g5 12x large
@kiranshivaraju we were able to do it using https://github.com/mistralai/mistral-finetune and using vast ai 1x A100 SXM4 instance having 80 GB VRAM.
Python Version
Pip Freeze
Reproduction Steps
I fined tuned mistral model instruct v0.3 model as per the instructions mentioned here https://github.com/mistralai/mistral-finetune. Attaching my train.jsonl and eval.jsonl file.
Now when I ask my fine tuned model following question. we have following categories of products \n\n- Uncategorized,Accessories,Automotive,Baby Products,baby-products -,Bags,Beauty & Personal Care,Bottoms,boy,Bras & Tanks,Cell Phones & Accessories,Clothing,clothing -,Collectibles & Fine Art,Collections,Decor,Eco Friendly,Eco Friendly|Clothing,Electronics,Erin Recommends,Erin Recommends|Clothing,Fashion,Fitness Equipment,Fitness Equipment|Collections,Gear,girl,Health & Household,Hoodies,Hoodies & Sweatshirts,Interior,Jackets,Lawn & Garden,Men,Mens Fashion,Music,Musical Instruments,New Luma Yoga Collection,New Luma Yoga Collection|Clothing,Pants,Patio,Performance Fabrics,Performance Fabrics|Clothing,Promotions,Raincoats,Shirts,Shoes,Shorts,shorts bottoms women,Sports & Outdoors,Sweater,T-shirts,Tanks,Tees,Tees|Clothing,Tools & Home Improvement,Tops,Video Games,Watches,Watches|Collection,Winter,Women,Women Sale,Women Sale|Clothing,Womens Fashion.\n\n Please return categories from this list those matches with user need\n\n input: I want to purchase tshirts
It returns _{'followup':'sure, here are some trending tshirts for you $item', 'intent': 'browseproducts', 'metadata': {'categories': ['tshirts']}}
But when same question I ask to Mistral instruct v0.3 here it returns correct matched categories as below Based on the provided list, the categories that match your need for purchasing t-shirts are:
T-shirts Tees Tees|Clothing You may find your desired t-shirts in any of these categories. Happy shopping!
Why my fine tuned model is not returning T-Shirts, Tees, Tees|Clothing in metadata.categories.
train_and_eval_files.zip
Expected Behavior
My fine tuned model should return T-Shirts, Tees, Tees|Clothing in metadata. categories.
Additional Context
command to fine tune the model - torchrun --nproc-per-node 1 --master_port 8009 -m train example/7B.yaml
attaching ![Uploading 7B_yaml.JPG…]() 7B.yaml file
Suggested Solutions
No response