Open rombodawg opened 6 months ago
This looks a lot like a partial or corrupted download of the model. Are you able to load it in other software from that path?
Same thing here, Ive been pounding my head on this issue for over 12 hours.
Huggingface Login successful
C:\Users\USER\mergekit>mergekit-yaml F:\AIBaseMergeModels\ProfessorSensei\mergekit-config.yml F:\AIBaseMergeModels\ProfessorSensei\ProfessorSensei-v1.0 --cuda --write-model-card --lazy-unpickle --low-cpu-memory --out-shard-size 1B --safe-serialization
Traceback (most recent call last):
File "
@cg123 Yea sorry i forgot to mention, i got it working, i didnt finish my download, it got interrupted, i did end up merging the model just fine.
@cg123 Yea sorry i forgot to mention, i got it working, i didnt finish my download, it got interrupted, i did end up merging the model just fine.
Glad you got it working!
@CaptainHer0 that looks like a syntax error in your YAML config file. Could you post the config you're trying to use?
### [solved] My YAML file was not nested properly. I've only used Python for less than a week. Lesson learned.
Something:
- something-else
another-thing
- something-else2
another-thing2
@cg123
Windows 11 Hone - 23H2 (OS build: 22631.3296) Python 3.12.2 (I have tried 3.12.1 & 3.10) PIP v24.0
*My inspiration to use these 4 models came from theprofessor155B. I wanted to include the entire updated meditron-70b model and the new WizardMath-7b with bfloat16. Maybe the linear merge is the wrong type for this?
Steps:
huggingface-cli login
Login successful
git clone https://github.com/cg123/mergekit.git
cd mergekit
pip install -e .
C:\Users\USER\mergekit>
mergekit-yaml F:\mergemodels\professorsensei\config.yaml F:\mergemodels\professorsensei\professorsensei-v1.0 --cuda --out-shard-size 1B --lazy-unpickle`
RESULTS IN ERRORS
"config.yaml" UTF-8
model_name = professorsensei-v1.0
merge_method linear
slices
sources
model F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
layer_range [0, 1]
model F:\mergemodels\epfl-llm\meditron-70b
layer_range [0, 1]
parameters
weight 0
sources
model F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
layer_range [1, 25]
sources
model F:\mergemodels\WizardLM\WizardMath-7B-V1.1
layer_range [1, 17]
sources
model F:\mergemodels\epfl-llm\meditron-70b
layer_range [1, 40]
sources
model F:\mergemodels\migtissera\Synthia-70B-v1.2b
layer_range [10, 25]
sources
model F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
layer_range [25, 45]
sources
model F:\mergemodels\WizardLM\WizardMath-7B-V1.1
layer_range [17, 35]
sources
model F:\mergemodels\migtissera\Synthia-70B-v1.2b
layer_range [45, 65]
sources
model F:\mergemodels\epfl-llm\meditron-70b
layer_range [40, 80]
sources
model F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
layer_range [45, 79]
sources
model F:\mergemodels\epfl-llm\meditron-70b
layer_range [79, 80]
model F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
layer_range [79, 80]
parameters
weight 0
dtype bfloat16
tokenizer_source model:F:\mergemodels\cognitivecomputations\dolphin-2.2-70b
I have tried referencing the HF repositories as well.
ERROR:
Traceback (most recent call last):
File "<frozen runpy>", line 198, in _run_module_as_main
File "<frozen runpy>", line 88, in _run_code
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Scripts\mergekit-yaml.exe\__main__.py", line 7, in <module>
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Lib\site-packages\click\core.py", line 1157, in __call__
return self.main(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Lib\site-packages\click\core.py", line 1078, in main
rv = self.invoke(ctx)
^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Lib\site-packages\click\core.py", line 1434, in invoke
return ctx.invoke(self.callback, **ctx.params)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Lib\site-packages\click\core.py", line 783, in invoke
return __callback(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\mergekit\mergekit\options.py", line 76, in wrapper
f(*args, **kwargs)
File "C:\Users\USER\mergekit\mergekit\scripts\run_yaml.py", line 44, in main
merge_config: MergeConfiguration = MergeConfiguration.model_validate(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python312\Lib\site-packages\pydantic\main.py", line 509, in model_validate
return cls.__pydantic_validator__.validate_python(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
**pydantic_core._pydantic_core.ValidationError: 1 validation error for MergeConfiguration
Input should be a valid dictionary or instance of MergeConfiguration [type=model_type, input_value='model_name = "professors...ations\\dolphin-2.2-70b', input_type=str]
For further information visit https://errors.pydantic.dev/2.6/v/model_type**
I have all the model(s) files downloaded on my SSD (F:) Thank You! 👍
When merging any starcoderV2 models this error always comes up "safetensors_rust.SafetensorError: Error while deserializing header: MetadataIncompleteBuffer"
Complete error bellow