Closed Kuckuck44 closed 9 months ago
Picked German LLAMA-Fork by "TheBloke" (known contributor on Hugging Face) https://huggingface.co/TheBloke/Llama-2-13B-German-Assistant-v4-GGUF/tree/main
Used desktop-version of GPT4All (https://gpt4all.io/index.html) to generate Text
GPT4All is a library / tool for easy model-interaction → Allows use of different models with low effort
Makes use / interchangeability of German baseline easy
How to import own models:
Results:
For further testing I will try llama-2-13b-german-assistant-v4.Q4_K_M.gguf, as it is flagged as "medium, balanced quality - recommended".
By the way - "TheBloke" only provided the "inferencable" model formats. The underlaying model is fine-tuned by Florian Zimmermeister.
To generate the first test-data, i will use the Prompt "Bitte generiere einen langen Aufsatz, basierend auf dem folgenden Anfang: "Hallo, mein Name ist"." The Max-Token paramter for the generated response has ben manually modified to 131.072, instead of the default 4096.
I will create a .txt file with at least 5.000 Words from different responses to that input...anything else will last forever :)
Just to clarify: the pretty silly modification of the "Max-Token" parameter was just in regard to some testing of GPT4All, because the used model behaved strangely at first.