Open AtakanTekparmak opened 1 month ago
The model also produces gibberish for me when I did the same. I'm also on an m1 max but with 32gb. Same thing, the text came back as all white. Python version: 3.11.1 PyTorch version: 2.0.1 NumPy version: 1.26.4 MLX version: 0.17.3
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
<antThinking>
I need to delve deeper into my actual thought processes, including the uncertainties, associations, and even potential biases that arise as I consider the query. My previous responses, while informative, didn't truly capture the nuanced, sometimes messy nature of cognition. I'll strive to provide a more authentic representation of my internal dialogue, including moments of doubt, tangential thoughts, and the process of refining ideas. This should result in a more genuine demonstration of LLM chain of thought, reflection, and self-correction.
</antThinking>
Which is greater, 9.9 or 9.11?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
=============== generating =================
I’m sorry�������urbed atmurbedardycowurbedardycowuçuçuçansıpast dungurbed�arbeday objected atm dung atm�وروبELINK�����������������������uç�������uç���dccáčcaccccıttarget: Number calculation> calculations suggest our opinion will definitely turn����������12220thelapse for such specific scenario it generally comesDown<eckuç��uç������ansıuç������������������.gr.gr thế.gr�������������������������������������������������������������������������������������������������������.gr.gr.gr.gr.gr.gr.gr.gr thế thế thế���������������uç���������������������������������������������������������������������.gr.gr.gr.gr.gr.gr.gr.gr.gr.gr.gr.```
This happens for me as well. I have identified that the issue comes from the sampler itself (possibly due to all the hyperparameters).
So, when in the sampler
method if you return the sampling immediately using a constant temperature, top_k, top_p, and min_p the output will be what you expect.
Example, if you use this code below, it gives the correct answer:
def sample(
gen_tokens: mx.array, logits: mx.array, attention_scores: mx.array, cfg: SamplerConfig,
clarifying_question_token: int = 2564, key: mx.array = None
) -> Tuple[mx.array, str, dict]:
return _sample(logits, temperature=0.6, top_p=0.97, top_k=27, min_p=0.2), COLORS["lelv"], metrics
You can always tune the hyper-parameters and see what works for the current state of the sampler since essentially this is a research code. You can also experiment around the sampling conditions to check if it can be improved.
I'm also trying to see how it can be improved. Hope this helps.
This happens for me as well. I have identified that the issue comes from the sampler itself (possibly due to all the hyperparameters).
So, when in the
sampler
method if you return the sampling immediately using a constant temperature, top_k, top_p, and min_p the output will be what you expect.Example, if you use this code below, it gives the correct answer:
def sample( gen_tokens: mx.array, logits: mx.array, attention_scores: mx.array, cfg: SamplerConfig, clarifying_question_token: int = 2564, key: mx.array = None ) -> Tuple[mx.array, str, dict]: return _sample(logits, temperature=0.6, top_p=0.97, top_k=27, min_p=0.2), COLORS["lelv"], metrics
You can always tune the hyper-parameters and see what works for the current state of the sampler since essentially this is a research code. You can also experiment around the sampling conditions to check if it can be improved.
I'm also trying to see how it can be improved. Hope this helps.
It seems like after the first token everything falls under the adaptive sampling block, you're right, it's probably related to the hyperparams in the SamplerConfig
.
Trying the example query 9.9 vs 9.11 query on a M1 Max 64GB, Python 3.11.2. I haven't ran the original entropix logo but the model output doesn't seem right. Only "I" at the beginning gets printed as green text, rest plain white
Also provided the installed packages below: