Skip to content

Conversation

1773226512
Copy link

Thank you for your great works. We have released the 4-bit GPTQ quantized LLaDA model on Hugging Face:

Based on the published evaluation code, we have evaluated the quantized base model. The results are as follows:

Dataset GPTQ-4bit FP16
MMLU 65.20 65.90
CMMLU 69.23 69.90
ARC-Challenge 45.48 47.90

@NieShenRuc
Copy link
Collaborator

I am so sorry for the late response. Thank you for your excellent work! May I ask what is the accuracy of quantized LLaDA model on GSM8K?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants