| uv run python custom_mlx_lm/custom_convert.py --hf-path . --mlx-path MobileLLM-R1-950M-mixed-4bit-mlx --dynamic-quant --target-bpw 4.5 --group-size 64 --report-ppl |
| Loading model from .... |
| Loading calibration data... |
| Token indices sequence length is longer than the specified maximum sequence length for this model (110205 > 32768). Running this sequence through the model will result in indexing errors |
| Calculating perplexity of original model... |
| Original PPL: 50.262 |
| Starting advanced mixed-precision quantization... |
| huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks... |
| To disable this warning, you can either: |
| - Avoid using `tokenizers` before the fork if possible |
| - Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false) |
| Estimating sensitivities: 100%|████████████████████████████████████| 54/54 [02:03<00:00, 2.28s/it] |
| Calculating perplexity of quantized model... |
| Quantized PPL: 59.059 |
|
|
| ✅ Model saved to MobileLLM-R1-950M-mixed-4bit-mlx |
|
|
| uv run python custom_mlx_lm/quant_summary.py --model-path MobileLLM-R1-950M-mixed-4bit-mlx --show 8 |
| Method: mixed_precision_dynamic |
| Group size: 64 |
| Total linear layers: 154 |
| 4-bit layers: 153 |
| 8-bit layers: 1 |
|
|
| Examples (8-bit): |
| - layers.0.attention.o_proj |
|
|
| Examples (4-bit): |
| - layers.0.attention.k_proj |
| - layers.0.attention.q_proj |
| - layers.0.attention.v_proj |
| - layers.0.feed_forward.down_proj |
| - layers.0.feed_forward.gate_proj |
| - layers.0.feed_forward.up_proj |
| - layers.1.attention.k_proj |
| - layers.1.attention.o_proj |
|
|
| weights.npz contains quantized tensors: True |