This model DeProgrammer/Jan-v3-4B-base-instruct-MNN-Q8 was converted to MNN format from janhq/Jan-v3-4B-base-instruct using llmexport.py in MNN version 3.4.0 with --quant_bit 8 but otherwise default settings.

Inference can be run via MNN, e.g., MNN Chat on Android.

Downloads last month
8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DeProgrammer/Jan-v3-4B-base-instruct-MNN-Q8

Quantized
(16)
this model