metadata
base_model: DavidAU/L3-MOE-4X8B-Grand-Horror-25B
language:
- en
library_name: transformers
quantized_by: mradermacher
tags:
- mergekit
- moe
- mixture of experts
- merge
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):