🖼️ Available 1 models from 1 repositories

Filter by type:

Filter by tags:

Operations in progress

vulpecula-4b (from the 'localai' repository)

Installation

meta-llama-3.1-8b-claude-imat

Meta-Llama-3.1-8B-Claude-iMat-GGUF: Quantized from Meta-Llama-3.1-8B-Claude fp16. Weighted quantizations were creating using fp16 GGUF and groups_merged.txt in 88 chunks and n_ctx=512. Static fp16 will also be included in repo. For a brief rundown of iMatrix quant performance, please see this PR. All quants are verified working prior to uploading to repo for your safety and convenience.

Repository: localaiLicense: llama3.1

Link #1Link #2