Model Gallery

Discover and install AI models from our curated collection

6 models available
1 repositories
Documentation

Find Your Perfect Model

Filter by Model Type

Browse by Tags

gemma-3-4b-it
Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. Gemma 3 models are multimodal, handling text and image input and generating text output, with open weights for both pre-trained variants and instruction-tuned variants. Gemma 3 has a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions. Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone. Gemma-3-4b-it is a 4 billion parameter model.

Repository: localaiLicense: gemma

gemma-3-4b-it-qat
This model corresponds to the 4B instruction-tuned version of the Gemma 3 model in GGUF format using Quantization Aware Training (QAT). The GGUF corresponds to Q4_0 quantization. Thanks to QAT, the model is able to preserve similar quality as bfloat16 while significantly reducing the memory requirements to load the model. You can find the half-precision version here.

Repository: localaiLicense: gemma

mlabonne_gemma-3-4b-it-abliterated
This is an uncensored version of google/gemma-3-4b-it created with a new abliteration technique. See this article to know more about abliteration.

Repository: localaiLicense: gemma

gemma-3-4b-it-uncensored-dbl-x-i1
Google's newest Gemma-3 model that has been uncensored by David_AU (maintains instruction following / model performance and adds 4 layers to the model) and re-enforced with a system prompt (optional) - see below.

Repository: localaiLicense: gemma

genericrpv3-4b
Model's part of the GRP / GenericRP series, that's V3 based on Gemma3 4B, licensed accordingly. It's a simple merge. To see intended behavious, see V2 or sum, card's more detailed. allura-org/Gemma-3-Glitter-4B: w0.5 huihui-ai/gemma-3-4b-it-abliterated: w0.25 Danielbrdz/Barcenas-4b: w0.25 Happy chatting or whatever.

Repository: localaiLicense: gemma

gemma-3-4b-it-max-horror-uncensored-dbl-x-imatrix
Google's newest Gemma-3 model that has been uncensored by David_AU (maintains instruction following / model performance and adds 4 layers to the model) and re-enforced with a system prompt (optional) - see below. The "Horror Imatrix" was built using Grand Horror 16B (at my repo). This adds a "tint" of horror to the model. 5 examples provided (NSFW / F-Bombs galore) below with prompts at IQ4XS (56 t/s on mid level card). Context: 128k. "MAXED" This means the embed and output tensor are set at "BF16" (full precision) for all quants. This enhances quality, depth and general performance at the cost of a slightly larger quant. "HORROR IMATRIX" A strong, in house built, imatrix dataset built by David_AU which results in better overall function, instruction following, output quality and stronger connections to ideas, concepts and the world in general. This combines with "MAXing" the quant to improve preformance.

Repository: localaiLicense: gemma