gemma-3-12b-it-ai-expert - A Fine-tuned Model for AI Core Technologies

馃 Hugging Face   |   馃 ModelScope

This model is a specialized expert on core Artificial Intelligence concepts, developed by performing Instruction Supervised Fine-Tuning (SFT) on the google/gemma-3-12b-it model.

The fine-tuning was conducted using Low-Rank Adaptation (LoRA), a parameter-efficient technique, on a custom-built dataset. This process adapted the model to provide high-quality, detailed responses specifically within the domains of:

  • Large Language Models (LLMs)
  • Retrieval-Augmented Generation (RAG)
  • AI Agents

The model was fine-tuned with LlaMA-Factory.

  • Developed by: real-jiakai
  • License: gemma
  • Finetuned from model : google/gemma-3-12b-it

Usage

from transformers import pipeline
import torch

pipe = pipeline(
    model="GXMZU/gemma-3-12b-it-ai-expert",
    device="cuda",
    torch_dtype=torch.bfloat16
)

messages = [
    {
        "role": "system",
        "content": [{"type": "text", "text": "You are an AI expert assistant(Focus on LLM, RAG, and Agent Domain) to help with technical questions. You should provide clear, accurate, and helpful responses."}]
    },
    {
        "role": "user",
        "content": [
            {"type": "text", "text": "What is MCP Protocol?"}
        ]
    }
]

output = pipe(messages, max_new_tokens=200, temperature=0.1)
print(output[0]["generated_text"][-1]["content"])

Performance

The primary objective of this fine-tuning is to adapt the model to a specialized domain, enhancing its performance on specific tasks by injecting relevant knowledge and terminology while preserving its foundational generalist capabilities.

Before Fine-tuning vs After Fine-tuning

The model demonstrates significant improvements in domain-specific tasks related to LLM, RAG, and AI Agents, as shown in the example below:

demo

Fine-tuning Procedure

Dataset

The model was fine-tuned on a custom, high-quality dataset. The dataset was carefully curated to cover three core areas:

  • Large Language Models (LLM)
  • Retrieval-Augmented Generation (RAG)
  • AI Agents

Citation

If you use this model in your work, please cite it as:

@misc{gemma-3-12b-it-ai-expert,
  author = {real-jiakai},
  title = {gemma-3-12b-it-ai-expert},
  year = 2025,
  url = {https://huggingface.co/GXMZU/gemma-3-12b-it-ai-expert},
  publisher = {Hugging Face}
}
@article{gemma_2025,
    title={Gemma 3},
    url={https://goo.gle/Gemma3Report},
    publisher={Kaggle},
    author={Gemma Team},
    year={2025}
}
Downloads last month
76
Safetensors
Model size
12B params
Tensor type
BF16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for GXMZU/gemma-3-12b-it-ai-expert

Finetuned
(153)
this model