Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 8 additions & 2 deletions Gemma/[Gemma_3]Gradio_LlamaCpp_Chatbot.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@
"\n",
"Author: Sitam Meur\n",
"\n",
"* GitHub: [github.com/sitamgithub-MSIT](https://github.com/sitamgithub-MSIT/)\n",
"* GitHub: [github.com/sitammeur](https://github.com/sitammeur/)\n",
"* X: [@sitammeur](https://x.com/sitammeur)\n",
"\n",
"Description: Google recently released Gemma 3 QAT—the [Quantization Aware Trained (QAT) Gemma 3](https://huggingface.co/collections/google/gemma-3-qat-67ee61ccacbf2be4195c265b) checkpoints. These models maintain similar quality to half precision while using three times less memory. This notebook demonstrates creating a user-friendly chat interface for the [gemma-3-1b-it-qat](https://huggingface.co/google/gemma-3-1b-it-qat-q4_0-gguf) text model using Llama.cpp (for inference) and Gradio (for user interface).\n",
Expand Down Expand Up @@ -128,7 +128,13 @@
},
"outputs": [],
"source": [
"!pip install -q huggingface_hub scikit-build-core llama-cpp-python llama-cpp-agent gradio"
"# Use the official pre-built wheels for llama-cpp-python with CPU support for compatibility\n",
"!pip install -q \\\n",
" huggingface_hub \\\n",
" scikit-build-core \\\n",
" llama-cpp-python==0.3.9 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu \\\n",
" \"llama-cpp-agent>=0.2.25\" \\\n",
" gradio==5.49.1"
Comment on lines +132 to +137
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

Pinning dependencies is a great practice for reproducibility. To improve clarity and security transparency, it's helpful to add a comment explaining the need for the --extra-index-url. This informs users that it's a trusted and necessary source for the specific package build.

!pip install -q \
  huggingface_hub \
  scikit-build-core \
  # Use the official pre-built wheels for llama-cpp-python with CPU support for compatibility
  llama-cpp-python==0.3.9 --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu \
  "llama-cpp-agent>=0.2.25" \
  gradio==5.49.1

]
},
{
Expand Down
Loading