diff --git a/Gemma/[Gemma_1]Using_with_Ollama.ipynb b/Gemma/[Gemma_1]Using_with_Ollama.ipynb index 1d6af236..a684bf7b 100644 --- a/Gemma/[Gemma_1]Using_with_Ollama.ipynb +++ b/Gemma/[Gemma_1]Using_with_Ollama.ipynb @@ -106,6 +106,7 @@ } ], "source": [ + "!sudo apt-get update && sudo apt-get install -y zstd\n", "!curl -fsSL https://ollama.com/install.sh | sh" ] }, diff --git a/Gemma/[Gemma_2]Using_with_Elasticsearch_and_LangChain.ipynb b/Gemma/[Gemma_2]Using_with_Elasticsearch_and_LangChain.ipynb index c712400d..4d0069f9 100644 --- a/Gemma/[Gemma_2]Using_with_Elasticsearch_and_LangChain.ipynb +++ b/Gemma/[Gemma_2]Using_with_Elasticsearch_and_LangChain.ipynb @@ -336,6 +336,7 @@ } ], "source": [ + "!sudo apt-get update && sudo apt-get install -y zstd\n", "!curl -fsSL https://ollama.com/install.sh | sh" ] }, diff --git a/Gemma/[Gemma_2]Using_with_Firebase_Genkit_and_Ollama.ipynb b/Gemma/[Gemma_2]Using_with_Firebase_Genkit_and_Ollama.ipynb index a4e098b6..7f23f189 100644 --- a/Gemma/[Gemma_2]Using_with_Firebase_Genkit_and_Ollama.ipynb +++ b/Gemma/[Gemma_2]Using_with_Firebase_Genkit_and_Ollama.ipynb @@ -311,6 +311,7 @@ "!pip install -q gradio\n", "\n", "# Install Ollama\n", + "!sudo apt-get update && sudo apt-get install -y zstd\n", "!curl -fsSL https://ollama.ai/install.sh | sh\n", "\n", "# Install Node.js\n", diff --git a/Gemma/[Gemma_2]Using_with_Ollama_Python.ipynb b/Gemma/[Gemma_2]Using_with_Ollama_Python.ipynb index 21209905..f67f6a8d 100644 --- a/Gemma/[Gemma_2]Using_with_Ollama_Python.ipynb +++ b/Gemma/[Gemma_2]Using_with_Ollama_Python.ipynb @@ -41,7 +41,7 @@ "\n", "Author: Sitam Meur\n", "\n", - "* GitHub: [github.com/sitamgithub-MSIT](https://github.com/sitamgithub-MSIT/)\n", + "* GitHub: [github.com/sitammeur](https://github.com/sitammeur/)\n", "* X: [@sitammeur](https://x.com/sitammeur)\n", "\n", "Description: This notebook demonstrates how you can run inference on a Gemma model using [Ollama Python library](https://github.com/ollama/ollama-python). The Ollama Python library provides the easiest way to integrate Python 3.8+ projects with Ollama.\n", @@ -95,6 +95,7 @@ }, "outputs": [], "source": [ + "!sudo apt-get update && sudo apt-get install -y zstd\n", "!curl -fsSL https://ollama.com/install.sh | sh" ] }, diff --git a/Gemma/[Gemma_3]Local_Agentic_RAG.ipynb b/Gemma/[Gemma_3]Local_Agentic_RAG.ipynb index 363206d3..3e0344bb 100644 --- a/Gemma/[Gemma_3]Local_Agentic_RAG.ipynb +++ b/Gemma/[Gemma_3]Local_Agentic_RAG.ipynb @@ -44,6 +44,15 @@ "- GitHub: [lucifertrj](https://github.com/lucifertrj/)\n", "- Twitter: [TRJ_0751](https://x.com/trj_0751)\n", "\n", + "
\n",
+ " Run in Google Colab\n",
+ " | \n",
+ "