From d1f9edb3421b7e8f5bccc38696b63bab4450e465 Mon Sep 17 00:00:00 2001 From: Bharat Kumar Jain <33486187+BharatKJain@users.noreply.github.com> Date: Sun, 9 Mar 2025 17:11:59 +0530 Subject: [PATCH 1/2] Update local-rag-app-llama3.2.ipynb, fix _host Signed-off-by: Bharat Kumar Jain <33486187+BharatKJain@users.noreply.github.com> --- local-rag-app-llama3.2.ipynb | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/local-rag-app-llama3.2.ipynb b/local-rag-app-llama3.2.ipynb index ec1ce59..c8e9b83 100644 --- a/local-rag-app-llama3.2.ipynb +++ b/local-rag-app-llama3.2.ipynb @@ -113,7 +113,7 @@ " VALUES (\n", " %(title)s,\n", " %(content)s,\n", - " ollama_embed('nomic-embed-text', concat(%(title)s, ' - ', %(content)s), _host=>'http://ollama:11434')\n", + " ollama_embed('nomic-embed-text', concat(%(title)s, ' - ', %(content)s), host => 'http://ollama:11434')\n", " )\n", " \"\"\", doc)\n", "\n", @@ -212,7 +212,7 @@ " \n", "# Embed the query using the ollama_embed function\n", "cur.execute(\"\"\"\n", - " SELECT ollama_embed('nomic-embed-text', %s, _host=>'http://ollama:11434');\n", + " SELECT ollama_embed('nomic-embed-text', %s, host => 'http://ollama:11434');\n", "\"\"\", (query,))\n", "query_embedding = cur.fetchone()[0]\n", "\n", @@ -267,7 +267,7 @@ "\n", "# Generate the response using the ollama_generate function\n", "cur.execute(\"\"\"\n", - " SELECT ollama_generate('llama3.2', %s, _host=>'http://ollama:11434');\n", + " SELECT ollama_generate('llama3.2', %s, host => 'http://ollama:11434');\n", "\"\"\", (f\"Query: {query}\\nContext: {context}\",))\n", " \n", "model_response = cur.fetchone()[0]\n", From da07feaab9efcfee90a0e5bcab42fdebfa946d33 Mon Sep 17 00:00:00 2001 From: Bharat Kumar Jain <33486187+BharatKJain@users.noreply.github.com> Date: Sun, 9 Mar 2025 17:16:55 +0530 Subject: [PATCH 2/2] Update local-rag-app-llama3.2.ipynb - add "ai." Signed-off-by: Bharat Kumar Jain <33486187+BharatKJain@users.noreply.github.com> --- local-rag-app-llama3.2.ipynb | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/local-rag-app-llama3.2.ipynb b/local-rag-app-llama3.2.ipynb index c8e9b83..becab29 100644 --- a/local-rag-app-llama3.2.ipynb +++ b/local-rag-app-llama3.2.ipynb @@ -113,7 +113,7 @@ " VALUES (\n", " %(title)s,\n", " %(content)s,\n", - " ollama_embed('nomic-embed-text', concat(%(title)s, ' - ', %(content)s), host => 'http://ollama:11434')\n", + " ai.ollama_embed('nomic-embed-text', concat(%(title)s, ' - ', %(content)s), host => 'http://ollama:11434')\n", " )\n", " \"\"\", doc)\n", "\n", @@ -212,7 +212,7 @@ " \n", "# Embed the query using the ollama_embed function\n", "cur.execute(\"\"\"\n", - " SELECT ollama_embed('nomic-embed-text', %s, host => 'http://ollama:11434');\n", + " SELECT ai.ollama_embed('nomic-embed-text', %s, host => 'http://ollama:11434');\n", "\"\"\", (query,))\n", "query_embedding = cur.fetchone()[0]\n", "\n", @@ -267,7 +267,7 @@ "\n", "# Generate the response using the ollama_generate function\n", "cur.execute(\"\"\"\n", - " SELECT ollama_generate('llama3.2', %s, host => 'http://ollama:11434');\n", + " SELECT ai.ollama_generate('llama3.2', %s, host => 'http://ollama:11434');\n", "\"\"\", (f\"Query: {query}\\nContext: {context}\",))\n", " \n", "model_response = cur.fetchone()[0]\n",