diff --git a/docs/source/user_guide_rag.rst b/docs/source/user_guide_rag.rst index f8c2c4216..b51c019a1 100644 --- a/docs/source/user_guide_rag.rst +++ b/docs/source/user_guide_rag.rst @@ -223,7 +223,10 @@ it can be queried using the following: .. code:: python from neo4j_graphrag.llm import OllamaLLM - llm = OllamaLLM(model_name="orca-mini") + llm = OllamaLLM( + model_name="orca-mini", + # host="...", # when using a remote server + ) llm.invoke("say something") diff --git a/examples/customize/embeddings/ollama_embeddings.py b/examples/customize/embeddings/ollama_embeddings.py index 243cf4b6e..7a460146c 100644 --- a/examples/customize/embeddings/ollama_embeddings.py +++ b/examples/customize/embeddings/ollama_embeddings.py @@ -6,6 +6,7 @@ embeder = OllamaEmbeddings( model="", + # host="...", # if using a remote server ) res = embeder.embed_query("my question") print(res[:10]) diff --git a/examples/customize/llms/ollama_llm.py b/examples/customize/llms/ollama_llm.py index f815d4db9..dfac09db4 100644 --- a/examples/customize/llms/ollama_llm.py +++ b/examples/customize/llms/ollama_llm.py @@ -6,6 +6,7 @@ llm = OllamaLLM( model_name="", + # host="...", # if using a remote server ) res: LLMResponse = llm.invoke("What is the additive color model?") print(res.content)