GraphRAG with Ollama - Install Local Models for RAG - Easiest Tutorial

  Рет қаралды 16,070

Fahd Mirza

Fahd Mirza

Күн бұрын

Пікірлер: 97
@fahdmirza
@fahdmirza 5 ай бұрын
Watch More GraphRAG Videos: 🔥GraphRAG with Ollama - Install Local Models for RAG - Easiest Tutorial kzbin.info/www/bejne/bIrYZ32mgbKKjNEsi=ONzq5rT1OSd0l4mD 🔥Install GraphRAG Locally - Build RAG Pipeline with Local and Global Search kzbin.info/www/bejne/iaqYfGl3rplprrcsi=g5eKWBsWg6zPaN7a 🔥GraphRAG with Groq - Install Locally with Local and Global Search kzbin.info/www/bejne/rpyneKOIasxsepIsi=QVfnD5tUSnxvPhAH 🔥GraphRAG with Llama.cpp Locally with Groq kzbin.info/www/bejne/b3jTY4SlZrN0ibs
@jianjieyin
@jianjieyin 5 ай бұрын
GraphRAG with Ollama, entity_extraction directory is not empty but errors come... Columns must be same length as key . How to solve?
@optimistic_dipak8632
@optimistic_dipak8632 2 ай бұрын
Excellent Video, Very Helpful!! Subscribed! Thank you.
@fahdmirza
@fahdmirza 2 ай бұрын
Awesome, thank you!
@optimistic_dipak8632
@optimistic_dipak8632 2 ай бұрын
​@@fahdmirza Hi Fahd, I am getting an error during the entity generation step. Below are few lines of that: "type": "error", "data": "Error Invoking LLM", "stack": "Traceback (most recent call last): File \"/scratch/dmeher/custom_env/miniforge/envs/graphrag_env/lib/python3.10/site-packages/httpx/_transports/default.py\", line 72, in map_httpcore_exceptions yield File \"/scratch/dmeher/custom_env/miniforge/envs/graphrag_env/lib/python3.10/site-packages/httpx/_transports/default.py\", line 377, in handle_async_request resp = await self._pool.handle_async_request(req) File \"/scratch/dmeher/custom_env/miniforge/envs/graphrag_env/lib/python3.10/site-packages/httpcore/_async/connection_pool.py\", line 216, in handle_async_request raise exc from None File \"/scratch/dmeher/custom_env/miniforge/envs/graphrag_env/lib/python3.10/site-packages/httpcore/_async/connection_pool.py\", line 196, in handle_async_request response = await connection.handle_async_request( File \"/scratch/dmeher/custom_env/miniforge/envs/graphrag_env/lib/python3.10/site-packages/httpcore/_async/connectio Any help with this would be greatly appreciated. Thanks!
@sergeziehi4816
@sergeziehi4816 5 ай бұрын
Graph RAG cost a lot indeed on API calls. One of your best video I do believe.thanks a lot
@fahdmirza
@fahdmirza 5 ай бұрын
thank you
@vinp2007
@vinp2007 3 ай бұрын
Hi great Video, i had a question if you already processed a PDF document for example however few weeks down the line you need to modify the pdf, how would you update the graph to make sure that the existing pdf is cleared and the updated one is the one that is being searched?
@TelB
@TelB 3 ай бұрын
I was having the api issue and also was in the 'ragtest' folder. I ran 'python3 -m graphrag.index --root .' whilst in the /ragtest directory instead of 'python3 -m graphrag.index --root ./ragtest' (15:14mins) and it seemed to work. Thank you for this video Fahd. Appreciate all the work and sharing!
@fahdmirza
@fahdmirza 3 ай бұрын
Thank you
@georgeknerr
@georgeknerr 5 ай бұрын
Excellent work - got a working example going!
@SaddamBinSyed
@SaddamBinSyed 4 ай бұрын
Hi @Fahad, This is simply excellent stuff. keep going
@Ayush-tl3ny
@Ayush-tl3ny 5 ай бұрын
Thank you so much for this video! You are Awesome ❤
@fahdmirza
@fahdmirza 5 ай бұрын
thank you
@arwaalyahyai8097
@arwaalyahyai8097 Ай бұрын
anyone has a problem with --method local? is ollama support embedding or not?
@lisag.9863
@lisag.9863 4 ай бұрын
Thank you for the great video! I got an error that says that "No text files found in input" even though my input does have a clear *.txt file. Do you know what could be the problem?
@IBCagents
@IBCagents 5 ай бұрын
Hi Fahd …, 1. Where does graphrag store the vectors and graphs in? I.e on local machine… 2. how do we transfer the entire graphrag app from the local machine to into the cloud….once we are done with ingestion and testing
@fahdmirza
@fahdmirza 5 ай бұрын
It has its own built-in vector store. For migration, I would suggest installing it from scratch in cloud.
@framefact4636
@framefact4636 4 ай бұрын
Thank you!, I am curious about visualize the knowledge graph, how to visualize it?
@xiaoluwang5019
@xiaoluwang5019 2 ай бұрын
22:50:49,145 graphrag.index.verbs.graph.clustering.cluster_graph WARNING Graph has no nodes 22:50:49,146 datashaper.workflow.workflow ERROR Error executing verb "cluster_graph" in create_base_entity_graph: Columns must be same length as key Traceback (most recent call last): a mistake happens like that? what's the problem may
@xiaoluwang5019
@xiaoluwang5019 2 ай бұрын
is that beacause of the Ollama model? or other reason
@fahdmirza
@fahdmirza 2 ай бұрын
yes likely
@aliyoussef7897
@aliyoussef7897 4 ай бұрын
nice! Thanks for sharing. in case we need to use local embedding model but not as a service (without Ollama), do we still need to pass the api_base?
@fahdmirza
@fahdmirza 3 ай бұрын
Yes, correct
@TheStuzenz
@TheStuzenz 5 ай бұрын
Nice video Fahd - GraphRAG looks really good! I plan on trying it out tonight. The querying against it looks quite expensive though. I wonder if they have built in any caching approach with the query engine. I guess I better do some reading.
@davidtindell950
@davidtindell950 5 ай бұрын
Thank You from a NEW Subscriber !
@fahdmirza
@fahdmirza 5 ай бұрын
Awesome, thank you!
@mikew2883
@mikew2883 5 ай бұрын
Excellent tutorial! I was wondering if you had a chance to work with the "graphrag-accelerator" Github project that Microsoft also put out. It says it can be used as an API that has all the GraphRAG functionality but in an API.
@fahdmirza
@fahdmirza 5 ай бұрын
I think graphrag-accelerator requires Azure. If its API based, I would rather go directly to OpenAI and I have already done a video on it.
@vitaliiturchenko8101
@vitaliiturchenko8101 5 ай бұрын
Thanks Fahd for your hard work! Very interesting!!! 1) is it possible to link GraphRag to the local ChromaDB database ? 2) local search also works in your method or only global search ?
@fahdmirza
@fahdmirza 5 ай бұрын
thaks. You would have to hack the source code to change the vector store. Yes local search also worked. Just have to replace global keyword with local.
@TheMariolino2005
@TheMariolino2005 5 ай бұрын
Great job! What if I want to add another document to the GraphRAG? Should I repeat the --init procedure or is there any other method? Great video, thank you.
@fahdmirza
@fahdmirza 5 ай бұрын
Yes, you would have to run the index procedure. Thanks.
@Salionca
@Salionca 5 ай бұрын
Thank you for the video.
@fahdmirza
@fahdmirza 5 ай бұрын
You're welcome
@giantworks1366
@giantworks1366 4 ай бұрын
Thank you for sharing the code.
@padhuLP
@padhuLP 5 ай бұрын
Good tutorial. Thank you for sharing the code.
@fahdmirza
@fahdmirza 5 ай бұрын
Thank you
@tollington9414
@tollington9414 5 ай бұрын
Good stuff!
@fahdmirza
@fahdmirza 5 ай бұрын
thanks
@AdityaSingh-in9lr
@AdityaSingh-in9lr 5 ай бұрын
hey, i got it working, but it is giving out of context answers when I do local search, any idea what could be wrong?
@zhengwu-jw6fm
@zhengwu-jw6fm 5 ай бұрын
When run the code 'python3 -m graphing.index --root./rattiest',showers occurred during the pipeline run, See logs for more details.What to solve this problem?
@fahdmirza
@fahdmirza 5 ай бұрын
plz check the logs in output directory and see what the error is. Also make sure that command is correct
@ShudhanshuShrotriya
@ShudhanshuShrotriya 12 күн бұрын
@@fahdmirza Hey Fahd, amazing video. I also received the same error. And upon checking the output log, I found this: datashaper.workflow.workflow ERROR Error executing verb "cluster_graph" in create_base_entity_graph: Columns must be same length as key Any idea as to why this error arises?
@revanthphanisaimedukonduru1177
@revanthphanisaimedukonduru1177 5 ай бұрын
Thanks for latest information, Can you please also add reference for this point , "GraphRAG don't support if its less than 32k context?" 7:22
@fahdmirza
@fahdmirza 5 ай бұрын
That's on basis of trial at the moment of creating video.
@shameekm2146
@shameekm2146 5 ай бұрын
Thanks for the video. Can i use mxbai from ollama for embedding purposes... or is there a limitation on that?
@fahdmirza
@fahdmirza 5 ай бұрын
sure you can use it.
@khriza4991
@khriza4991 5 ай бұрын
Thank you for the video. I'm facing the same error as another commenter mentioned: '❌ Errors occurred during the pipeline run, see logs for more details.' Where can I find the logs?
@fahdmirza
@fahdmirza 5 ай бұрын
Sure, go to this directory ~/ragtest/output/20240711-055438/reports . The date directory would vary as per your run. You would log files there. Thanks.
@jiangnanfan8944
@jiangnanfan8944 5 ай бұрын
@@fahdmirza raise ValueError(\"Columns must be same length as key\") ValueError: Columns must be same length as key ", "source": "Columns must be same length as key", "details": null , I FACE SAME ERROR , AND I FOUND THE LOG FILES , THEY SAID
@Gadgetwars
@Gadgetwars 5 ай бұрын
@@jiangnanfan8944 I also face the same error "ValueError(\"Columns must be same length as key\", "details": null)
@giantworks1366
@giantworks1366 4 ай бұрын
@@Gadgetwars same error :(
@chrishau5556
@chrishau5556 5 ай бұрын
Does this solution still works for anybody ?
@richardobiri2642
@richardobiri2642 4 ай бұрын
Awesome
@fahdmirza
@fahdmirza 4 ай бұрын
thanks mate
@fahdmirza
@fahdmirza 5 ай бұрын
🔥Install GraphRAG Locally - Build RAG Pipeline with Local and Global Search kzbin.info/www/bejne/iaqYfGl3rplprrcsi=f-o9SyqE62OgNU14
@am4_m037
@am4_m037 3 ай бұрын
I keep seeing this any idea ? Errors occurred during the pipeline run, see logs for more details
@fahdmirza
@fahdmirza Ай бұрын
There are similar videos on the channel, please search on the channel, thanks.
@arwaalyahyai8097
@arwaalyahyai8097 Ай бұрын
I could not find openai_embeddings_llm.py in my machine, anyone can help please?
@fahdmirza
@fahdmirza Ай бұрын
have you cloned repo?
@arwaalyahyai8097
@arwaalyahyai8097 Ай бұрын
@@fahdmirza yes, I found the file, but now i got another error which is error invoking LLM, do you have idea how to solve it?
@codelucky
@codelucky 5 ай бұрын
Can you create a video on how to use GraphRAG with the GROQ API? Looks like nobody has done it yet. Thank you.
@fahdmirza
@fahdmirza 5 ай бұрын
yeah just did. Thanks.
@codelucky
@codelucky 5 ай бұрын
@@fahdmirza Thanks, I appreciate your work.
@sharankumar31
@sharankumar31 5 ай бұрын
Kindly could you show, how to use this Graph RAG with CSV data. Will be super helpful
@fahdmirza
@fahdmirza 5 ай бұрын
Its the same process as any data. The cleaner your data is, the better your responses will be.
@IBCagents
@IBCagents 5 ай бұрын
Langchain+neo4j+chroma = MS graphrag …. Correct?
@fahdmirza
@fahdmirza 5 ай бұрын
Please explore this repo github.com/microsoft/graphrag for underlying tech. Thanks.
@themax2go
@themax2go 5 ай бұрын
@fahdmirza FYI on your webpage linked w/ the commands and code snippets for this vid, you have "model: nomic_embed_text" yet "ollama pull nomic-embed-text" which leads to: Error embedding chunk {'OpenAIEmbedding': 'Error code: 404 - {\'error\': "model \'nomic_embed_text\' not found, try pulling it first"}'}
@aa-xn5hc
@aa-xn5hc 5 ай бұрын
API key for Ollama should be "ollama". also, no need to do the embeddings locally because their cost is not high. The main objective should be to to do the LLM part with Ollama and then enquire both globally and locally.
@fahdmirza
@fahdmirza 5 ай бұрын
That can be done too in various ways, but the purpose of this video to do it all in Ollama. Thanks for comment.
@davidbeauchamp5867
@davidbeauchamp5867 5 ай бұрын
would you change this in the .env file or directly in the setting.yaml. I have the same issue as above where _config.py requires API key
@narendrasingh-tg1mb
@narendrasingh-tg1mb 5 ай бұрын
hi fahd thanks for video, getting this error : File "C:\Users\Narendrasingh\.conda\envs\graphollama\Lib\site-packages\graphrag\config\create_graphrag_config.py", line 229, in create_graphrag_config raise ApiKeyMissingError graphrag.config.errors.ApiKeyMissingError: API Key is required for Completion API. Please set either the OPENAI_API_KEY, GRAPHRAG_API_KEY or GRAPHRAG_LLM_API_KEY environment variable. ⠋ GraphRAG Indexer
@davidbeauchamp5867
@davidbeauchamp5867 5 ай бұрын
Same issue here...below states to use "ollama" as API key. In which file should this be indicated?
@EngineerMustaphaSahli
@EngineerMustaphaSahli 5 ай бұрын
Thanks for sharing! ... Anyone else suffering from this error: "openai.APITimeoutError: Request timed out." ??
@YoussefMohamed-fn6wl
@YoussefMohamed-fn6wl 5 ай бұрын
first of all thank you, ZeroDivisionError: Weights sum to zero, can't be normalized when using local method
@fahdmirza
@fahdmirza 5 ай бұрын
which model you are using?
@aravindchakrahari8966
@aravindchakrahari8966 5 ай бұрын
I got the same error as well while using local method. And also, Error embedding chunk {'OpenAIEmbedding': "'NoneType' object is not iterable"} I am using mistral and nomic-embed-text:latest for embeddings.
@ayushjadia6527
@ayushjadia6527 5 ай бұрын
I am also getting same error while using local method
@Ayush-tl3ny
@Ayush-tl3ny 5 ай бұрын
same error with groq api llama3 8b and nomic embed text, any solution to this?
@gnsnwr
@gnsnwr 5 ай бұрын
Issue is that `--method local` does not work out of the box with open source embedding models. It is because of the way how OpenAI's `text-embedding-3-small` model is working. It is using token IDs as input, while open source models like `nomic-embed-text` are working with text as input. So you need to convert token IDs to text before using open source models. Solution is to add one line to package's `graphrag/query/llm/oai/embedding.py` "embed" function : ```python ... def embed(self, text: str, **kwargs: Any) -> list[float]: """ Embed text using OpenAI Embedding's sync function. For text longer than max_tokens, chunk texts into max_tokens, embed each chunk, then combine using weighted average. Please refer to: github.com/openai/openai-cookbook/blob/main/examples/Embedding_long_inputs.ipynb """ token_chunks = chunk_text( text=text, token_encoder=self.token_encoder, max_tokens=self.max_tokens ) chunk_embeddings = [] chunk_lens = [] for chunk in token_chunks: # decode chunk from token ids to text (added line after row 83) chunk = self.token_encoder.decode(chunk) try: embedding, chunk_len = self._embed_with_retry(chunk, **kwargs) chunk_embeddings.append(embedding) chunk_lens.append(chunk_len) # TODO: catch a more specific exception except Exception as e: # noqa BLE001 self._reporter.error( message="Error embedding chunk", details={self.__class__.__name__: str(e)}, ) continue chunk_embeddings = np.average(chunk_embeddings, axis=0, weights=chunk_lens) chunk_embeddings = chunk_embeddings / np.linalg.norm(chunk_embeddings) return chunk_embeddings.tolist() ... ```
@shawnkratos1347
@shawnkratos1347 5 ай бұрын
You only did global search what about local. That is only half the rag. I got this far and thought you figured it out
@fahdmirza
@fahdmirza 5 ай бұрын
Its the same process, you just need to replace global with local
@shawnkratos1347
@shawnkratos1347 5 ай бұрын
@@fahdmirza no it fails to build community reports:just tested again with mistral to make sure i have the exact same set up as you. look in the index-engine.log. 5:48:44,679 graphrag.index.graph.extractors.community_reports.community_reports_extractor ERROR error generating community report Traceback (most recent call last): File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/index/graph/extractors/community_reports/community_reports_extractor.py", line 58, in __call__ await self._llm( File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/json_parsing_llm.py", line 34, in __call__ result = await self._delegate(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_token_replacing_llm.py", line 37, in __call__ return await self._delegate(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_history_tracking_llm.py", line 33, in __call__ output = await self._delegate(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/caching_llm.py", line 104, in __call__ result = await self._delegate(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 177, in __call__ result, start = await execute_with_retry() File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 159, in execute_with_retry async for attempt in retryer: File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/asyncio/__init__.py", line 166, in __anext__ do = await self.iter(retry_state=self._retry_state) File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/asyncio/__init__.py", line 153, in iter result = await action(retry_state) File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/_utils.py", line 99, in inner return call(*args, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/tenacity/__init__.py", line 398, in self._add_action_func(lambda rs: rs.outcome.result()) File "/usr/lib/python3.10/concurrent/futures/_base.py", line 451, in result return self.__get_result() File "/usr/lib/python3.10/concurrent/futures/_base.py", line 403, in __get_result raise self._exception File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 165, in execute_with_retry return await do_attempt(), start File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/rate_limiting_llm.py", line 147, in do_attempt return await self._delegate(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/base/base_llm.py", line 48, in __call__ return await self._invoke_json(input, **kwargs) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/llm/openai/openai_chat_llm.py", line 90, in _invoke_json raise RuntimeError(FAILED_TO_CREATE_JSON_ERROR).......python -m graphrag.query --root . --method global "what are the top themes in this story?" INFO: Reading settings from settings.yaml creating llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_chat", 'model': 'mistral', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/v1', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': True, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25} SUCCESS: Global Search Response: In the story, the main themes revolve around the transition of young people from formal education to practical work, specifically through apprenticeship under Ebenezer Scrooge. This transition is evident in various scenes and actions [Data: Scenes (1, 2, 3); Actions (4)]. During their apprenticeship, the young people are engaged in specific tasks or responsibilities that are likely related to Scrooge's business [Data: Actions (1-5)]. It is also suggested that Scrooge may act as a mentor or supervisor to these apprentices during this period [Data: Relationships (1-23)]. The young people are involved in various activities related to their apprenticeship, which could include tasks such as bookkeeping, accounting, or business management [Data: Actions (1-5)]. However, the exact nature of these activities is not explicitly detailed in the provided data. It is important to note that the information provided is based on the analysis of multiple reports and does not necessarily cover all aspects of the story. For a more comprehensive understanding, additional research or analysis may be required. shawn@pop-os:~/Documents/GRAPHRAG$ python -m graphrag.query --root . --method local "who is scrooge, and what are his main relationships?" INFO: Reading settings from settings.yaml creating llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_chat", 'model': 'mistral', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/v1', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': True, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25} creating embedding llm client with {'api_key': 'REDACTED,len=56', 'type': "openai_embedding", 'model': 'nomic_embed_text', 'max_tokens': 4000, 'request_timeout': 180.0, 'api_base': 'localhost:11434/api', 'api_version': None, 'organization': None, 'proxy': None, 'cognitive_services_endpoint': None, 'deployment_name': None, 'model_supports_json': None, 'tokens_per_minute': 0, 'requests_per_minute': 0, 'max_retries': 10, 'max_retry_wait': 10.0, 'sleep_on_rate_limit_recommendation': True, 'concurrent_requests': 25} Error embedding chunk {'OpenAIEmbedding': 'Error code: 404 - {\'error\': "model \'nomic_embed_text\' not found, try pulling it first"}'} Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/__main__.py", line 75, in run_local_search( File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/cli.py", line 154, in run_local_search result = search_engine.search(query=query) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/structured_search/local_search/search.py", line 118, in search context_text, context_records = self.context_builder.build_context( File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/structured_search/local_search/mixed_context.py", line 139, in build_context selected_entities = map_query_to_entities( File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/context_builder/entity_extraction.py", line 55, in map_query_to_entities search_results = text_embedding_vectorstore.similarity_search_by_text( File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/vector_stores/lancedb.py", line 118, in similarity_search_by_text query_embedding = text_embedder(text) File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/context_builder/entity_extraction.py", line 57, in text_embedder=lambda t: text_embedder.embed(t), File "/home/shawn/.local/lib/python3.10/site-packages/graphrag/query/llm/oai/embedding.py", line 96, in embed chunk_embeddings = np.average(chunk_embeddings, axis=0, weights=chunk_lens) File "/home/shawn/.local/lib/python3.10/site-packages/numpy/lib/function_base.py", line 550, in average raise ZeroDivisionError( ZeroDivisionError: Weights sum to zero, can't be normalized
@Thinker-i8d
@Thinker-i8d 5 ай бұрын
perfect job. but when i try to use graphrag with ollama, error happened. logs.json shows: {"type": "error", "data": "Error Invoking LLM", "stack": "Traceback (most recent call last), and the index-engine.log shows:graphrag.index.reporting.file_workflow_callbacks INFO Error Invoking LLM does anyone know how to fix this error??
@karthickdurai2157
@karthickdurai2157 3 ай бұрын
same issue
@optimistic_dipak8632
@optimistic_dipak8632 2 ай бұрын
Reload the input
@themax2go
@themax2go 5 ай бұрын
python -m graphrag.query --root ./ --method local "explain relationships between the people in the story" leads to: ./graphrag/lib/python3.12/site-packages/numpy/lib/function_base.py", line 550, in average raise ZeroDivisionError(ZeroDivisionError: Weights sum to zero, can't be normalized - and before that: Error embedding chunk {'OpenAIEmbedding': "'NoneType' object is not iterable"}
@LuZhenxian
@LuZhenxian 5 ай бұрын
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/zhenxian/Documents/XXG/github/graphrag-main/graphrag/config/create_graphrag_config.py", line 231, in create_graphrag_config raise ApiKeyMissingError graphrag.config.errors.ApiKeyMissingError: API Key is required for Completion API. Please set either the OPENAI_API_KEY, GRAPHRAG_API_KEY or GRAPHRAG_LLM_API_KEY environment variable.
@donzhu4996
@donzhu4996 5 ай бұрын
got the same error
Local GraphRAG with LLaMa 3.1 - LangChain, Ollama & Neo4j
15:01
Coding Crash Courses
Рет қаралды 34 М.
Run ALL Your AI Locally in Minutes (LLMs, RAG, and more)
20:19
Cole Medin
Рет қаралды 301 М.
1% vs 100% #beatbox #tiktok
01:10
BeatboxJCOP
Рет қаралды 67 МЛН
So Cute 🥰 who is better?
00:15
dednahype
Рет қаралды 19 МЛН
How Strong Is Tape?
00:24
Stokes Twins
Рет қаралды 96 МЛН
Quilt Challenge, No Skills, Just Luck#Funnyfamily #Partygames #Funny
00:32
Family Games Media
Рет қаралды 55 МЛН
Local LightRAG: A GraphRAG Alternative but Fully Local with Ollama
18:55
Prompt Engineering
Рет қаралды 46 М.
Microsoft Trellis turns any image into a 3D Model INSTANTLY
3:10
halkony2012
Рет қаралды 3,8 М.
Unlimited AI Agents running locally with Ollama & AnythingLLM
15:21
Tim Carambat
Рет қаралды 165 М.
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
15 INSANE Use Cases for NEW Claude Sonnet 3.5! (Outperforms GPT-4o)
28:54
GraphRAG Advanced: Avoid Overspending with These Tips
12:41
Mervin Praison
Рет қаралды 9 М.
LangGraph Simplified: Master Custom AI Agent Creation
43:51
Data Centric
Рет қаралды 42 М.
1% vs 100% #beatbox #tiktok
01:10
BeatboxJCOP
Рет қаралды 67 МЛН