英文:
Llama_index issue behind HTTP request
问题
我遇到了一个问题,使用Llama_Index来使用先前为自定义内容ChatGPT查询生成的索引。
我使用以下代码生成了索引:
from llama_index import SimpleDirectoryReader, GPTListIndex, readers, GPTSimpleVectorIndex, LLMPredictor, PromptHelper, ServiceContext
from langchain import OpenAI
import sys
import os
def construct_index():
# 设置最大输入大小
max_input_size = 40960
# 设置输出令牌的数量
num_outputs = 20000
# 设置最大块重叠
max_chunk_overlap = 20
# 设置块大小限制
chunk_size_limit = 600
# 定义提示助手
prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
# 定义LLM
llm_predictor = LLMPredictor(llm=OpenAI(temperature=0.5, model_name="text-davinci-003", max_tokens=num_outputs))
documents = SimpleDirectoryReader("./data").load_data()
service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)
index = GPTSimpleVectorIndex.from_documents(documents, service_context=service_context)
index.save_to_disk('./index.json')
return index
要使用查询,我使用以下代码:
def ask_ai(query):
index = GPTSimpleVectorIndex.load_from_disk('./index.json')
response = index.query(query)
return response.response
这是一段常用的代码,实际上在虚拟环境中运行时(py ./index.py
)是有效的,只需添加一行来默认调用construct_index
或ask_ai
方法。
然而,当我尝试将其放在HTTP请求后面,例如在AWS Lambda或Flask API中,ask_ai
在load_from_disk
方法中失败,两种方法都出现相同的错误:
ERROR:app:Exception on /api/query [POST]
Traceback (most recent call last):
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\flask\app.py", line 2190, in wsgi_app
response = self.full_dispatch_request()
...
KeyError: 'docs'
对于Flask的实现,我从这里获取了代码:
https://github.com/bbachi/python-flask-restapi
只需在AppService中留下一个方法(ask_ai
),然后从控制器中的一个路由中调用它。AWS容器化Lambda的Docker文件如下:
FROM public.ecr.aws/lambda/python:3.10
COPY requirements.txt ./
RUN pip3 install -r requirements.txt
COPY myfunction.py ./
COPY index.json ./
CMD ["myfunction.lambda_handler"]
在这两种情况下,index.json
包含上面提供的代码生成的索引,而对于Lambda,myfunction.py
包含 lambda_handler
方法后面的 ask_ai
方法实现。
是否有人遇到过这样的问题?非常感谢提前提供的任何帮助!
英文:
I'm having an issue using Llama_Index to use an index previously generated for custom content ChatGPT queries.
I generated the index with the following code:
from llama_index import SimpleDirectoryReader, GPTListIndex, readers, GPTSimpleVectorIndex, LLMPredictor, PromptHelper, ServiceContext
from langchain import OpenAI
import sys
import os
def construct_index():
# set maximum input size
max_input_size = 40960
# set number of output tokens
num_outputs = 20000
# set maximum chunk overlap
max_chunk_overlap = 20
# set chunk size limit
chunk_size_limit = 600
# define prompt helper
prompt_helper = PromptHelper(max_input_size, num_outputs, max_chunk_overlap, chunk_size_limit=chunk_size_limit)
# define LLM
llm_predictor = LLMPredictor(llm=OpenAI(temperature=0.5, model_name="text-davinci-003", max_tokens=num_outputs))
documents = SimpleDirectoryReader("./data").load_data()
service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper)
index = GPTSimpleVectorIndex.from_documents(documents, service_context=service_context)
index.save_to_disk('./index.json')
return index
To use query something the code I use is the following:
def ask_ai(query):
index = GPTSimpleVectorIndex.load_from_disk('./index.json')
response = index.query(query)
return response.response
This is a popular code that in fact works when I run it in a virtual environment (py ./index.py) adding a line to call by default the construct_index or ask_ai methods.
However, when I try to put it behind an HTTP request, like in an AWS Lambda or Flask API, the ask_ai fails in the load_from_disk method for both approaches with the same error:
ERROR:app:Exception on /api/query [POST]
Traceback (most recent call last):
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\flask\app.py", line 2190, in wsgi_app
response = self.full_dispatch_request()
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\flask\app.py", line 1486, in full_dispatch_request
rv = self.handle_user_exception(e)
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\flask\app.py", line 1484, in full_dispatch_request
rv = self.dispatch_request()
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\flask\app.py", line 1469, in dispatch_request
return self.ensure_sync(self.view_functions[rule.endpoint])(**view_args)
File "C:\pmi\python-flask-restapi\app.py", line 17, in qyery_chatgpt
return appService.ask_ai(query)
File "C:\pmi\python-flask-restapi\app_service.py", line 10, in ask_ai
index = GPTSimpleVectorIndex.load_from_disk('index.json')
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\llama_index\indices\base.py", line 352, in load_from_disk
return cls.load_from_string(file_contents, **kwargs)
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\llama_index\indices\base.py", line 328, in load_from_string
return cls.load_from_dict(result_dict, **kwargs)
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\llama_index\indices\vector_store\base.py", line 260, in load_from_dict
return super().load_from_dict(result_dict, **config_dict, **kwargs)
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\llama_index\indices\base.py", line 304, in load_from_dict
docstore = DocumentStore.load_from_dict(result_dict[DOCSTORE_KEY])
File "C:\pmi\python-flask-restapi\flaskapi\lib\site-packages\llama_index\docstore.py", line 72, in load_from_dict
for doc_id, doc_dict in docs_dict["docs"].items():
KeyError: 'docs'
For the implementation of Flask I took it from here:
https://github.com/bbachi/python-flask-restapi
Just by leaving a single method in the AppService (the ask_ai) and invoking it from a single route in the controller. The docker file for the AWS container-based lambda is:
FROM public.ecr.aws/lambda/python:3.10
COPY requirements.txt ./
RUN pip3 install -r requirements.txt
COPY myfunction.py ./
COPY index.json ./
CMD ["myfunction.lambda_handler"]
In both cases the index.json holds the index generated with the above-provided code and for the lambda the myfunction.py contains the ask_ai method implementation behind the lambda_handler method.
Has someone faced such issue? any help is much appreciated in advance!
答案1
得分: 0
我使用Chat GPT来检查这个问题,它帮助我找到了原因,实际上这是我的一个大错误,因为我愚蠢地使用了与我后来用来加载的Llama_Index不同版本的索引,我没有注意到这一点。
如果有人遇到相同的问题,请尽量使用版本0.5.6。
英文:
I used Chat GPT to examine this issue and it led me to find out the cause, actually it was a big mistake on my side because I dumbly generated the index with a different version of Llama_Index than the one I was later using to load it, I haven't noticed that.
If someone's facing the same issue try to stick to version 0.5.6.
通过集体智慧和协作来改善编程学习和解决问题的方式。致力于成为全球开发者共同参与的知识库,让每个人都能够通过互相帮助和分享经验来进步。
评论