[Question]: vector_store.add(nodes) gives me error = opensearchpy.exceptions.ConnectionTimeout: ConnectionTimeout caused by - TimeoutError() #13110
Labels
question
Further information is requested
Question Validation
Question
I am trying to add nodes in vectorsearch vector database, I am following this-:
https://docs.llamaindex.ai/en/stable/examples/low_level/ingestion/
endpoint = f"https://admin:admin@0.0.0.0:9200"
idx = "sample-index"
text_field = "text"
embedding_field = "vector_field"
client = OpensearchVectorClient(
endpoint, idx, dim=384, embedding_field=embedding_field, text_field=text_field, use_ssl = False, verify_certs = False
)
vector_store = OpensearchVectorStore(client)
#storage_context = StorageContext.from_defaults(vector_store=vector_store)
documents = SimpleDirectoryReader("/home/seceon/opensearch_lm_index/textdocs").load_data()
embeddings = HuggingFaceEmbedding()
splitter = SentenceSplitter(
chunk_size=700,
chunk_overlap=300,
)
nodes = splitter.get_nodes_from_documents(documents)
llm = HuggingFaceLLM(
model_name=model_id,
tokenizer_name=model_id,
context_window=3900,
max_new_tokens=256,
model_kwargs={"torch_dtype": torch.float16},
# tokenizer_kwargs={},
generate_kwargs={"temperature": 0.3, "top_p": 0.95},
device_map="auto",
)
Settings.llm = llm
Settings.embed_model = embeddings
Settings.context_window=4096,
Settings.num_output=1000,
Settings.chunk_overlap_ratio=0.3,
Settings.chunk_size_limit=None
for node in nodes:
node_embedding = embed_model.get_text_embedding(
node.get_content(metadata_mode="all")
)
node.embedding = node_embedding
vector_store.add(nodes)
#######################################################
Traceback (most recent call last):
File "/home/seceon/opensearch_lm_index/koshish.py", line 90, in
vector_store.add(nodes)
File "/usr/local/lib/python3.10/site-packages/llama_index/vector_stores/opensearch/base.py", line 476, in add
return asyncio.get_event_loop().run_until_complete(
File "/usr/local/lib/python3.10/asyncio/base_events.py", line 646, in run_until_complete
return future.result()
File "/usr/local/lib/python3.10/site-packages/llama_index/vector_stores/opensearch/base.py", line 492, in async_add
await self._client.index_results(nodes)
File "/usr/local/lib/python3.10/site-packages/llama_index/vector_stores/opensearch/base.py", line 347, in index_results
return await _bulk_ingest_embeddings(
File "/usr/local/lib/python3.10/site-packages/llama_index/vector_stores/opensearch/base.py", line 110, in _bulk_ingest_embeddings
await client.indices.refresh(index=index_name)
File "/usr/local/lib/python3.10/site-packages/opensearchpy/_async/client/indices.py", line 92, in refresh
return await self.transport.perform_request(
File "/usr/local/lib/python3.10/site-packages/opensearchpy/_async/transport.py", line 428, in perform_request
raise e
File "/usr/local/lib/python3.10/site-packages/opensearchpy/_async/transport.py", line 390, in perform_request
status, headers_response, data = await connection.perform_request(
File "/usr/local/lib/python3.10/site-packages/opensearchpy/_async/http_aiohttp.py", line 329, in perform_request
raise ConnectionTimeout("TIMEOUT", str(e), e)
opensearchpy.exceptions.ConnectionTimeout: ConnectionTimeout caused by - TimeoutError()
The text was updated successfully, but these errors were encountered: