Replies: 1 comment 1 reply
-
🤖 Hello @azurewtl, I'm here to help you with your questions about the LlamaIndex repository. Let's dive into your queries. The The method As for the concern about memory consumption, it's a valid point. However, the design choice to store all {vector_id:node_id} mapping information in the index is a trade-off between memory usage and the ability to retrieve original data. If memory consumption becomes a problem, you might need to consider other strategies such as sharding the index across multiple machines, or using a more memory-efficient data structure for storing the index. In the context shared, the As for the I hope this helps! If you have any more questions, feel free to ask. Best, Sources
This response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research. Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant. If you want to continue the conversation, start your reply with @dosu-bot. |
Beta Was this translation helpful? Give feedback.
-
What is the intention of design of storing all {vector_id:node_id} mapping information in index such as vector_index, or is it necessary?
Shouldn't all the node information been storage in
index_store
? Otherwise it will eat up python memory stack if I instantiate multiple Index over the same storage_context, or with too many of small nodes?In vector index, I found
insert_nodes
takes two step, which make me really confused:Step[1]: update the mapping of
nodes_dict
inVectorStoreIndex._index_struct
Step[2]: update the data in physical database from
_storage_context.index_store
Beta Was this translation helpful? Give feedback.
All reactions