Replies: 1 comment
-
🤖 Hello, Thank you for reaching out and providing detailed information about your issue. The In your case, you've set If your long text file is not being split into multiple nodes as expected, it could be due to the size of your text or the settings of your Also, please note that if the metadata length is longer than the chunk size, an error will be raised. If the resulting chunks are less than 50 tokens, a warning will be printed. Consider increasing the chunk size or decreasing the size of your metadata to avoid this. Here is a reference to a similar issue that was solved: [Question]: How to split or seperate nodes Doucment data?. The solution was to implement a function to split large documents into smaller chunks before passing the documents to the I hope this helps! If you have any other questions or need further clarification, please let me know. Best, Sources
This response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research. Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant. If you want to continue the conversation, start your reply with @dosu-bot. |
Beta Was this translation helpful? Give feedback.
-
Question Validation
Question
Hello,
I have dive in your docs but i have some issue related to chunk and text spilt. As I think, the text split will split 1 document to many node and will store in docstore. But when i trying to test with the file with long text. It doesn't happen.
Here is my code to trying to setting the Text spiliter :
text_splitter = TokenTextSplitter( separator=" ", chunk_size=5, chunk_overlap=1, backup_separators=["\n"] ) parser= SimpleNodeParser( text_splitter=text_splitter ) index = VectorStoreIndex( documents, service_context=service_context, show_progress=True )
Please help me
Beta Was this translation helpful? Give feedback.
All reactions