LlamaCPP #9023
Replies: 1 comment 1 reply
-
🤖 Hello @pjbruno327, I'm a friendly bot here to assist with your issues and concerns related to the LlamaIndex repository. While we wait for a human maintainer, I'll do my best to help with bug fixes, answer your questions, and help you become a contributor. Let's go ahead and dive into your issue. Yes, it is possible to swap the OpenAI LLM with llamaCPP in the LlamaIndex codebase. The LlamaIndex codebase is designed to work with multiple language learning models (LLMs), including OpenAI LLM and llamaCPP. To use llamaCPP instead of OpenAI LLM, you need to initialize the llama_cpp = LlamaCPP(model_url="https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML/resolve/main/llama-2-13b-chat.ggmlv3.q4_0.bin") In this example, the As for using the RAG (Retrieval-Augmented Generation) model with multiple documents, I wasn't able to find a direct answer within the repository. However, the LlamaIndex codebase provides several techniques for integrating the LlamaCPP class with multiple documents using the RAG model. These techniques include decoupling chunks used for retrieval vs. chunks used for synthesis, structured retrieval for larger document sets, dynamically retrieving chunks depending on your task, and optimizing context embeddings. You can find more details about these techniques in the production_rag.md document in the repository. I hope this helps! If you have any further questions or need more clarification, please let me know. SourcesThis response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research. Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant. If you want to continue the conversation, start your reply with @dosu-bot. |
Beta Was this translation helpful? Give feedback.
-
Is there a way to swap the OpenAI LLM with llamaCPP to use rag with multiple documents?
Beta Was this translation helpful? Give feedback.
All reactions