Replies: 1 comment
-
🤖 Hi @dinonovak! 👋 I'm Dosu, an AI bot here to help you with your LlamaIndex concerns while we wait for a human maintainer. I'm here to help solve bugs, answer questions, and assist you in contributing to this project. Let's tackle these tech hiccups together! 🚀 Based on the information available in the LlamaIndex repository, the As for the Regarding the interaction between LlamaIndex and the Mistral 7x8B Instruct served by vllm, I couldn't find any specific information in the repository. It's possible that you might need to customize the LLM definition for Mistral/Mixtral Instruct, but without more information, it's hard to say for sure. I hope this helps! If you have any more questions or if there's anything else you'd like me to look into, please let me know. SourcesThis response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research. Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant. If you want to continue the conversation, start your reply with @dosu-bot. |
Beta Was this translation helpful? Give feedback.
-
I am using mistral 8x7B served via vllm.
When using it in llama-index with OpenAIlike model definition it looks like it is not finishing messages with token.
DoI need to customize llm definition specifically for mistral/mixtral instruct ?
I am using model with is_chat_model = True and all works, but il looks like all questions are considered part of the same chat.
How can I force end of conversation and do I need to send token somehow?
Beta Was this translation helpful? Give feedback.
All reactions