Hi everyone! Happy 2025! :tada: I am trying to se...
# community-help
d
Hi everyone! Happy 2025! 🎉 I am trying to setup Conversational RAG to enable retrieval of restaurants and corresponding products. My embeddings and semantic search seem to work as expected, except for the LLM response. I created a model with a simple prompt:
Copy code
conv_model = {
    "id": "conv-model-1",
    "model_name": "openai/gpt-4-turbo",
    "history_collection": "conversation_store",
    "api_key": "mykey",
    "system_prompt": "You are an assistant for question-answering of a food delivery service. You can only make conversations based on the provided context.",        
    "max_bytes": 16384
}
I am using Python API to get the results:
Copy code
result = client.collections['menus'].documents.search({
      'q': 'i am craving something sweet, which restaurants can i order from?',
      'query_by': 'embedding',
      'exclude_fields': 'embedding,
      'conversation_model_id': "conv-model-1",
      'conversation': True,
    })
But no matter what I try, I always get the same answer: "It looks like I don't have specific restaurant data available right now. I recommend checking the app or website you're using for a list of local restaurants that offer sweet dishes or desserts. You can usually find a variety of options there!" However, the
result['hits']
return correct results (a bunch of bakeries and candy stores). Is it some misconfiguration on my end?
j
May I know what model you’re using for the embedding?
d
@Jason Bosco I am using
ts/paraphrase-multilingual-mpnet-base-v2
j
Could you also share one sample document from your dataset?
d
I was able to make it work by increasing
max_bytes
and excluding more fields from the document. When the menus are too long I suppose the context gets cut off too early for the response to be able to use the documents properly
👌 1