Discussion about this post

User's avatar
Anon's avatar

Largely the accepted way of doing this for now, but it's far from perfect. In my experience your system prompt has to be very exact in order to stop it hallucinating. The other problem is with chunking the data itself - if you're passing several large documents and the chunks are of set length, it can and often will infer bad context just based on chunk cutoffs. Small, semi-structured individual documents seem to perform far better than chunking.

Expand full comment
RLLMing in time.'s avatar

Can you an end to end project that covers:

RAG assistance In LLM based chatbots for context and reducing mistakes

Fine tuning open source LLM mode

Mixe media vector embedding

Re ranking things like that..

It will be onebigassproject for sure.

Expand full comment
7 more comments...

No posts