Replies: 2 comments
-
You can see exactly how the the prompt is crafted by previewing it with
I haven't decided yet. I would like to leverage the fact that you can tag your instructions and add commentary, in hopes that it will create a much more accurate vectorization. Right now, the approach I have in my head is to vectorize only references and memoize the vectorization, rebuilding the index only when the references were actually changed. We can chunk the references and make them fit the model context constraints. I know RAG is something that will definitely be added sometime, but at the moment there are other more crucial features I would like to have that will help me work on the package, such as linked instructions. RAG will essentially try to free you from having to supply query tags, which is very lovely. Converted to discussion. |
Beta Was this translation helpful? Give feedback.
-
Thanks for the explantion. |
Beta Was this translation helpful? Give feedback.
-
I’ll install and try asap but wanted to ask, are “instructions” text that’s sent with the prompt as is, assuming it won’t be truncated to fit the context? I’m guessing with Gemini models and others with huge context window it’s not a problem but with ollama, which I believe defaults to 2k this could be an issue (I think gptel might have it hardcoded for 4k or 8k if I remember correctly).
And in this context, what are your plans for rag? I’ve been trying elysa for a while but lately it simply hangs vectorizing the prompt, or is very slow. But the idea sounds cool.
Beta Was this translation helpful? Give feedback.
All reactions