There are model management tools that leverage a RAG type system (Ollama uses “knowledge base”, Msty uses “knowledge stacks”, LM Studio only allows you to attach documents to prompts). However, it’s not really making them an “expert”.Okay thank you too for the information. Now, given the small size of the model, realistically, what can be done with such models? I guess they aren’t as capable as ChatGPT 4, or even 3.5, but can I feed them info, documentation, uni subjects, to make it an “expert” in that and let it evaluate me or help me with my studies? That’s a mere example.
Now, if this small models aren’t very useful / versatile, then I guess I won’t let this be a factor that influences me to get a 32GB machine instead of a 24GB one.
They rely on RAG (Retrieval Augmented Generation) in which, from your prompt, the model management tool attempts to retrieve relevant citations from the knowledge base (your documents) that enables the LLM to “augment” its “generation” of its response by looking at those citations first, before generating its answer from its parameters.
I’ve not found a truly reliable or accurate RAG system yet. I’d like to use LLMs that are experts on my stories and characters but I’ve not been happy with what they retrieve from a knowledge base filled with my documents.
I think the only real way of creating a workable “expert” is in fine tuning a model, but you need a pretty powerful computer to do that.
However, if all you need to do is attach one or two short documents (in my case a single scene or an act), you can attach them to the prompt and quiz the LLM about it and it’ll probably do a pretty good job. But giving them too much info leads to inaccuracies and overlooked info.
This is all from my experimentation with Ollama’s “knowledge base” and trying to have LM Studio manage large context. Msty might be different. I will be looking more at RAG options as I get time.