Also some feedback, a bit more technical, since I was trying to see how it works, more of a suggestion I suppose
It looks like you’re looping through the documents and asking it for known tags, right? ({str(db.current_library.tags)}.)
I don’t know if I would do this through a chat completion and a chat response, there are special functions for keyword-like searching, like embeddings. It’s a lot faster, and also probably way cheaper, since you’re paying barely anything for embeddings compared to chat tokens
Also some feedback, a bit more technical, since I was trying to see how it works, more of a suggestion I suppose
It looks like you’re looping through the documents and asking it for known tags, right? (
{str(db.current_library.tags)}.
)I don’t know if I would do this through a chat completion and a chat response, there are special functions for keyword-like searching, like embeddings. It’s a lot faster, and also probably way cheaper, since you’re paying barely anything for embeddings compared to chat tokens
So the common way to do something like this in AI would be to use Vectors and embeddings: https://platform.openai.com/docs/guides/embeddings
So - you’d ask for an embedding (A vector) for all your tags first. Then you ask for embeddings of your document.
Then you can do a Nearest Neighbor Search for the tags, and see how closely they match
Cool! But one problem: I’m not using OpenAI. It supports Mistral, ollama and xtekky’s gpt4free
It’s called embeddings in other models as well:
https://huggingface.co/blog/getting-started-with-embeddings
https://ollama.com/blog/embedding-models
Embeddings are not unique to openai.