This is a quick prototype I built for semantic search and factual question answering using embeddings and GPT-3.
It tries to solve the LLM hallucination issue by guiding it only to answer questions from the given context instead of making things up. If you ask something not covered in an episode, it should say that it doesn't know rather than providing a plausible, but potentially incorrect response.
It uses Whisper to transcribe, text-embedding-ada-002 to embed, Pinecone.io to search, and text-davinci-003 to generate the answer.
More examples and explanations here: https://twitter.com/rileytomasek/status/1603854647575384067
Tip - you don’t actually need GPT-3 level embedding for a decent semantic search. Sentence transformers paired with one of their models is good enough.
I like this: https://huggingface.co/sentence-transformers/multi-qa-MiniLM... - since it’s very light.
Also, perhaps I am an idiot but I just used Postgres array field to store my embeddings array to keep things simple and free.