Comments

You must log in or register to comment.

_Arsenie_Boca_ t1_jdbsl4b wrote

What are the costs for all the services? I assume GPT-4 is billed per request and Pinecone per hour?

3

Icko_ t1_jdecnjx wrote

Sure:

  1. Suppose you had 1 million embeddings of sentences, and one vector you want the closest sentence to. If the vectors were a single number, you could just do a binary search, and you'd be done. If they are higher dimensionality, it's a lot more involved. Pinecone is a paid product doing this. Faiss is a library by facebook, which is very good too, but is free.
  2. Recently, Facebook released the LLama models. They are large language models. ChatGPT is also a LLM, but after pretraining on a text corpus, you train it with human instructions, which is costly and time-consuming. Stanford took the LLama models, and trained them with ChatGPT. The result is pretty good not AS good, but pretty good. They called it "Alpaca".
2

saintshing t1_jdgwgt7 wrote

I heard of people talking about using ANNOY for approximate nearest neighbor search. How is ANNOY compared to pinecone and faiss? Are pinecone and faiss self-hostable?

1

Always1Max t1_jdkn18v wrote

could there be something like this, but for code?

1

fletchertyler914 t1_je9o94n wrote

I just found this a few days ago and actually used it as a prototype base to learn the ropes, so thanks op! I ended up gutting the ingest flow in favor of an additional upload api route to make it more flexible, but overall it was a good example/guide to follow. Nice work.

1