Discussion on OpenAI Embeddings Response Time and Alternatives
TLDR Walter discusses the varying response time when using OpenAI embeddings. Jason confirms this issue and Manish suggests using the E5 model in Typesense RC as a better alternative.
1
Jun 28, 2023 (3 months ago)
Walter
04:42 PMI've noticed that the response time of search requests can be anywhere from 1-4s when using the open ai embeddings. Has that been your experience as well?
Jason
04:50 PMWalter
04:53 PMJason
06:06 PMManish
07:05 PMWalter
07:22 PM1
Typesense
Indexed 2779 threads (79% resolved)
Similar Threads
Finding Similar Documents Using JSON and Embeddings
Manish wants to find similar JSON documents and asks for advice. Jason suggests using Sentence-BERT with vector query and provides guidance on working with OpenAI embeddings and Typesense. They discuss upcoming Typesense features and alternative models.
Optimum Cluster for 1M Documents with OpenAI Embedding
Denny inquired about the ideal cluster configuration for handling 1M documents with openAI embedding. Jason recommended a specific configuration, explained record size calculation, and clarified embedding generation speed factors and the conditions that trigger openAI.
Integrating OpenAI Embeddings with DocSearch Scraper
Marcos was looking for how to use OpenAI embeddings with DocSearch. Jason guided with an update to the scraper config, and suggested the GTE built-in model for generic use.