(no title)
patresh | 2 years ago
Error handling when you get rate limited, the token has expired or the token length is too long would be problematic, and from a security point of view it requires your DB to directly call OpenAI which can also be risky.
Personally I haven't used that many Postgres extensions, so perhaps these risks are mitigated somehow that I don't know?
chuckhend|2 years ago
On Tembo cloud, we deploy this as part of the VectorDB and RAG Stacks. So you get a dedicated Postgres instance, and a container next to Postgres that hosts the text-to-embeddings transformers. The API calls/data never leave your namespace.
infecto|2 years ago