Technical Considerations
Embeddings
What is an embedding?
Embeddings are numerical representations of concepts converted to number sequences, which make it easy for computers to understand the relationships between those concepts. They are capable of capturing the context of a word in a document, its semantic and syntactic similarity, and its relation with other words.
How can I select different types of embeddings in Stack AI?
Embeddings can be selected by the user in two different section of Stack AI platform.
- The most intuitive place is in a
vector store
. As explained in section, the input will be vectorized and indexed in a vector database for later usage in an AI model (so only the relevant chunks of the input are sent to the LLM). Document search
elements are also customizable with respect to their embeddings.
Which are the models available?
Below a list of the embeddings models integrated into Stack AI’s platform.
MODEL | COMPANY | DESCRIPTION | LINK |
---|---|---|---|
text-embedding-3-large | OpenAI | newest and most performant embedding models are now available, with lower costs, higher multilingual performance, and new parameters to control the overall size. | More Info |
text-embedding-3-small | OpenAI | Outperforms previous OpenAI’s most capable model, Davinci, at most tasks, while being priced 99.8% lower | More Info |
text-embedding-ada-002 | OpenAI | Outperforms previous OpenAI’s most capable model, Davinci, at most tasks, while being priced 99.8% lower | More Info |
bert-base-cased | Embeddings based on Bidirectional Encoder Representations from Transformers (BERT) | More Info | |
palm2 | Vertex AI PaLM API supports Gecko for Embeddings | More Info | |
all-mpnet-base-v2 | Open Source | Sentence-transformers model that maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. | More Info |