What is an embedding?

Embeddings are numerical representations of concepts converted to number sequences, which make it easy for computers to understand the relationships between those concepts. They are capable of capturing the context of a word in a document, its semantic and syntactic similarity, and its relation with other words.

How can I select different types of embeddings in Stack AI?

Embeddings can be selected by the user in two different section of Stack AI platform.

  • The most intuitive place is in a vector store. As explained in section, the input will be vectorized and indexed in a vector database for later usage in an AI model (so only the relevant chunks of the input are sent to the LLM).
  • Document search elements are also customizable with respect to their embeddings.

Which are the models available?

Below a list of the embeddings models integrated into Stack AI’s platform.

MODELCOMPANYDESCRIPTIONLINK
text-embedding-3-largeOpenAInewest and most performant embedding models are now available, with lower costs, higher multilingual performance, and new parameters to control the overall size.More Info
text-embedding-3-smallOpenAIOutperforms previous OpenAI’s most capable model, Davinci, at most tasks, while being priced 99.8% lowerMore Info
text-embedding-ada-002OpenAIOutperforms previous OpenAI’s most capable model, Davinci, at most tasks, while being priced 99.8% lowerMore Info
bert-base-casedGoogleEmbeddings based on Bidirectional Encoder Representations from Transformers (BERT)More Info
palm2GoogleVertex AI PaLM API supports Gecko for EmbeddingsMore Info
all-mpnet-base-v2Open SourceSentence-transformers model that maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.More Info