Skip to main content

Vertex AI Search / Discovery Engine

This has had a few names etc. Vertex AI Search and Conversation but its this service: https://cloud.google.com/enterprise-search

Not to be confused with Vertex AI Vector Search which is the new name for Matching Engine (confused yet?)

Serverless chunking

Regardless, it is a very low code way to create RAG apps. It takes care of the chunking and indexing meaning you can just send in your documents and use them within your GenAI apps smoothly.

An example config is shown below:

kind: vacConfig
apiVersion: v1
vac:
sample_vector:
llm: vertex
agent: edmonbrain
memory:
- discovery_engine_vertex_ai_search:
vectorstore: vertex_ai_search # or discovery_engine

Since no chunking is necessary, documents are not indexed via the embedding service, but directly sent to the Vertex AI Search data store, with the same id as the VAC name e.g. sample_vector in above example. Make new Data Stores by creating new VACs.

Sunholo Multivac

Get in touch to see if we can help with your GenAI project.

Contact us

Other Links

Sunholo Multivac - GenAIOps

Copyright ©

Holosun ApS 2024