To create a multimodal RAG (Retrieval-Augmented Generation) using multimodal embeddings, ensure that you have the correct models deployed and that they are accessible in your Azure environment. Here are some steps to consider:
- Deployment of Cohere Models: Make sure that the Cohere embedding models are provisioned as serverless API deployments. You can use an ARM/Bicep template for this task. If you are unable to find the serverless option, double-check that you are looking in the correct sections of the Azure Foundry portal or your project settings.
- Regional Availability: Confirm that both your Azure OpenAI resource and Azure AI Search service are created in the same region. While integrated vectorization does not strictly require this, having them in the same region can improve performance and reduce latency.
- Model Selection: When adding knowledge to your agent workflow in the Foundry portal, ensure you are selecting the correct embedding models that are supported for your project. The supported models include
text-embedding-ada-002,text-embedding-3-small, andtext-embedding-3-large, among others. - Public Access: Ensure that all resources have public access enabled so that the Azure portal can access them. This is crucial for the wizard to run successfully.
If you have followed these steps and are still encountering issues, consider reaching out to Azure support for further assistance.
References: