Build a multi-LLM chat application with Azure Container Apps

Ғылым және технология

In this demo, explore how to leverage GPU workload profiles in ACA to run your own model backend, and easily switch, compare, and speed up your inference times. You will also explore how to leverage LlamaIndex to ingest data on-demand, and host models using Ollama. Then finally, decompose the application as a set of microservices written in Python, deployed on ACA.
#microsoftreactor #multillm #llms #azurecontainerapps #azure #chatapp
[eventID:22137]

Пікірлер: 1

  • @changtimwu
    @changtimwuАй бұрын

    Excellent example. It appears that ACA is more portable than AWA when it comes to migrating the app to an on-premises/local infrastructure.

Келесі