Showing posts with label LLM. Show all posts
Showing posts with label LLM. Show all posts

Thursday, July 6, 2023

deploy falcon 7b & 40b on amazon sagemaker example

 https://github.com/aws/amazon-sagemaker-examples/blob/main/inference/generativeai/llm-workshop/lab10-falcon-40b-and-7b/falcon-40b-deepspeed.ipynb 


https://youtu.be/-IV1NTGy6Mg 

https://www.philschmid.de/sagemaker-falcon-llm 

How cache can be enabled for embeded text as well for search query results in Azure AI ?

 Great question, Rahul! Caching in the context of Azure AI (especially when using **RAG pipelines with Azure OpenAI + Azure AI Search**) can...