Hi Stephen
All the models might not need high end GPU.
Some of them are available with Serverless APIs, internal computes at Azure (OpenAI models) and some can run on even memory optimized computes.
You can filter out model with Serverless APIs with "Deployment Option= Serverless API" filter.
Yes, but you might to need to raise GPU for certain LLM models as the deployment requires.
Reference- https://learn.microsoft.com/en-us/azure/machine-learning/concept-model-catalog?view=azureml-api-2
Hope it addresses your query.
Please accept this answer and say yes if it helped.
Thank you.