Gets the list of supported models for the given vendor.

Request

POST https://{hostname}:{port}
    /cli/llm/models
Accept: {contentType}
Table 1. Header parameters
Parameter Type Required Description
Accept
  • application/json
  • application/yaml
true
This command takes a JSON request string or file. Use the following template for the request:
{
  "accessTokenEndpoint": "The uri for getting an access 
  token for authenticating with the LLM if neccessary.",
  "apiKey": "The API Key provided by the LLM Vendor. This 
  may be required for Vendors such as Watsonx and OpenAI.",
  "endpoint": "The base URI pointing to the LLM provider 
  API. Required.",
  "organization": "The organization if exists.",
  "project": "The project id if exists.",
  "vendor": "The name of the LLM Vendor. Options for this 
  are returned by the getVendors endpoint. Required."
}

Related CLI command: getLLMVendorModels.