Gets the list of supported models for the given vendor.
Request
POST https://{hostname}:{port}
/cli/llm/models
Accept: {contentType}
| Parameter | Type | Required | Description |
|---|---|---|---|
| Accept |
|
true |
This command takes a JSON request string or file. Use the following template for the request:
{
"accessTokenEndpoint": "The uri for getting an access
token for authenticating with the LLM if neccessary.",
"apiKey": "The API Key provided by the LLM Vendor. This
may be required for Vendors such as Watsonx and OpenAI.",
"endpoint": "The base URI pointing to the LLM provider
API. Required.",
"organization": "The organization if exists.",
"project": "The project id if exists.",
"vendor": "The name of the LLM Vendor. Options for this
are returned by the getVendors endpoint. Required."
}
Related CLI command: getLLMVendorModels.