Google Vertex
The vertex provider is compatible with Google's Vertex AI offering, which offers access to models such as bison.
You can use it by specifying any of the available stable or latest model versions offered by Vertex AI. These include:
vertex:chat-bisonvertex:chat-bison@001vertex:chat-bison-32kvertex:chat-bison-32k@001vertex:codechat-bisonvertex:codechat-bison@001vertex:codechat-bison-32kvertex:codechat-bison-32k@001
Supported environment variables:
VERTEX_API_KEY(required) - gcloud API tokenVERTEX_PROJECT_ID(required) - gcloud project IDVERTEX_REGION- gcloud region, defaults tous-central1VERTEX_PUBLISHER- model publisher, defaults togoogleVERTEX_API_HOST- used to override the full Google API host, e.g. for an LLM proxy, defaults to{region}-aiplatform.googleapis.com
The Vertex provider also supports various configuration options such as context, examples, temperature, maxOutputTokens, and more, which can be used to customize the the behavior of the model like so:
providers:
- id: vertex:chat-bison-32k
config:
temperature: 0
maxOutputTokens: 1024