- 1.18.0 (latest)
- 1.17.0
- 1.16.0
- 1.15.0
- 1.14.0
- 1.13.1
- 1.12.0
- 1.11.0
- 1.10.0
- 1.9.0
- 1.8.0
- 1.7.0
- 1.6.0
- 1.5.0
- 1.4.0
- 1.3.0
- 1.2.0
- 1.1.0
- 1.0.0
- 0.39.0
- 0.38.0
- 0.37.1
- 0.32.0
- 0.31.0
- 0.30.0
- 0.29.0
- 0.28.0
- 0.27.0
- 0.26.2
- 0.25.0
- 0.24.0
- 0.23.0
- 0.22.0
- 0.21.0
- 0.20.0
- 0.19.0
- 0.18.0
- 0.17.0
- 0.16.0
- 0.15.0
- 0.13.0
- 0.12.0
- 0.11.1
- 0.10.0
Reference documentation and code samples for the Google Cloud Ai Platform V1 Client class ComputeTokensRequest.
Request message for ComputeTokens RPC call.
Generated from protobuf message google.cloud.aiplatform.v1.ComputeTokensRequest
Namespace
Google \ Cloud \ AIPlatform \ V1Methods
__construct
Constructor.
Parameters | |
---|---|
Name | Description |
data |
array
Optional. Data for populating the Message object. |
↳ endpoint |
string
Required. The name of the Endpoint requested to get lists of tokens and token ids. |
↳ instances |
array<Google\Protobuf\Value>
Optional. The instances that are the input to token computing API call. Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models. |
↳ model |
string
Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers//models/ |
↳ contents |
array<Content>
Optional. Input content. |
getEndpoint
Required. The name of the Endpoint requested to get lists of tokens and token ids.
Returns | |
---|---|
Type | Description |
string |
setEndpoint
Required. The name of the Endpoint requested to get lists of tokens and token ids.
Parameter | |
---|---|
Name | Description |
var |
string
|
Returns | |
---|---|
Type | Description |
$this |
getInstances
Optional. The instances that are the input to token computing API call.
Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models.
Returns | |
---|---|
Type | Description |
Google\Protobuf\Internal\RepeatedField |
setInstances
Optional. The instances that are the input to token computing API call.
Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models.
Parameter | |
---|---|
Name | Description |
var |
array<Google\Protobuf\Value>
|
Returns | |
---|---|
Type | Description |
$this |
getModel
Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers//models/
Returns | |
---|---|
Type | Description |
string |
setModel
Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers//models/
Parameter | |
---|---|
Name | Description |
var |
string
|
Returns | |
---|---|
Type | Description |
$this |
getContents
Optional. Input content.
Returns | |
---|---|
Type | Description |
Google\Protobuf\Internal\RepeatedField |
setContents
Optional. Input content.
Parameter | |
---|---|
Name | Description |
var |
array<Content>
|
Returns | |
---|---|
Type | Description |
$this |
static::build
Parameters | |
---|---|
Name | Description |
endpoint |
string
Required. The name of the Endpoint requested to get lists of tokens and token ids. Please see LlmUtilityServiceClient::endpointName() for help formatting this field. |
instances |
array<Google\Protobuf\Value>
Optional. The instances that are the input to token computing API call. Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models. |
Returns | |
---|---|
Type | Description |
ComputeTokensRequest |