PromptTokensDetails

class oci.generative_ai_inference.models.PromptTokensDetails(**kwargs)

Bases: object

Breakdown of tokens used in the prompt.

Methods

__init__(**kwargs) Initializes a new PromptTokensDetails object with values from keyword arguments.

Attributes

cached_tokens Gets the cached_tokens of this PromptTokensDetails.
__init__(**kwargs)

Initializes a new PromptTokensDetails object with values from keyword arguments. The following keyword arguments are supported (corresponding to the getters/setters of this class):

Parameters:cached_tokens (int) – The value to assign to the cached_tokens property of this PromptTokensDetails.
cached_tokens

Gets the cached_tokens of this PromptTokensDetails. Cached tokens present in the prompt.

Returns:The cached_tokens of this PromptTokensDetails.
Return type:int