Wrapper for Python class TokenizerLike.
Summary
Functions
Python method TokenizerLike.apply_chat_template.
Python method TokenizerLike.convert_ids_to_tokens.
Python method TokenizerLike.convert_tokens_to_ids.
Python method TokenizerLike.convert_tokens_to_string.
Python method TokenizerLike.decode.
Python method TokenizerLike.encode.
Python method TokenizerLike.from_pretrained.
Python method TokenizerLike.get_added_vocab.
Python method TokenizerLike.get_vocab.
Constructs TokenizerLike.
Python method TokenizerLike.num_special_tokens_to_add.
Types
Functions
@spec all_special_ids(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec all_special_tokens(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec apply_chat_template(SnakeBridge.Ref.t(), [term()], [term()], keyword()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.apply_chat_template.
Parameters
messages(list(term()))tools(term() default: None)kwargs(term())
Returns
term()
@spec bos_token_id(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec convert_ids_to_tokens(SnakeBridge.Ref.t(), [integer()], [term()], keyword()) :: {:ok, [String.t()]} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.convert_ids_to_tokens.
Parameters
ids(list(integer()))skip_special_tokens(boolean() default: False)
Returns
list(String.t())
@spec convert_tokens_to_ids(SnakeBridge.Ref.t(), term(), keyword()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.convert_tokens_to_ids.
Parameters
tokens(term())
Returns
term()
@spec convert_tokens_to_string(SnakeBridge.Ref.t(), [String.t()], keyword()) :: {:ok, String.t()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.convert_tokens_to_string.
Parameters
tokens(list(String.t()))
Returns
String.t()
@spec decode(SnakeBridge.Ref.t(), term(), [term()], keyword()) :: {:ok, String.t()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.decode.
Parameters
ids(term())skip_special_tokens(boolean() default: False)
Returns
String.t()
@spec encode(SnakeBridge.Ref.t(), String.t(), [term()], keyword()) :: {:ok, [integer()]} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.encode.
Parameters
text(String.t())truncation(term() default: None)max_length(term() default: None)add_special_tokens(boolean() default: True)
Returns
list(integer())
@spec eos_token_id(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec from_pretrained(SnakeBridge.Ref.t(), term(), [term()], keyword()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.from_pretrained.
Parameters
path_or_repo_id(term())args(term())trust_remote_code(boolean() keyword-only default: False)revision(term() keyword-only default: None)download_dir(term() keyword-only default: None)kwargs(term())
Returns
term()
@spec get_added_vocab( SnakeBridge.Ref.t(), keyword() ) :: {:ok, %{optional(String.t()) => integer()}} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.get_added_vocab.
Returns
%{optional(String.t()) => integer()}
@spec get_vocab( SnakeBridge.Ref.t(), keyword() ) :: {:ok, %{optional(String.t()) => integer()}} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.get_vocab.
Returns
%{optional(String.t()) => integer()}
@spec is_fast(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec max_token_id(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec new( [term()], keyword() ) :: {:ok, SnakeBridge.Ref.t()} | {:error, Snakepit.Error.t()}
Constructs TokenizerLike.
Parameters
args(term())kwargs(term())
@spec num_special_tokens_to_add( SnakeBridge.Ref.t(), keyword() ) :: {:ok, integer()} | {:error, Snakepit.Error.t()}
Python method TokenizerLike.num_special_tokens_to_add.
Returns
integer()
@spec pad_token_id(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec truncation_side(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}
@spec vocab_size(SnakeBridge.Ref.t()) :: {:ok, term()} | {:error, Snakepit.Error.t()}