# -*- coding: utf-8 -*-
# Generated by the protocol buffer compiler.  DO NOT EDIT!
# source: mediapipe/tasks/cc/genai/inference/proto/llm_params.proto
# Protobuf Python Version: 4.25.1
"""Generated protocol buffer code."""
from google.protobuf import descriptor as _descriptor
from google.protobuf import descriptor_pool as _descriptor_pool
from google.protobuf import symbol_database as _symbol_database
from google.protobuf.internal import builder as _builder
# @@protoc_insertion_point(imports)

_sym_db = _symbol_database.Default()


from mediapipe.tasks.cc.genai.inference.proto import transformer_params_pb2 as mediapipe_dot_tasks_dot_cc_dot_genai_dot_inference_dot_proto_dot_transformer__params__pb2


DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n9mediapipe/tasks/cc/genai/inference/proto/llm_params.proto\x12\x10odml.infra.proto\x1a\x41mediapipe/tasks/cc/genai/inference/proto/transformer_params.proto\"\xa1\x03\n\rLlmParameters\x12G\n\x16transformer_parameters\x18\x01 \x01(\x0b\x32\'.odml.infra.proto.TransformerParameters\x12\x12\n\nvocab_size\x18\x02 \x01(\x05\x12\x18\n\x0estart_token_id\x18\x04 \x01(\x05H\x00\x12\x15\n\x0bstart_token\x18\x06 \x01(\tH\x00\x12\x13\n\x0bstop_tokens\x18\x05 \x03(\t\x12]\n\x1binput_output_normalizations\x18\x07 \x03(\x0e\x32\x38.odml.infra.proto.LlmParameters.InputOutputNormalization\"s\n\x18InputOutputNormalization\x12&\n\"INPUT_OUTPUT_NORMALIZATION_UNKNOWN\x10\x00\x12/\n+INPUT_OUTPUT_NORMALIZATION_BYTES_TO_UNICODE\x10\x01\x42\x13\n\x11start_token_unionJ\x04\x08\x03\x10\x04*\x8a\x02\n\x0cLlmModelType\x12\x1a\n\x16LLM_MODEL_TYPE_UNKNOWN\x10\x00\x12\x1f\n\x1bLLM_MODEL_TYPE_FALCON_RW_1B\x10\x05\x12\x1b\n\x17LLM_MODEL_TYPE_GEMMA_2B\x10\x06\x12\x1b\n\x17LLM_MODEL_TYPE_GEMMA_7B\x10\x0c\x12#\n\x1fLLM_MODEL_TYPE_STABLELM_4E1T_3B\x10\x08\x12\x18\n\x14LLM_MODEL_TYPE_PHI_2\x10\x0b\x12\x1a\n\x16LLM_MODEL_TYPE_TF_LITE\x10\x64\"\x04\x08\x01\x10\x01\"\x04\x08\x02\x10\x02\"\x04\x08\x03\x10\x03\"\x04\x08\x04\x10\x04\"\x04\x08\x07\x10\x07\"\x04\x08\t\x10\t\"\x04\x08\n\x10\nB1\n\x1b\x63om.google.odml.infra.protoB\x12LLMParametersProtob\x06proto3')

_globals = globals()
_builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
_builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'mediapipe.tasks.cc.genai.inference.proto.llm_params_pb2', _globals)
if _descriptor._USE_C_DESCRIPTORS == False:
  _globals['DESCRIPTOR']._options = None
  _globals['DESCRIPTOR']._serialized_options = b'\n\033com.google.odml.infra.protoB\022LLMParametersProto'
  _globals['_LLMMODELTYPE']._serialized_start=567
  _globals['_LLMMODELTYPE']._serialized_end=833
  _globals['_LLMPARAMETERS']._serialized_start=147
  _globals['_LLMPARAMETERS']._serialized_end=564
  _globals['_LLMPARAMETERS_INPUTOUTPUTNORMALIZATION']._serialized_start=422
  _globals['_LLMPARAMETERS_INPUTOUTPUTNORMALIZATION']._serialized_end=537
# @@protoc_insertion_point(module_scope)
