ReaderConfig#
- class langchain_community.llms.titan_takeoff.ReaderConfig[source]#
Bases:
BaseModel
Configuration for the reader to be deployed in Titan Takeoff API.
Create a new model by parsing and validating input data from keyword arguments.
Raises ValidationError if the input data cannot be parsed to form a valid model.
- param consumer_group: str = 'primary'#
The consumer group to place the reader into
- param max_batch_size: int = 4#
The max batch size for continuous batching of requests
- param max_seq_length: int = 512#
The maximum sequence length to use for inference, defaults to 512
- param model_name: str [Required]#
The name of the model to use
- param tensor_parallel: int | None = None#
The number of gpus you would like your model to be split across