trinity.common.models package
Submodules
- trinity.common.models.model module
InferenceModel
ModelWrapper
ModelWrapper.__init__()
ModelWrapper.generate()
ModelWrapper.generate_async()
ModelWrapper.chat()
ModelWrapper.chat_async()
ModelWrapper.logprobs()
ModelWrapper.logprobs_async()
ModelWrapper.convert_messages_to_experience()
ModelWrapper.convert_messages_to_experience_async()
ModelWrapper.model_version
ModelWrapper.get_openai_client()
ModelWrapper.extract_experience_from_history()
convert_api_output_to_experience()
extract_logprobs()
- trinity.common.models.openai_api module
- trinity.common.models.utils module
- trinity.common.models.vllm_async_model module
- trinity.common.models.vllm_model module
vLLMRolloutModel
vLLMRolloutModel.__init__()
vLLMRolloutModel.chat()
vLLMRolloutModel.generate()
vLLMRolloutModel.logprobs()
vLLMRolloutModel.convert_messages_to_experience()
vLLMRolloutModel.shutdown()
vLLMRolloutModel.sync_model()
vLLMRolloutModel.init_process_group()
vLLMRolloutModel.run_api_server()
vLLMRolloutModel.has_api_server()
vLLMRolloutModel.api_server_ready()
vLLMRolloutModel.reset_prefix_cache()
vLLMRolloutModel.get_model_version()
vLLMRolloutModel.sleep()
vLLMRolloutModel.wake_up()
- trinity.common.models.vllm_worker module
Module contents
- trinity.common.models.create_inference_models(config: Config) Tuple[List[InferenceModel], List[List[InferenceModel]]] [source]
Create engine_num rollout models.
Each model has tensor_parallel_size workers.