Inference Server API
- class mlops_client.inference_server.inferenceserver_client.InferenceServer(headers: Dict[str, str], host: str, client: Optional[ClientSession] = None)
- async create_rest(model_path: str, num_instances: int, sources: List[FeatureSource]) Any
Creates a REST inference server for a given model.
- Args:
model_path: Path to the model file.
num_instances: Number of instances to create.
sources: A list of FeatureSource objects.
- Returns:
A list of InferenceServer objects.
- async list(offset: Optional[str] = None, limit: Optional[int] = None) Any
List all inference servers.
- Parameters:
offset: An optional string that represents the starting item, should be the value of ‘next’ field from the previous response.
limit: An optional integer to limit the number of returned items.
- Returns:
A list of InferenceServer objects.