Inference Server API

class mlops_client.inference_server.inferenceserver_client.InferenceServer(headers: Dict[str, str], host: str, client: Optional[ClientSession] = None)
async create_rest(model_path: str, num_instances: int, sources: List[FeatureSource]) Any

Creates a REST inference server for a given model.

Args:
  • model_path: Path to the model file.

  • num_instances: Number of instances to create.

  • sources: A list of FeatureSource objects.

Returns:
  • A list of InferenceServer objects.

async list(offset: Optional[str] = None, limit: Optional[int] = None) Any

List all inference servers.

Parameters:
  • offset: An optional string that represents the starting item, should be the value of ‘next’ field from the previous response.

  • limit: An optional integer to limit the number of returned items.

Returns:
  • A list of InferenceServer objects.