Skip to content

cortexlabs/cortex

Repository files navigation


Run inference at scale

Cortex is an open source platform for large-scale inference workloads.


Model serving infrastructure

  • Supports deploying TensorFlow, PyTorch, sklearn and other models as realtime or batch APIs.
  • Ensures high availability with availability zones and automated instance restarts.
  • Runs inference on on-demand instances or spot instances with on-demand backups.
  • Autoscales to handle production workloads with support for overprovisioning.

Configure a cluster

# cluster.yaml region: us-east-1 instance_type: g4dn.xlarge min_instances: 10 max_instances: 100 spot: true

Spin up on your AWS or GCP account

$ cortex cluster up --config cluster.yaml ○ configuring autoscaling ✓ ○ configuring networking ✓ ○ configuring logging ✓ cortex is ready! 

Reproducible deployments

  • Package dependencies, code, and configuration for reproducible deployments.
  • Configure compute, autoscaling, and networking for each API.
  • Integrate with your data science platform or CI/CD system.
  • Deploy custom Docker images or use the pre-built defaults.

Define an API

class PythonPredictor: def __init__(self, config): from transformers import pipeline self.model = pipeline(task="text-generation") def predict(self, payload): return self.model(payload["text"])[0] requirements = ["tensorflow", "transformers"]

Configure an API

api_spec = { "name": "text-generator", "kind": "RealtimeAPI", "compute": { "gpu": 1, "mem": "8Gi" }, "autoscaling": { "min_replicas": 1, "max_replicas": 10 } }

Scalable machine learning APIs

  • Scale to handle production workloads with request-based autoscaling.
  • Stream performance metrics and logs to any monitoring tool.
  • Serve many models efficiently with multi-model caching.
  • Use rolling updates to update APIs without downtime.
  • Configure traffic splitting for A/B testing.

Deploy to your cluster

import cortex cx = cortex.client("aws") cx.create_api(api_spec, predictor=PythonPredictor, requirements=requirements) # creating https://example.com/text-generator

Consume your API

$ curl https://example.com/text-generator -X POST -H "Content-Type: application/json" -d '{"text": "hello world"}'

Get started

About

Production infrastructure for machine learning at scale

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Contributors 22