Sesterce Cloud Doc
  • 👋Welcome on Sesterce Cloud
    • 🚀Get Started!
      • Account creation
      • Manage your account
      • Payment & Billing
        • Invoicing
  • 🚀Compute instances
    • Compute Instance configuration
      • Persistent storage (volumes)
      • SSH Keys
    • Terminal connection
  • 💬AI Inference instances
    • Inference Instance configuration
      • Select your Flavor
      • Select your regions
      • Autoscaling limits
    • Edit an inference instance
    • Chat with Endpoint
  • ▶️Manage your instances
  • 🔗API Reference
    • Authentication
    • GPU Cloud instances
    • SSH Keys
    • Volumes
    • Inference Instances
  • 📗Tutorials
    • Expose AI model from Hugging Face using vLLM
Powered by GitBook
On this page
  • How to edit an inference instance?
  • What can I edit from a running inference instance?

Was this helpful?

  1. AI Inference instances

Edit an inference instance

PreviousAutoscaling limitsNextChat with Endpoint

Last updated 2 months ago

Was this helpful?

How to edit an inference instance?

Even after launching, it is possible to update the settings of your inference instances. This is particularly relevant if you need to edit your autoscalling limits, according to the usage of your endpoint.

From Inference, select "Edit". Then, fill up the form with the new parameters to want to pass into your instance.

What can I edit from a running inference instance?

Here are the element that can be customized from an Running inference instance:

  • GPU/CPU Flavor

  • Region

  • Startup Command

  • Containers and autoscaling triggers

  • Environment variables

💬