LogoLogo
  • Welcome to Release
  • Getting started
    • Quickstart
    • Create an account
    • Prepare to use Release
    • Create an application
      • Create custom application
      • Create from template
      • Servers vs runnables
    • Create an environment
  • Guides and examples
    • Domains and DNS
      • Manage domains
      • DNS and nameservers
        • Configure GoDaddy
        • Configure Cloudflare
        • Configure Namecheap
        • Other DNS hosts
      • Routing traffic
    • Example applications
      • Full stack voting app
      • Flask and RDS counter app
      • Static site with Gatsby
      • Golang with Postgres and Nginx
      • WordPress with MySQL
      • Spring and PostgreSQL
      • Terraform and Flask
      • OpenTelemetry demo
      • Load balancer with hostname
      • Static JavaScript service
      • SSH bastion access to services
      • ngrok and OAuth for private tunnels
      • Using OAuth Proxy
      • Hybrid Docker and static site
      • App Imports: Connecting two applications
      • Example library
    • Running instances
      • Cron jobs
      • Jobs
      • Using Helm charts
      • Using terminal
      • Viewing logs
      • Troubleshooting
        • ImagePullBackoff error
        • CrashLoopBackoff error
        • Exit codes
        • OOM: out of memory
    • Advanced guides
      • Containers guide
      • Application guide
      • Kubernetes guide
      • Create a cluster
      • Upgrade a cluster
      • Managing node groups
      • Patch node groups
      • Hostnames and rules
      • Serve traffic on multiple ports
      • Configure access to your K8s cluster
      • Designing for multiple environments
      • Microservices architecture
      • Monitoring your clusters
      • Performance tuning
      • Visibility and monitoring
      • Working with data
        • Container-based data
        • Seeding and migration
        • Cloud-provided data
        • Golden images
        • Third party
      • Pausing Instant Datasets
        • Application pausing schedules
        • Pause/resume environments
      • Infrastructure as code
        • Terraform
  • Reference documentation
    • Account settings
      • Account info
      • Managing users
      • Build settings
        • Build arguments
        • Build SSH keys
      • Add integrations
      • View clusters and cloud integrations
      • Add datasets
      • Environment handles
    • Workflows in Release
      • Stages of workflows
      • Serial deployments
      • Parallel deployments
      • Rolling deployments
      • Rainbow deployments
    • Networking
      • Network architecture (AWS)
      • Network architecture (GCP)
      • Ingresses
      • IP addresses
      • Cloud-provided services
      • Third-party services
    • Release environment versioning
    • Application settings
      • Application Template
        • Schema definition
      • Default environment variables
      • GitHub
      • Pull requests
      • GitOps
      • Just-in-time file mounts
      • Primary App Link
      • Create application FAQ
      • App-level build arguments
      • Parameters
      • Workspaces
    • End-to-end testing
    • Environment settings
      • Environment configuration
      • Environment variables
        • Environment variable mappings
        • Secrets vaults
        • Using Secrets with GitOps
        • Kubernetes Secrets as environment variables
        • Managing legacy Release Secrets
    • Environment expiration
    • Environment presets
    • Instant datasets on AWS
    • Instant datasets on GCP
    • Instant dataset tasks
      • Tonic Cloud
      • Tonic On-Premise
    • Cloud resources
    • Static service deployment
    • Helm
      • Getting started
      • Version-controlled Helm charts
      • Open-source charts
      • Building Docker images
      • Ingress and networking
      • Configuration
    • GitOps
    • The .release.yaml file
    • Docker Compose conversion support
    • Reference examples
      • Adding and removing services
      • Managing service resources
      • Adding database containers to the Application Template
      • Stock Off-The-Shelf Examples
    • Release API
      • Account Authentication
      • Environments API
        • Create
        • Get
        • Setup
        • Patch
      • User Authentication
      • Environment Presets API
        • Get Environment Preset List
        • Get Environment Preset
        • Put Environment Preset
  • Background concepts
    • How Release works
  • Frequently asked questions
    • Release FAQ
    • AWS FAQ
    • Docker FAQ
    • JavaScript FAQ
  • Integrations
    • Integrations overview
      • Artifactory integration
      • Cloud integrations (AWS)
        • AWS guides
        • Grant access to AWS resources
        • AWS how to increase EIP quota
        • Control your EKS fleet with systems manager
        • Managing STS access
        • AWS Permissions Boundaries
        • Private ECR Repositories
        • Using an Existing AWS VPC
        • Using an Existing EKS Cluster
      • Docker Hub integration
      • LaunchDarkly integration
      • Private registries
      • Slack integration
      • Cloud integrations (GCP)
        • GCP Permissions Boundary
      • Datadog Agent
      • Doppler Secrets Manager
      • AWS Secrets Management
    • Source control integrations
      • GitHub
        • Pull request comments
        • Pull request labels
        • GitHub deployments
        • GitHub statuses
        • Remove GitHub integration
      • Bitbucket
      • GitLab
    • Monitoring and logging add-ons
      • Datadog
      • New Relic
      • ELK (Elasticsearch, Logstash, and Kibana)
  • Release Delivery
    • Create new customer integration
    • Delivery guide
    • Release to customer account access controls
    • Delivery FAQs
  • Release Instant Datasets
    • Introduction
    • Quickstart
    • Security
      • AWS Instant Dataset security
    • FAQ
    • API
  • CLI
    • Getting started
    • Installation
    • Configuration
    • CLI usage example
    • Remote development environments
    • Command reference
      • release accounts
        • release accounts list
        • release accounts select
      • release ai
        • release ai chat
        • release ai config-delete
        • release ai config-init
        • release ai config-select
        • release ai config-upsert
      • release apps
        • release apps list
        • release apps select
      • release auth
        • release auth login
        • release auth logout
      • release builds
        • release builds create
      • release clusters
        • release clusters exec
        • release clusters kubeconfig
        • release clusters shell
      • release datasets
        • release datasets list
        • release datasets refresh
      • release deploys
        • release deploys create
        • release deploys list
      • release development
        • release development logs
        • release development start
      • release environments
        • release environments config-get
        • release environments config-set
        • release environments create
        • release environments delete
        • release environments get
        • release environments list
        • release environments vars-get
      • release gitops
        • release gitops init
        • release gitops validate
      • release instances
        • release instances exec
        • release instances logs
        • release instances terminal
  • Release.ai
    • Release.ai Introduction
    • Getting Started
    • Release.ai Templates
    • Template Configuration Basics
    • Using GPU Resources
    • Custom Workflows
    • Fine Tuning LlamaX
    • Serving Inference
Powered by GitBook
On this page
  • Node sizing
  • Autoscaling
  • Node autoscaling
  • Pod autoscaling
  • Resource management

Was this helpful?

  1. Guides and examples
  2. Advanced guides

Kubernetes guide

PreviousApplication guideNextCreate a cluster

Last updated 2 years ago

Was this helpful?

Originally designed by Google, Kubernetes is an open-source system for automating deployment, scaling, and management of containerized applications.

While Release dramatically simplifies operating Kubernetes for our users, it is important to have at least a basic understanding of some of its key concepts and terms as they relate to Release.

  • Pod: One or more containers that run with shared storage and networking.

  • Node: The cloud computing instances that run your Pods (for example, an AWS EC2 instance).

  • Cluster: Manages Pods and related resources across a set of Nodes.

  • Ingress: Routes HTTP and HTTPS traffic from outside the cluster to services within it.

Release manages the creation and administration of Kubernetes clusters within your cloud account. Configuration of your cluster, such as Node instance types, autoscaling of Nodes, Kubernetes version upgrades, and deployment of resources into your cluster, are all managed within the Release platform.

You can learn more about Kubernetes from .

Node sizing

To ensure best performance and to keep your cloud costs in check, it’s important to size the Nodes within your Kubernetes cluster appropriately for the applications they’ll be running. This involves selecting a cloud instance type that fits the needs of your workloads.

Some factors to consider include:

  • How much memory does your application need, under your expected usage?

  • Does your application tend to be more memory- or CPU-bound?

  • Does your application need access to high-performance local storage?

  • How sensitive are the environments within this cluster to variance in performance?

  • What are the minimum and maximum number of environments that will be running at a time?

Overlaying these considerations onto your budget will help you choose an appropriate instance type for your Nodes. This is also something that can and should be tuned as your needs or usage patterns change.

Autoscaling

At a high level, there are two major categories of autoscaling to consider within a Kubernetes cluster:

  • Node autoscaling

  • Pod autoscaling

Node autoscaling

Release manages Node autoscaling on your behalf. If the Kubernetes cluster does not have sufficient resources available to run a workload, then new Nodes will be automatically provisioned and added to the cluster, up to a set maximum. Likewise, Nodes will be automatically terminated and removed from the cluster, down to a set minimum, as environments are removed.

You can set a comfortable minimum and maximum number of Nodes and adjust it over time.

Pod autoscaling

Pod autoscaling is a little more complicated, because of the many application-specific factors that need to be taken into account. To give a few examples:

  • How can we tell that the system is in a state that it needs more or fewer Pods?

  • Which types of Pods do we need more or fewer of to respond to the changing load?

  • How can we avoid overwhelming limited resources (for example, DB connections)?

Release can support most popular Pod autoscaling solutions that exist in the Kubernetes ecosystem, but they are configured at the Kubernetes layer and require additional effort and expertise to effectively integrate them into your application. If you’d like more information on Pod autoscaling, contact us.

Typically, Pod autoscaling is something that most end users will not worry about or even consider until they deploy to production environments. Some customers may also want to configure and test a production-like environment for performance testing and tuning. These are the most common use cases for Pod autoscaling.

Resource management

Kubernetes provides capabilities to ensure that, when an application environment is created, its services are run on Nodes that have sufficient memory and CPU available. Kubernetes generally refers to this type of configuration as a resource request. It is also possible to monitor ongoing resource usage and restart Pods if they exceed the configured limit. Requests are can be thought of as a minimum set of guaranteed resources, while limits are can be thought of as the maximum advisable values.

For more information on Node sizing, see the . may also be helpful, particularly with regards to pricing.

Tuning resource requests and limits is key to efficiently managing your cluster’s resources and maintaining the stability and performance of your applications. Since resource configuration is heavily dependent on your application’s unique needs, this is something that should be carefully considered and adjusted over time. For more information on service resource tuning, refer to our documentation on .

their documentation
AWS EKS documentation
This third-party EC2 instance comparison tool
managing service resources