Instill AI changelog

Serverless Model Serving: Optimize AI deployment

changelog cover

The Serverless Model Serving feature on ⚗️ Instill Model offers a more cost-efficient solution for AI deployment by eliminating the need to manage the underlying infrastructure. ⚗️ Instill Model handles the compute resource provisioning, auto-scaling and availability, allowing you, the developers to focus on model development and deployment 🎉.

Key Benefits:

  • Scalability: Automated resource allocation scales to handle varying workloads efficiently.

  • Cost-Efficiency: Compute resources are provisioned on-demand and scaled down to zero when idling.

  • Simplified Management: No need for server management, free up your teams to focus on improving models and applications.

  • Quicker Deployment: Seamless integration with 🔮 Instill Core for faster full-stack AI development with data pipeline and knowledge bases.

We've also made significant performance enhancements:

  • Faster image push times

  • Reduced model cold-start times

We’re continuously enhancing ⚗️ Instill Model and will soon introduce Dedicated Model Serving for production-purpose model serving service.

☁️ Instill Cloud users can host public models for free and we offer monthly free Instill Credits for ☁️ Instill Cloud users to run any public models hosted on ☁️ Instill Cloud.

More at: Instill AI