Instill AI changelog
Serverless Model Serving: Optimize AI deployment
The Serverless Model Serving feature on ⚗️ Instill Model offers a more cost-efficient solution for AI deployment by eliminating the need to manage the underlying infrastructure. ⚗️ Instill Model handles the compute resource provisioning, auto-scaling and availability, allowing you, the developers to focus on model development and deployment 🎉.
Key Benefits:
Scalability: Automated resource allocation scales to handle varying workloads efficiently.
Cost-Efficiency: Compute resources are provisioned on-demand and scaled down to zero when idling.
Simplified Management: No need for server management, free up your teams to focus on improving models and applications.
Quicker Deployment: Seamless integration with 🔮 Instill Core for faster full-stack AI development with data pipeline and knowledge bases.
We've also made significant performance enhancements:
Faster image push times
Reduced model cold-start times
We’re continuously enhancing ⚗️ Instill Model and will soon introduce Dedicated Model Serving for production-purpose model serving service.
☁️ Instill Cloud users can host public models for free and we offer monthly free Instill Credits for ☁️ Instill Cloud users to run any public models hosted on ☁️ Instill Cloud.
More at: Instill AI