Instill AI changelog
Run-on-Event: Send Slack message on GitHub star event
Introducing the new Run-on-Event feature on 💧 Instill VDP to enhance pipeline automation by listening to events from other services so that VDP can provide corresponding actions.
Features
A star created event from GitHub
A new message event from Slack
Run-on-Event enables you to:
Trigger downstream processes by listening to events
Automate notifications & alerts
Enable real-time data synchronization across different services
Send a message to Slack on a GitHub star event
Pre-requisites:
When someone stars your GitHub repo, automatically send a message to your designated Slack channel.
Customize your event messages
See the example recipe below:
# VDP Version
version: v1beta
on:
github:
type: github
event: EVENT_STAR_CREATED
config:
repository: <USERNAME/YOUR-GITHUB-REPO> # i.e. instill-ai/instill-core
setup: ${connection.<YOUR-GITHUB-CONNECTION>}
variable:
repository:
title: repository
format: string
listen:
- ${on.github.message.repository.full-name}
user:
title: user
format: string
listen:
- ${on.github.message.sender.login}
description:
title: user
format: string
listen:
- ${on.github.message.repository.description}
stargazers-count:
title: stargazers-count
format: number
listen:
- ${on.github.message.repository.stargazers-count}
output:
text:
value: ${variable.repository} ${variable.user}
component:
slack:
type: slack
input:
channel-name: <YOUR-SLACK-CHANNEL-NAME> # i.e. notify-github-star
message: |
${variable.user} star ${variable.repository}
Repo Description: ${variable.description}
Stargazers Count: ${variable.stargazers-count}
as-user: false
condition:
setup: ${connection.<YOUR-SLACK-CONNECTION>}
task: TASK_WRITE_MESSAGE
The GitHub star and Slack message examples demonstrate how Run-On-Event can bridge different services, allowing VDP to create responsive, event-driven workflows that enhance overall system automation and integration.
Learn more about this feature by joining our Discord where we showcase these new features.
Release 48: Improvement & fixes
Other
Fix bug for cloning features
To file an issue please use this template or raise them in our Discord community.
One-click integrations with GitHub, Slack & Google Drive
OAuth is an open-standard authorization protocol or framework that provides applications the ability for “secure designated access.
Today, we upgraded the GitHub, Slack and Google Drive Components to use OAuth for a quicker and simpler authentication process to connect to your external apps.
To get started:
Sign in to the Console
Navigate to your top right Profile > Setting > Integrations
Search for GitHub, Slack or Google Drive and click connect
Key Benefits:
Secure Connections: Access your resources safely without sharing passwords.
Easy Access Management: Quickly control who can access connected services.
Automatic Setup: No manual IDs needed—connections are set up automatically.
Seamless Integration: Connect to top platforms in just a click, simplifying your workflow.
We will be adding more integrations very soon so stay tuned!
Learn more about this feature by joining our Discord where we showcase these new features.
Release 47: Improvement & fixes
Pipeline Builder
[Improvement] Add auto-save for pipeline when "Run" button clicked
Python SDK
[Improvement] Remove attribute naming
_service
[Improvement] Support specifying target namespace in endpoint functions
[Improvement] Align initialization functions between each service client
[Improvement] Remove configuration system
Other
[Bug] Fix iterator component dependency has no connection in preview canvas
[Bug] Fix error encountered when model running at "Scaling" stage after 15 minutes
[Improvement] Add
*/*
format type togenerateInputsPayload
[Improvement] Add Instill Model scaling hint for the VDP Component
[Improvement] Update the landing page after onboarding to the "Explore" page.
[Improvement] Update inconsistent logo sizes in the "About" Page
To file an issue please use this template or raise them in our Discord community.
Visit Your Dashboard
The Dashboard is a centralized view for observing all your pipeline and model run activity and Instill Credit consumption costs. This view includes a chart for data visualization and a table for your run history logs.
This means you can easily review and calculate any future costs associated with your pipeline and model usage. Simply log into your Instill Cloud account and go to the Dashboard tab to check it out!
Learn more about this feature by joining our Discord where we showcase these new features.
Ready-to-use AI Assistants on Instill Cloud
Today, we are releasing Instill App, where you can create ready-to-use AI Assistants that seamlessly connect with your knowledge bases stored in Instill Artifact. This chatbot interface enables you to easily interact with any documents you have uploaded, fast-tracking the infrastructure for achieving RAG applications.
To get started:
Create a Catalog in Artifact
Upload your files
Create App using either:
export INSTILL_API_TOKEN=******** curl -X POST 'https://api.instill.tech/v1alpha/namespaces/NAMESPACE_ID/apps' \ --header "Content-Type: application/json" \ --header "Authorization: Bearer $INSTILL_API_TOKEN" \ --data '{ "id": "customer-chatbot", "description": "Answers customer enquiries about Instill AI.", "tags": ["AI", "Assistant"] }'
Go to Applications and chat with your AI Assistant
This feature is exclusively available on ☁️ Instill Cloud and Instill Enterprise.
Learn more about this feature by joining our Discord where we showcase these new features.
Release 46: Improvement & fixes
Pipeline Builder
[Improvement] Implement Auto-Save for Pipeline When "Run" Button Clicked
Python SDK
[Improvement] Remove Attribute Naming
_service
[Improvement] Support specifying target namespace in endpoint functions
[Improvement] Align initialization functions between each service client
[Improvement] Remove configuration system
Other
[Bug] Fix iterator component dependency has no connection in preview canvas
[Bug] Fix Error Encountered When Model Running at "Scaling" Stage After 15 Minutes
[Improvement] Add
*/*
format type togenerateInputsPayload
[Improvement] Instill Model Scaling Hint for the VDP Component
[Improvement] Update the landing page after onboarding to the "Explore" page.
[Improvement] Inconsistent Logo Sizes in the "About" Page
To file an issue please use this template or raise them in our Discord community.
Instill App Component
We are introducing Instill App Component on💧 Instill VDP to easily access your conversation history from your AI Assistants created from Instill App.
Features
Support tasks:
Read Chat History
Write Message
To request a new component, please go here and click "Request A Component".
Release 45: Improvement & fixes
Unified Data Processing
[Bug] Fix JSON operator jq filtering
[Bug] Fix ask-the-catalog Pipeline Output Display Chunks as 'object'
Pipeline Builder
[Feat.] Add per component output in editor for debugging
[Bug] Fix JSON output missing in Playground and pipeline builder for Crawler
[Bug] Fix no hint when renaming does not work
[Bug] Fix only Partial Credit-Related Errors Msg Displayed in the Entire Pipeline
[Bug] Fix logic to display correct UI for JSON output
[Improvement] Format UX for full visibility of Ask task when scrolling
[Improvement] Add smart hint before user types any character (`CTL+space`)
Other
[Bug] Fix button not functioning in pipeline preview tab
[Bug] Fix CMDO will be broken after adding one random component
[Bug] Fix org name and logo disappear after creating an organization
[Bug] Use the "Response title" instead of the "Response key" in the Pipeline Playground Output view
[Bug] Fix dialogue windows jitter after clicking the Clone button
[Improvement] Replace UI for not found Public Pipelines & Models to the Playground Tab
To file an issue please use this template or raise them in our Discord community.
Instill Integrations
We are improving vendor connection resources by introducing Instill Integrations to easily setup and contain the connection details of any 3rd party vendors (i.e. OpenAI API key, Pinecone URL and API key, etc.) on Instill VDP. This will ensure higher observability by ensuring you see what vendor has been authorized for each integration and improve overall user experience when applied in the pipeline builder.
In contrast to Secret Management, this new integration feature will separate the component setup from the variable which means:
Secrets will be used for avoiding repetition and will not be tied to any component type
Integrations will be used to store connection setup details all in one place
To setup an integration, you can either do it through the Integrations page or within the pipeline builder.
Configuring integrations
The Integrations page will allow you to easily configure and protect your external connections all in one place, allowing you to quickly search & filter available integrations and configure new ones.
To setup your personal new integration, simply follow these steps:
Go to your Profile > Settings > Integrations
Click
Connect
on any available 3rd party vendorFill in the connection details
To setup your organizations new integration, simply follow these steps:
Go to your Profile > Your Organization > Organization > Integrations
Click
Connect
on any available 3rd party vendorFill in the connection details
Using integrations
When writing a pipeline recipe in the pipeline builder, you can reference the connection in the setup
section within your component declaration:
component:
library-db:
type: sql
input:
table-name: books
update-data: ${variable.books}
setup: ${connection.my-library-1} # <-- reference connection
task: TASK_UPDATE
Next, we will also enable connecting with 3rd party services via OAuth.
Learn more about this feature by joining our Discord where we showcase these new features.
Release 44: Improvement & fixes
Instill Artifact
[Improvement] Rename all APIs to follow a new structure
Pipeline Builder
[Bug] Enable Asana being added as a component correctly
[Bug] Fix component spinning after pipeline successfully runs
[Bug] Correct component indentation display when adding a component
[Improvement] Center code in the left editor panel when clicking a preview component
[Improvement] Make the left sidebar in the Low-Code Editor adjustable
[Improvement] Enhance the "Add Component" experience by removing null entries
[Improvement] Round component borders and increase text size
[Improvement] Enable description hints for each key
Python SDK
[Bug] Fix issue with triggering the organization pipeline using
InstillClient
andtrigger_org_pipeline_release()
Other
[Bug] Add missing Public/Private labels in some scenarios
[Bug] Fix bugs when converting JPG to text
[Bug] Eliminate white padding issue in the image preview widget
[Improvement] Remove the version "test" button
To file an issue please use this template or raise them in our Discord community.
Instill SDK for Python Devs
The Instill SDK provides a Python-based interface for interacting with Instill Core features, including data, pipeline and model operations.
Please note: Our Python SDK supports up to and including Python 3.11.
Data - Instill Artifact
Perform key data operations such as:
List, create, update, and delete catalogs
Upload, process, and manage files within a catalog
View and search file chunks
Ask catalog to retrieve relevant information based on files
Pipeline - Instill VDP
Handle pipeline tasks:
Create and run pipelines
Model - Instill Model
Work with AI models:
Create and serve models
Run model for inference
Check out Instill AI Cookbooks to learn how to use the Python SDK:
The Python SDK is undergoing continuous improvements. We'd love to hear your feedback! Please share your thoughts in the #feedback channel on our Discord community.
Asana Component
Asana is a web and mobile "work management" platform designed to help teams organize, track, and manage their work
We added an Asana Component on💧 Instill VDP to empower project management workflow automation. Build unique pipeline automation that you can mix and match across various vendors using Instill Components like data, application and AI.
Features
Support tasks:
Goal
Task
Project
Portfolio
To request a new component, please go here and click "Request A Component".
Freshdesk Component
Freshdesk is a cloud-based customer support platform that was founded with the mission of enabling companies of all sizes to provide great customer service. Our goal is simple: make it easy for brands to talk to their customers and make it easy for users to get in touch with businesses.
We added a Freshdesk Component on💧 Instill VDP to enable more Customer Relationship Management (CRM) capabilities when building your pipelines. This application component focuses on using Freshdesk's ticketing system.
Features
Support tasks:
Get Ticket
Create Ticket
Reply To Ticket
Create Ticket Note
Get All Conversations
Get Contact
Create Contact
Get Company
Create Company
Get All
Get Product
Get Group
Get Agent
Get Skill
Get Role
To request a new component, please go here and click "Request A Component".
Chroma Component
Chroma is an AI-native open-source embedding database.
We added a Chroma Component on💧 Instill VDP to expand your vector database options and enable you to build and search Chroma's' vector databases.
Features
Support tasks:
Batch Upsert
Upsert
Query
Delete
Create Collection
Delete Collection
To request a new component, please go here and click "Request A Component".
Release 43: Improvement & fixes
VDP components
[Improvement] Instill Artifact Component Tasks I/O Design
[Improvement] Add create issue to Jira Component
Pipeline builder
[Bug] Auto-save Synchronization Between Frontend and Backend May Cause Users to Revert to a Previous State
[Bug] Pipeline Overview Version Dropdown Fails to Display All Versions
[Bug] Streaming is not Working on Safari Browser
[Bug] Broken Documentation Link for Component Instill Model
[Bug] Output Image Cannot display on the Console
[Improvement] Reduce the font size of code blocks in the Pipeline Editor
[Improvement] User Can Use
Control+S
orCmd+S
to Manual-Save Recipe[Improvement] Add the Preview from Pipeline Editor to the Overview Page
[Improvement] Pipeline Editor Output Default UI Need to Adjustment
[Improvement] User does not know how to make a variable multiline
To file an issue please use this template or raise them in our Discord community.
Revamping the Pipeline Builder
We are introducing a completely new way of building your pipelines on Instill VDP to shift our tool to focus on better supporting developers in the AI and data domain. The goal of VDP is to help you dynamically build versatile pipelines for complex and unique use cases that require minimal maintenance with high data observability.
Recipe as Code
The new pipeline builder UI allows you to edit YAML recipes next to the pipeline preview for improved debugging and data flow visualization. This means you can edit recipe.yml
files in your own editors like VSCode and simply import them into VDP, streamlining the recipe sharing process. Additionally, the editor offers smart hints and syntax highlighting, making it easier to build and modify recipes efficiently.
Watch this tutorial on how to build recipes in the new pipeline editor:
Quick tips:
Use
Control+O
orCmd+O
shortcut to add components easilyUse
Control+K
orCmd+K
shortcut to search for thingsUse
Control+S
orCmd+S
shortcut to save your recipe manuallyUse
Control+Enter
orCmd+Enter
shortcut to run the pipelineClick the
Getting Started
button at the bottom-left corner to view the guide in the pipeline preview panel.
High Observability
The pipeline preview offers a visual representation of data flow between components, striking a balance between code-level control and visual management. This feature is particularly beneficial for data engineers working on complex pipelines, as it provides real-time streaming feedback to simplify debugging. The preview also enhances component presentation, making it easier to build, test, and maintain sophisticated data pipelines efficiently.
You can try exploring the recipe of the structured-text-summarizer pipeline to transform a paragraph of unstructured text into structured JSON output in the new pipeline editor.
Learn more about this feature by joining our Discord where we showcase these new features.
Pipeline Streaming Mode
The new pipeline editor contains a pipeline preview with a streaming feature which is useful for you to observe how data flows through the components when running your pipeline. It will provide visibility into the status of each component in your pipeline run.
You can enable streaming mode by:
Create or use an existing pipeline such as structured-web-insights
Clone this pipeline by clicking
Clone
in the top rightGo into editor mode by clicking
Edit
in the top rightSetup your Google Search Engine (see setup docs or README)
Fill in the input form and hit
Run
Learn more about this feature by joining our Discord where we showcase these new features.
Runs for Logging Pipeline and Model Triggers
We are introducing Runs, a logging system for pipeline and model executions. This feature provides execution logs for the currently selected pipeline or model that will help you better track credit usage, monitor performance and get snapshots of each executions metadata which is useful for debugging.
To get started, simply follow these steps:
Pipeline Logging Runs
Go to any pipeline such as the contract-reviewer
In the Playground, run the pipeline
Go to "Runs" tab to view your history of pipeline executions
Click into any Run ID to view specific metadata related to this pipeline execution
Clicking into a specific Run ID will include useful information including run duration, credit consumption, trigger source, recipe, input/output results, and each components metadata.
Model Logging Runs
Go to any Instill Model such as the instill-ai/llava-1-6-13b.
In the Playground, run the model.
Go to "Runs" tab to view your history of model executions
Click into any Run ID to view specific metadata related to this model execution
More at: Instill AI
Release 42: Improvement & fixes
Improvements
Added create issue task to Jira Component
Added new fields in Website Crawl
Incorporated I/O tasks in Instill Artifact Component
Added Excel pivot table to Markdown conversion in Document Component
Implemented code snippet extraction in Web operator
Enhanced SQL component with SSL/TLS input (base64 encoded) and engine relocation
Expanded HubSpot Component with deal/ticket updates and task retrieval
Added XLSX to Text/Markdown conversion support for Catalog
Introduced PDF to Images conversion in Document component
Improved Artifact component with base64 upload, multiple file support, and filename output
Added file upload interruption notification
Bug Fixes
Fixed mobile device login/signup buttons
Corrected Public Model Playground Page button wording
Resolved text highlight omission issue
Fixed blank Floating Window UI bug
Addressed Catalog file deletion issue
To report a bug or improvement, please create an issue on GitHub here.
Zilliz Component
Powered by open-source Milvus, Zilliz delivers the most performant and cost-effective vector database for AI at any scale.
We added Zilliz Component on💧 Instill VDP to expand your vector database options and enable you to build and search Zilliz's vector database.
Features
Support tasks:
Vector Search
Upsert
Batch Upsert
Delete
Create Collection
Drop Collection
Create Partition
Drop Partition
To request a new component, please go here and click "Request A Component".
Milvus Component
Milvus is an open-source vector database built for GenAI applications. Install with pip, perform high-speed searches, and scale to tens of billions of vectors with minimal performance loss.
We added Milvus Component on💧 Instill VDP to expand your vector database options and enable you to build and search Milvus' vector databases.
Features
Support tasks:
Vector Search
Upsert
Delete
Drop Collection
Create Partition
Drop Partition
To request a new component, please go here and click "Request A Component".
Top up your Instill Credits
All Instill Cloud users automatically receive 10,000 Instill Credits for free each month. Instill Credits help reduce the need to pre-configure any AI components by skipping the need for creating accounts or API keys on 3rd party services.
If you run out, simply follow these steps to top up:
Personal Credits:
Click the top-right avatar
Go to Settings > Billing > Credits
Organization Credits:
Click the top-right avatar
Go to Settings > Organization
Select a specific organization
Go to Billing > Credits
Learn more about this feature by joining our Discord where we showcase these new features.
WhatsApp: Automate Sending Messages
We added WhatsApp Component on 💧 Instill VDP to help you automate workflows and send template messages on WhatsApp. Construct useful workflows that can automate your work and unstructured data with any AI Component and then send the result as a message to anyone on WhatsApp.
Features
Support tasks:
Send Text Based Template Message
Send Media Based Template Message
Send Location Based Template Message
Send Authentication Template Message
Send Text Message
Send Media Message
Send Location Message
Send Contact Message
Send Interactive Call To Action Url Button Message
To request a new component, please go here and click "Request A Component".
Release 41: Improvement & fixes
Improvements
Removed the bottom credits balance
Prevented deletion of catalogs under organizations
Enhanced web operator scrape quality for markdown
Bug Fixes
Resolved issue preventing users from scrolling in nodes within the pipeline-builder
Corrected dual indicator display on Console
Addressed organization credit consumption problem
Fixed frontend not rendering backend error messages correctly
Repaired broken Explore functionality in the Navigation Bar
Resolved issue with iterator over
pinecone-kb.output.matches
Array
More at: Instill AI
Elasticsearch: Scalable Data Store and Vector Database
Elasticsearch is a distributed, RESTful engine with powerful search, analytics, and a scalable vector database, ideal for full-text and semantic search, as well as personalized recommendations.
We added Elasticsearch Component on 💧 Instill VDP to enable powerful and flexible solution for indexing, searching, updating, and deleting data within your Elasticsearch databases.
Features
Support tasks:
Search
Vector Search
Index
Multi Index
Update
Delete
Create Index
Delete Index
To request a new component, please go here and click "Request A Component".
Qdrant: High-Performance Vector Search at Scale
Qdrant is a vector database and similarity search engine designed to handle high-dimensional vectors for performance and massive-scale AI applications.
We added Qdrant Component on💧 Instill VDP to integrate with the Qdrant vector search and storage solution and provide you with more flexibility and choice when it comes to vector database solutions.
Features
Support tasks:
Vector Search
Batch Upsert
Upsert
Delete
Create Collection
Delete Collection
More at: Instill AI
OpenAI: Structured Outputs
We added OpenAI's latest gpt-4o-2024-08-06 model to OpenAI Component on 💧 Instill VDP so you can build more reliable AI pipelines with Structured Outputs. This update enables you to define JSON schemas for model outputs, ensuring consistent, easily processable, and accurate responses that streamline workflows and enhance error handling. It is ideal for data analysts, developers, business intelligence teams, and customer service automation, particularly when working with large volumes of text data or building robust AI-powered applications that require reliable, structured information.
Features
OpenAI's latest model: gpt-4o-2024-08-06
New response format option: json_schema
☁️ Instill Cloud demo
More at: Instill AI
MongoDB: Versatile Data Movement for Document Databases
MongoDB is a NoSQL database that stores data in flexible, JSON-like documents, allowing for easy scalability and dynamic schema design, making it well-suited for handling large volumes of unstructured or semi-structured data in modern applications.
We added MongoDB Component on 💧 Instill VDP to enhance your data pipelines and connecting to your MongoDB databases.
Features
Support tasks:
Insert
Insert Many
Find
Update
Delete
Drop Collection
Drop Database
Create Search Index
Drop Search Index
Vector Search
More at: Instill AI
Weaviate: AI-Native Vector Database
Weaviate is a vector database combining versatile vector and hybrid search capabilities with scalable, real-time operations on multi-modal data.
We added Weaviate Component on 💧 Instill VDP to expand your vector database options and enhance capabilities for managing and querying vector data efficiently.
Features
Support tasks:
Vector search
Batch insert
Insert
Update
Delete
Delete Collection
To request a new component, please go here and click "Request A Component".
Fireworks AI: Fastest Inference for Generative AI
Fireworks AI is a generative AI inference platform to run and customize models with industry-leading speed and production readiness.
We added Fireworks AI Component on 💧 Instill VDP to unlock a wide range of natural language processing capabilities for your applications.
Features
Support the following tasks:
Text Generation Chat
Text Embeddings
Models:
llama3.1 (405B, 70B, 9B), llama3 (70B, 8B), gemma2-9b, phi-3-vision-128K, deepseek-coder, qwen2 (72B), nomic-ai/nomic-embed, thenlper/gte, WhereIsAI/UAE
☁️ Instill Cloud demo:
More at: Instill AI
Groq: Fast, Affordable, and Energy Efficient AI
Groq solutions are based on the Language Processing Unit (LPU), a new category of processor and LPUs run Large Language Models (LLMs) at substantially faster speeds and, on an architectural level, up to 10x better energy efficiency compared to GPUs.
We added Groq Component on 💧 Instill VDP to help you achieve fast AI inference at scale for your AI applications.
Features
Support text generation chat task with models: llama3.1, llama3, gemma2, gemma, and more
☁️ Instill Cloud demo
To request a new component, please go here and click "Request A Component".
Artifact Component
We recently launched 💾 Instill Artifact, the final piece that completes Instill Core as a full-stack AI infrastructure tool. Instill Artifact allows you to unify unstructured data from various sources on one platform and prepare it for AI and RAG tasks. Check out our blog for more details.
We've introduced the Artifact Component to 💧 Instill VDP, giving you the power to effortlessly build data pipelines that turn your files and documents into RAG-ready data. With this, you can easily retrieve relevant data from your catalog for AI chatbots, search, and recommendation systems. Plus, you can directly query your catalog and get instant, LLM-powered answers from your data.
Features
Support tasks:
Upload Files
Get Files Metadata
Get Chunks Metadata
Get File In Markdown
Match File Status
Search Chunks
Query
☁️ Instill Cloud demo
More at: Instill AI
Introducing Instill Artifact: Getting your data AI-ready
We're excited to introduce 💾 Instill Artifact, a new feature for orchestrating unstructured data like documents, images, audio, and video into Instill Catalog - a unified AI-ready Augmented Data Catalog format for unstructured data and AI applications.
💾 Instill Artifact can help with:
Data Preparation: Create high-quality AI-ready data catalogs from uploaded files.
Simplicity: Use low-latency API calls or Instill Console's one-click interface.
Transparency: Access and control data at catalog, file, or chunk levels.
Data Integrity: Ensure AI application reliability with markdown-based source-of-truth.
Scalability: Automate unstructured data transformation & growing data volumes.
Versatility: Process various unstructured, semi-structured, and structured data types and sources.
Full-Stack AI Integration: Seamlessly connect with 💧 Instill VDP and ⚗️ Instill Model.
This release allows you to:
Create your own Knowledge Base
Upload files in formats like PDF, Markdown, and TXT
View metadata for each Knowledge Base
Examine "Catalog" details for uploaded files
This feature completes 🔮 Instill Core and lays the groundwork for developing full-stack AI solutions by ensuring your data is unified, clean, and ready for insight extraction.
More at: Instill AI
Powerful CRM with HubSpot + AI Components
Adding the new HubSpot Component enables you to empower your CRM and marketing workflow use cases together with other AI Components using 💧 Instill VDP. This component will support the following tasks (see docs):
Get, Create Contact
Get, Create Deal
Get, Create Company
Get, Create Ticket
Get Thread
Insert Message
Retrieve Association
With these tasks, you will be able to create workflow automations with AI such as:
Get message thread from HubSpot, connect it to other AI Components and summarise into action items
Apply sentiment analysis to retrieved details of contacts
and more!
Try demo on ☁️ Instill Cloud:
More at: Instill AI
PM Workflow Automations with Jira + AI Components
As a Project Manager (PM), organizing tickets in your kanban board can get overwhelming really quickly, so we added the new Jira Component to support tasks like:
Get boards, issues, epic, sprint
Using 💧 Instill VDP, the Jira Component can connect with existing AI Components to solve the issue of scattered information. Easily add AI to automatically align tickets in your existing project management tool.
Try demo on ☁️ Instill Cloud:
More at: Instill AI
Multimedia Processing with Audio and Video Components
Data comes in many formats beyond text, so we've expanded our Operator Components to support a wider range of unstructured data processing methods. This updates introduces 2 new operators - Audio Component and Video Component.
These improvements significantly expand our platform's capabilities to split audio files into multiple chunks and sub-sample videos to provide you with more robust tools for unstructured data manipulation on 🔮 Instill Core.
Try demo on ☁️ Instill Cloud:
More at: Instill AI
SQL Component
We have included the SQL Component in this release to help you connect to your SQL databases: MySQL, PostgreSQL, MariaDB, Firebird, and Oracle. This integration supports SQL in your data pipelines on 💧 Instill VDP to allow for more flexible data movement and support the following tasks:
Insert
Update
Select
Delete
Create Table
Drop Table
More at: Instill AI
Local Full-Stack AI with Ollama Component
The Ollama Component is a key addition to 🔮 Instill Core, extending our full-stack AI platform to your local laptops! Self-hosting your own 🔮 Instill Core is ideal for secure, efficient, and flexible AI development and deployment.
With this integration, you can access and utilize AI models served on the Ollama for tasks such as:
Text generation chat
Text embedding
For setup, refer to our deployment documentation and the Ollama server tutorial.
Note: You will need to adjust your Docker network settings to ensure connectivity between 🔮 Instill Core and Ollama containers.
More at: Instill AI
GPT-4o mini: Cost-efficient small LLM
OpenAI recently released the most cost-efficient small LLM on the market, GPT-4o mini, so we updated our OpenAI Component to support the new model, unlocking the latest AI powers including:
Low cost & latency for diverse tasks
Text & vision API support
Enhanced non-English text processing
99.5% cost reduction per 1,000 Input Tokens vs GPT-4 (see docs)
Try the demo on ☁️ Instill Cloud:
More at: Instill AI
Mistral: Accelerate building AI
We have added Mistral Component, a powerful new addition to our AI Components family on 🔮 Instill Core.
Key Features:
Text Generation Chat
Text Embeddings
This component enables seamless integration with AI models served on the Mistral Platform, significantly accelerating AI development processes.
Benefits:
Streamlined access to Mistral's advanced AI models
Enhanced text generation and embedding capabilities
Simplified AI integration for developers
Leverage the Mistral component to boost your AI-driven applications and services. Explore the new possibilities in text generation and embeddings today!
Try this short-film script writer pipeline on ☁️ Instill Cloud.
Cohere: Advanced NLP Capabilities
We've expanded our list of AI Components with the new Cohere Component to enable powerful Natural Language Processing (NLP) capabilities on 🔮 Instill Core.
Key Additions:
Text Generation Chat
Text Embeddings
Text Reranking
Expanded NLP Capabilities:
Leverage Cohere's Large Language Models (LLMs) on 🔮 Instill Core for various natural language use cases:
Classification
Semantic search
Paraphrasing
Summarization
Content generation
Try it on ☁️ Instill Cloud:
Simple Prompting Pipeline: Text Generation Chat with Cohere's AI model.
CVE-Search Pipeline: Text Reranking for optimized searching on the NIST (National Institute of Standards and Technology) database.
This integration empowers you to harness Cohere's advanced Language Models directly within 🔮 Instill Core, opening up new possibilities for sophisticated NLP applications.
We're eager to see how you'll use these new capabilities in your projects!
More at: Instill AI
Serverless Model Serving: Optimize AI deployment
The Serverless Model Serving feature on ⚗️ Instill Model offers a more cost-efficient solution for AI deployment by eliminating the need to manage the underlying infrastructure. ⚗️ Instill Model handles the compute resource provisioning, auto-scaling and availability, allowing you, the developers to focus on model development and deployment 🎉.
Key Benefits:
Scalability: Automated resource allocation scales to handle varying workloads efficiently.
Cost-Efficiency: Compute resources are provisioned on-demand and scaled down to zero when idling.
Simplified Management: No need for server management, free up your teams to focus on improving models and applications.
Quicker Deployment: Seamless integration with 🔮 Instill Core for faster full-stack AI development with data pipeline and knowledge bases.
We've also made significant performance enhancements:
Faster image push times
Reduced model cold-start times
We’re continuously enhancing ⚗️ Instill Model and will soon introduce Dedicated Model Serving for production-purpose model serving service.
☁️ Instill Cloud users can host public models for free and we offer monthly free Instill Credits for ☁️ Instill Cloud users to run any public models hosted on ☁️ Instill Cloud.
More at: Instill AI
GitHub: Boost developer workflow
The new GitHub Component on 🔮 Instill Core allows developers to integrate and automate various GitHub operations, boosting development workflow efficiency. Key tasks include:
Listing and retrieving pull requests and commits
Managing review comments
Handling issues—listing, retrieving, and creating
Setting up automated notifications and actions via webhooks
Try it on ☁️ Instill Cloud:
We will soon introduce a new event-triggering mechanism for 💧Instill VDP which can be coupled with this to build even more powerful workflow automations.
More at: Instill AI
Anthropic's Claude 3.5 Sonnet is here
We've added Anthropic as a new AI component on 🔮 Instill Core, allowing users to connect to the AI models served on the Anthropic Platform. Get access to models including the Claude-3-family and most recently, the Claude-3.5-Sonnet. Claude 3.5 Sonnet sets new standards in reasoning, knowledge, and coding, excelling in nuance, humour, and high-quality content. We currently support these tasks.
To quickly get started,
Go to this pipeline on ☁️ Instill Cloud
Type in your input like "What is Instill Core?"
Clone this pipeline & start building your AI projects using the Anthropic AI component 🙌
Build Smarter Emails
You've got a new update! The Email Component on 🔮 Instill Core introduces ways to send and read your emails from the Mail Protocol. You can connect to different email servers through the Email component including Gmail, Outlook, Yahoo, and iCloud.
💌 Try sending an email using this pipeline on ☁️ Instill Cloud.
Share Resources Efficiently via Namespace Switch
Our new Namespace Switch makes it easier for you to toggle between your individual or organization's namespace. It makes browsing all available resources much quicker and helps you use your Instill Credit more efficiently.
To get started,
Join or create a new organization.
Click the namespace toggle next to the logo in the top left corner to switch between your different namespaces.
Done! You can now use the resources of the currently selected namespace 🎉.
Save Your Time with Instill Credit
Our new feature, Instill Credit, makes it easy to adopt Instill Cloud by reducing the time needed to build and set up pipelines.
You can create AI components with out-of-the-box configurations, avoiding the need for external accounts or API keys.
Start Right Away
Get started right away with 10,000 free credits every month 🤩. That means you can dive right in and get started without worrying about costs.
We've integrated Instill Credit seamlessly into AI components, so accessing models from providers like OpenAI and Stability AI is now a breeze. Wondering if you're using Instill Credit? Just look for Instill credentials preset using ${secrets.INSTILL_SECRET} as input in the supported components.
👉 Learn more about how Instill Credit works. We're continuously expanding support for more AI components with Instill Credit.
If you prefer to use your own keys, simply create a secret on your profile secrets page and reference it in the component configuration.
More Credits on the Way
But wait, there's more! We're gearing up to introduce Credit top-up and auto-billing options soon. Stay tuned for seamless operations without worrying about running out of credits.
Boost Workforce with Slack
We've added a new Instill Component, Slack, to Instill Core, and our team can't wait to use it! This addition enables you to build pipelines to read and write messages from your Slack workspace, making your workflow smoother and communication even better.
With this new feature, you can do cool stuff like:
Send Slack messages to LLM and get back helpful insights and actions powered by AI
Automate routine tasks, freeing up your team to focus on higher-value work
Try It Now!
Create and install a Slack app
Follow the Slack API tutorial to create and install a Slack app in your workspace. You'll get a Slack API bot token like "xoxb-******", which we'll need later.
Set up your Slack token as a secret
Go to Instill Console, click on your Avatar in the top right corner, then go to Settings > Secrets > Create Secret. Use the token you got in the first step to create a secret.
Build a pipeline with the Slack component
Go to the pipeline builder page, click on Component+ at the top left and Select Slack. Use the secret you created in step 2 as the token input in the Slack connection section. Now you're all set to start sending and receiving messages. Just make sure your Slack app is in the right channel.
So go ahead, have some fun with it! Whether you're using it to summarize messages, crack a few jokes, or just boost your team's productivity, we hope it helps you create a more connected and efficient work environment.
Introducing Secret Management for Connector Configuration
We've made some big improvements to how you set up your pipeline connectors, especially when it comes to managing sensitive information like API keys.
Now, you can create secrets, which are encrypted bits of sensitive data. These secrets are reusable and secure. You can use these secrets in your pipelines, but only if you specifically reference them in the connector setup.
When you head to the Settings > Secrets tab, you'll see a complete list of all your secrets. Once you've set them up, you can easily use them in your pipeline connectors.
Let's walk through setting up the OpenAI connector as an example:
To start, go to Settings > Secrets, and click on Create Secret. Enter your OpenAI API key with the ID openai-api-key.
Now, drag and drop the OpenAI connector onto the canvas. Then, access the OpenAI Connection section by clicking on More. Input ${ and select your OpenAI secret from the dropdown ${secrets.openai-api-key}.
With the OpenAI connector configured and your own API key in place, you're all set to build your pipeline. Any costs incurred will go directly to the service provider.
What's Next 🔜
In addition to the above Bring Your Own API Key (BYOK) feature, we're thrilled to introduce the upcoming Instill Credit for Instill Cloud! As a bonus, FREE Instill Credits will be available to all Instill Cloud users. Once it's ready, you won't need to register your own 3rd-party vendor account; these credits will grant access to AI models like OpenAI's GPT-4 without any additional setup required.
Stay tuned for more updates!
Instill Cloud Expands: Now Available in North America 🌎
With our recent deployment in the United States, Instill Cloud marks a significant milestone in our global expansion journey. We're extending our commitment to providing seamless cloud solutions across continents.
From Europe to Asia and now North America, our goal remains unchanged: to empower businesses worldwide with reliable, high-performance cloud services. As we continue to grow and evolve, stay tuned for more updates on how Instill Cloud is transforming the way organizations build AI applications.
Llama 3 8B Instruct Model Now Available on Instill Cloud
Just two days ago, Meta unveiled its most powerful openly available LLM to date, Llama 3. For more information, visit here.
Today, you can access the Llama 3 8B Instruct model on Instill Cloud!
How to Use
Here are the steps to use it:
Log in to Instill Cloud.
Use our provided pipeline to compare the capabilities of Llama 3 and Llama 2 side by side.
Alternatively, you can directly call the model. See the API request below:
Remember to replace <INSTILL CLOUD API TOKEN> in the API request with your own Instill Cloud API Token.
Instill Cloud Goes Global: From Europe to Asia 🌏
Instill Cloud platform is now deployed across multiple regions!
From the Netherlands in Europe, we've extended our footprint to Singapore in Asia, enhancing accessibility and performance for our users across diverse geographical locations. The US region will join our global network next, further enhancing accessibility for users across North America.
👉 Try out Instill Cloud today and stay tuned for updates as we continue to extend the reach of the Instill Cloud platform to serve you better.
Introducing Iterators for Batch Processing in Your Pipeline
We're excited to introduce a new component type: Iterators. Iterators are designed to process each data element individually within an array or list. This functionality is crucial for efficiently handling data in batches within your pipeline.
Iterator ( [ 🌽 , 🐮 , 🐓 ] , cook) => [ 🍿 , 🥩 , 🍗 ]
Enhance Your Workflows with Iterators
Imagine you're summarizing a webpage with extensive content. With an iterator, you can process the substantial content of a webpage, breaking it down into manageable chunks segmented by tokens. This enables you to efficiently generate concise summaries for each section of the page.
👉 Try out the pipeline Website Summarizer with Iterator.
LLaVA Level-Up: From 7B to 13B 🔼
We've leveled up LLaVA model on Instill Cloud from LLaVA 7B to LLaVA 13B!
👉 Dive into our updated tutorial to discover how to unlock the enhanced capabilities of LLaVA.
👉 We've built a visual assistant pipeline using LLaVA 13B. Feel free to try it out by uploading an image and ask one question about the image.
Introducing LLaVA 🌋: Your Multimodal Assistant
📣 The latest LLaVA model, LLaVA-v1.6-7B, is now accessible on Instill Cloud!
What's LLaVA?
LLaVA stands for Large Language and Vision Assistant, an open-source multimodal model fine-tuned on multimodal instruction-following data. Despite its training on a relatively small dataset, LLaVA demonstrates remarkable proficiency in comprehending images and answering questions about them. Its capabilities resemble those of multimodal models like GPT-4 with Vision (GPT-4V) from OpenAI.
What's New in LLaVA 1.6?
According to the original blog post, LLaVA-v1.6 boasts several enhancements compared to LLaVA-v1.5:
Enhanced Visual Perception: LLaVA now supports images with up to 4x more pixels, allowing it to capture finer visual details. It accommodates three aspect ratios, with resolutions of up to 672x672, 336x1344, and 1344x336.
Improved Visual Reasoning and OCR: LLaVA's visual reasoning and Optical Character Recognition (OCR) capabilities have been significantly enhanced, thanks to an improved mixture of visual instruction tuning data.
Better Visual Conversations: LLaVA now excels in various scenarios, offering better support for different applications. It also demonstrates improved world knowledge and logical reasoning.
Efficient Deployment: LLaVA ensures efficient deployment and inference, leveraging SGLang for streamlined processes.
👉 Dive into our tutorial to learn how to leverage LLaVA's capabilities effectively.
New and Improved Smart Hint
We've improved the Smart Hint feature to offer contextual advice and suggestions throughout your pipeline.
Here's how it functions:
Firstly, Smart Hint informs users about the type of content an input field can accept, indicated through placeholder text, making it clear whether references are allowed.
Upon interacting with the input field, Smart Hint reveals the specific type of data it supports, such as String, Number, Object, etc.
Moreover, when a user enters ${ into an input field, Smart Hint intelligently filters and presents only the relevant hints that are compatible with the field.
This structured guidance method simplifies the process of building your pipeline, enhancing productivity and ease of use.
jq JSON Processor Integration
Take control of your JSON data with seamless jq filter integration within the JSON Operator. Apply custom jq filters directly to your data streams, enabling advanced manipulation and extraction with precision.
To get started, head to the JSON Operator and select TASK_JQ.
Video Input Support
Unlock limitless possibilities with the addition of video input support. Seamlessly ingest and process video content within your workflows, enabling dynamic data analysis, transformation, and the use of models that support video content.
Improvements
Improved performance and stability ensure a seamless user experience.
Enhanced user interface with refined visual cues and streamlined navigation for increased productivity
Thank You for Your Contributions 🙌
We highly value your feedback, so feel free to initiate a conversation on GitHub Discussions or join us in the #general channel on Discord.
Huge thanks to rsmelo92 for their contribution!
Instill Python SDK: Pipeline Creation made Easier
We are thrilled to announce a significant upgrade to our Python SDK, aimed at streamlining and enhancing the pipeline creation and resource configuration process for connectors and operators.
One of the key challenges users faced in previous versions was the difficulty in understanding the required configurations for connectors and operators. With the implementation of JSON schema validation, type hints, and helper classes that define all the necessary fields, users no longer need to refer to extensive documentation to configure resources. This ensures a more straightforward and error-free setup.
This update marks a significant step forward in improving user experience and reducing friction in the configuration process. We believe that these enhancements will empower users to make end-to-end use of the Python SDK; resource configuration, pipeline creation, debugging/editing/updating, and deployment from within their tech stack.
You can find a complete pipeline setup example with Python SDK on our GitHub repo.
Your feedback is crucial to us, so please don't hesitate to share your thoughts and experiences with the improved configuration workflow. We're committed to continually enhancing our platform to meet your needs and provide you with the best possible user experience.
To set up and get started with the SDKs, head over to their respective GitHub Repos:
New Pipeline Component Reference Schema: ${}
We're excited to introduce an enhanced Component Reference Schema that will make pipeline building even more straightforward. With this update, referencing a component field is as simple as using ${}. When establishing connections between components, there's no need to distinguish between {} or {{}} anymore. Just type ${ in any form field within a component, and our Smart Hint feature will promptly suggest available connections.
Rest assured, you won't need to manually update your existing pipelines due to this change from our previous reference schemas {} and {{}}. We've seamlessly migrated all pipelines to this new format.
Explore Our OpenAPI Documentation
We have just launched our OpenAPI Documentation, which contains all the essential information required for smooth integration with our APIs. Please visit our OpenAPI Documentation to access comprehensive details. You can even insert your Instill Cloud API token in the Authorization Header to experiment with the APIs on our platform.
Bearer <INSTILL CLOUD API TOKEN>
User-friendly Error Messages
We understand the importance of clear communication, especially when errors occur. That's why we've revamped our error messages to be more user-friendly and informative. You'll now see which component in the pipeline the error is related to, allowing you to quickly identify and address issues, and speeding up the debugging process.
Instill Cloud Console URL Migration
We have successfully migrated the Instill Cloud console URL from https://console.instill.tech to https://instill.tech. Please make sure to update your bookmarks accordingly.
Thank You for Your Contribution 🙌
We highly value your feedback, so feel free to initiate a conversation on GitHub Discussions or join us in the #general channel on Discord.
Huge thanks to @chenhunghan and @HariBhandari07 for their contributions!
Introducing Zephyr-7b, Stable Diffusion XL, ControlNet Canny on Instill Model 🔮
Open-source models empower you and your businesses to leverage them within your own infrastructure, granting you complete control over your data and ensuring the privacy of sensitive information within your network. This approach mitigates the risk of data breaches and unauthorized access, affording you the flexibility of not being bound to a single hosting vendor.
With Instill Model, you gain access to a wide range of cutting-edge open-source models. Now, you can seamlessly use Zephyr-7b, Stable Diffusion XL, and ControlNet Canny in an integrated environment for FREE.
Zephyr-7b is a fine-tuned version of Mistral-7b created by Hugging Face. It is trained on public datasets but also optimized with knowledge distillation techniques. It's lightweight and fast, although 'intent alignment' may suffer. Read more about it and distillation techniques here.
Stable Diffusion XL empowers you to craft detailed images with concise prompts and even generate text within those images!
ControlNet Canny allows you to control the outputs of Stable Diffusion models and hence manipulate images in innovative ways.
We are committed to expanding our offering of open-source models, so stay tuned for more!
Instill VDP, Now on Beta! 🦾
In this Beta release, we are delighted to announce that the Instill VDP API has now achieved a state of stability and reliability. This marks a pivotal moment for Instill VDP, and we are committed to ensuring the continued robustness of our API.
We recognize the importance of a seamless transition for our users and pledge to avoid any disruptive or breaking changes. This commitment means that you can confidently build upon the current version of the Instill VDP API, knowing that your existing integrations and workflows will remain intact.
This achievement represents a major milestone in our journey to provide the best experience for our users. We understand the critical role that stability and consistency play in your development efforts, and we are excited to take this step forward together with you.
As we move forward, we will keep striving to enhance the features of Instill VDP, listening to your feedback, and working closely with our community to deliver even more value and innovation. Thank you for being part of this exciting journey with us.
Support JSON input
Our Start operator now supports JSON inputs. Leveraging JSON as an input format enables the handling of semi-structured data, providing a well-organized representation of information. This is especially valuable when working with a REST API connector as a payload or when utilizing the Pinecone connector to insert records that include metadata in the form of key-value pairs.
Use Redis as a chat history store
We've introduced a Redis connector to facilitate chat history functionality. Redis, a widely used in-memory data store, enables the efficient implementation of a chat history system. Messages from a given chat session, identified by a unique Session ID, can be stored consistently, complete with their associated roles like "system," "user," and "assistant." Moreover, you have the flexibility to configure the retrieval of the latest K messages and integrate them seamlessly with a Large Language Model (LLM), such as OpenAI's GPT models, allowing you to develop a chatbot with full control over your data.
Improve RAG capabilities
We now support metadata in the Pinecone connector to enhance the efficiency and precision of your vector searches. With the ability to associate metadata key-value pairs with your vectors, you can now access not only the most similar records after your data has been indexed but also retrieve the associated metadata. Additionally, you can define filter expressions during queries, enabling you to refine your vector searches based on metadata. This functionality proves invaluable when dealing with extensive datasets where precision and efficiency in retrieval are paramount.
Unlock Instill VDP's full potential with Instill Hub!
Welcome to Instill Hub, your go-to platform for all things Instill VDP! Here, you have the opportunity to explore and use pipelines that have been contributed by fellow members. These pipelines are designed to provide you with the most up-to-date and efficient AI capabilities. You can seamlessly trigger these pipelines for your projects, or if you're feeling creative, you can even clone them and customize them to suit your unique needs.
But that's not all – we strongly encourage you to share your pipelines with the Instill Hub community. By doing so, you'll be contributing to the growth and success of our platform, helping others in the community benefit from your expertise.
So, whether you're eager to explore, eager to contribute, or simply excited to learn from others, Instill Hub is here to foster a supportive and collaborative community around Instill VDP. Join Instill Cloud today and let's work together to make AI innovation accessible to everyone!