November 11, 2024
VESSL Storage supports Google Cloud Storage (GCS)

You can now integrate VESSL Storage with GCS for scalable and secure data storage. Try connecting with GCS now.

Go to VESSL Storage documentation for GCS connection
November 1, 2024
Introducing the All-New VESSL Storage Update

We are excited to announce the release of our new VESSL Storage update. In the previous version, we faced several challenges:

  • Ambiguous definitions: The definitions of Storage, Dataset, Model, and Artifact were not clear.
  • Complex operations: The methods for importing, mounting, and exporting were not straightforward.
  • Unclear artifact functionality: Understanding the function and role of Artifacts was difficult.

To resolve the above issues, we proactively developed followings:

Key Updates

  • Unified “Volume” concept: We’ve integrated artifacts, models, logs, and datasets into a single unit called Volume.
  • Seamless integration with workloads: Easily integrate Volumes with runs, workspaces, services, and pipelines.
  • Enhanced storage components:
    • Support for VESSL Storage and External Storage: Users can store Volumes in VESSL Storage or external storage solutions like AWS S3, GCP Storage (to be released in early November), and on-premise NFS systems. Files and directories are fully managed in VESSL Storage through automatic storage provisioning.
      • VESSL Storage: Optimized for use with VESSL features and ready to use immediately without any integration process.
      • External Storage: Allows users to retain data ownership and use data on VESSL without data migration, offering both high security and convenience.
  • Simplified import, mount, and export operations in the VESSL features:
    • Previous version:
      • Import: Code / HuggingFace / Dataset / Model / GCS / S3 / Files / Artifact
    • Current version:
      • Import: Code / HuggingFace / Volume / Model
      • Mount: Volume / GCS Fuse
      • Export: Volume / Model
  • Manage exported data in Volumes: In VESSL Storage’s Volumes, users can view exported data such as logs, metrics, and model checkpoints from VESSL Runs and Workspaces.
October 8, 2024
VESSL AI Raises $12M in Series A Funding

VESSL AI has successfully raised $12 million in Series A funding, bringing our total funding to $16.8 million.

Full article on TechCrunch

This funding reinforces our commitment to advancing AI orchestration and integrated MLOps. We extend our gratitude to our customers, partners, and investors: A Ventures (Series A lead), Ubiquoss, Mirae Asset, Sirius Investment, SJ Investment Partners, Wooshin Venture Investment, Shinhan Venture Investment, Oracle, Hyundai Motors, and Upstage.

October 4, 2024
CLI update announcement

We have updated our CLI commands to enhance functionality and improve user experience. As part of the recent updates, including the renaming of VESSL Serve to VESSL Service and the Pipeline GA, the following changes have been made to the VESSL CLI:

Deprecated Commands

  1. vessl serve update
  2. vessl serve revision list, vessl serve revision show, vessl serve revision terminate
  3. vessl serve gateway show

New Commands

  • vessl service create
  • vessl service list
  • vessl service read
  • vessl service terminate
  • vessl service scale
  • vessl service split-traffic
  • vessl service create-yaml

Notes

  • If you have scripts or workflows using the deprecated commands, please update them to use the new commands.
  • For more information on each command, use the --help option. For example: vessl service create --help
August 31, 2024
Pricing plan section updated in VESSL documentation

We’ve created the pricing plan section in the Resources section. Also, our pricing plan (GCP, AWS) details have been updated in the VESSL documentation. Users now have more clarity on compute options and corresponding costs. Pro users continue to receive 100 credits every month, with each credit equivalent to $1.00. For more details, refer to the updated Pricing & Compute section in the documentation.

August 5, 2024
Pipeline general availability (GA) update

VESSL Pipeline has reached general availability, designed to enhance the execution of complex ML workflows such as LLM fine-tuning, data preprocessing, and batch inference.

  • Key features:
    • Drag-and-drop GUI: Intuitive interface for modifying and visualizing pipeline flows.
    • Infra-as-code: YAML-based code interface integration, enabling effective version management of pipeline modifications.
    • High visibility and debuggability: Improved debugging capabilities with natural separation of task stages, including endpoint access and re-execution for failed tasks.
    • Human-in-the-loop: Built-in support for scenarios requiring user intervention, such as intermediate result feedback and decision-making based on input/output.

Three early adopters have successfully integrated their services into the running pipeline. If you’re interested in integrating the VESSL Pipeline, please contact our sales team.

Learn more about Pipeline

With the Pipeline section, you can explore it in detail.

July 12, 2024
Website V2 Update

We are excited to announce the launch of VESSL 2.0, introducing a sleek, user-centric interface designed to streamline the MLOps experience.

  • New features:
    • Self-service user interface: Seamless transition from model exploration to deployment, with tools like VESSL Hub for testing and fine-tuning open-source models, and VESSL Service for creating scalable APIs.
    • Service revision creation through web console: Previously only available through CLI, users can now create service revisions in both Provisioned and Serverless Mode through the UI.
    • CMD+K navigation: Quick access to any entity within VESSL, enhancing productivity and efficiency.

With VESSL 2.0, you can enjoy the sleek new interface, intuitive web console, and powerful CMD+K navigation. Visit our website now.

VESSL offers a unified interface across multiple cloud providers and on-premise servers, facilitating large-scale machine learning deployments.

May 12, 2024
Serverless deployment

Our serverless deployment infrastructure is the easiest way to scale inference workloads on remote GPUs. With continuous batching, effortless autoscaling, fast cold start, full observability, and more, your APIs are production-ready for full-spectrum AI & LLM applications.

  • Key features:
    • Cost efficiency: Serverless Mode operates on a scale-to-zero basis, ensuring that users only pay for the resources they actually use.
    • Automatic scaling: Real-time scaling based on workload demands without the need for complex configurations.
    • Simplified deployment: Minimal configuration required, making deployment accessible to all users.
    • High availability and resilience: Fast startup times (average 17 seconds) and robust infrastructure ensure high availability with minimal cold starts.

Refer to our docs to put custom Llama 3 in action Text Generation Inference (TGI), in 3 simple steps.

  1. Create a remote GPU-accelerated container
  2. Create an endpoint with Llama 3 from Hugging Face
  3. Send an HTTPS request to the deployed service
April 12, 2024
Announcing VESSL Serve

VESSL Service is the easiest way to Deploy custom models & generative AI applications and scale inference with ease. Deploy any models, to any clouds, at any scale in minutes without wasting hours on API servers, load balancing, automatic scaling, and more. Read our release post or try out the Llama 3.1 example to learn more.

March 11, 2024
Cloud storage support for VESSL Run

Import your data from and export results to a cloud storage like AWS S3 and GCP GCS for your run. You can also bring your own cloud storage by adding the credential of your cloud storage on our improved Secrets page. Refer to our docs for a step-by-step guide.

Google Cloud Storage FUSE

We are bringing FUSE support for GCS. FUSE helps you work with object storage through familiar filesystem operations without needing to directly use the proprietary GCS SDKs.

January 31, 2024
New get started guide

We’ve updated our documentation with a new get started guide. The new guide covers everything from product overview to the latest use cases of our product in Gen AI & LLM.

Follow along our new guide here.

New & Improved

  • Added a new managed cloud option built on Google Cloud
  • Renamed our default managed Docker images to torch:2.1.0-cuda12.2-r3
December 28, 2023
Announcing VESSL Hub

VESSL Hub is a collection of one-click recipes for the latest open-source models like Llama2, Mistral 7B, and StableDiffusion. Built on our fullstack AI infrastructure, Hub provides the easiest way to explore and deploy models.

Fine-tune and deploy the latest models on our production-grade fullstack cloud infrastructure with just a single click. Read about the release on our blog or try it out now at vessl.ai/hub.