Skip to main content
Back to Product Updates

What's new in Q1 2025

· 3 min read
  • Introducing Prompt Studio
  • Generative AI and AI agents in workflows
  • LLM deployment
  • LLM connection management
  • Operator to call endpoints
  • Improved navigation of cloud storage contents
  • Workload history
  • Project-specific coding environments
  • Improved UX for notebooks
  • GPU support for Workspaces

Introducing Prompt Studio

Engineer prompts for high-quality LLM applications in our new Prompt Studio interface, and easily re-use these prompts in workflows.

Our integrated prompt engineering tool enhances the AI Cloud platform by centralizing and streamlining prompt development, allowing users to craft high-quality LLM applications by fine-tuning prompts and integrating them into workflows. Read more

Prompt Studio


Generative AI and AI agents in workflows

Use pre-configured operators to leverage LLMs and build agents in visual workflows, allowing to interact with and execute various tasks on the platform automatically.

Prompt operator and Agent operator available in Workflow Designer under Prompt Engineering folder. This collection of operators includes the pre-configured and the configurable ones. Read more

Generative AI


LLM deployment

Easily deploy public and private Hugging Face LLMs as REST API endpoints, and use them in workflows.

Integrate both public and private Huggingface models into projects with ease, by deploying them and build workflows around them with the Prompt and Agent Operators (for this you'll need a working LLM connection). Read more

LLM deployment


LLM connection management

Centrally manage and share connections to both locally deployed and external API-based LLMs, and re-use connections in Prompt Studio and workflows.

Prompt Studio and Agent operators are built to be LLM agnostic, meaning you can work with any LLM via the LLM Connection. Note: for your local LLM deployments, the connection will be automatically created, this is coming with our next release soon. Until then you have to manually create it. Read more

LLM connection


Operator to call endpoints

Easily connect to REST API deployments within and outside of AI Cloud. Select accessible deployments via a dropdown parameter, or manually add API details as needed.

Call Endpoint operator was added to Workflow Designer to simplify calling any API deployments within your workflow. The parameter dropdown helps to find what is accessible to you across projects, but you can add details manually if you want.

Operator to call endpoints


Improved navigation of cloud storage contents

Browse and select files from cloud storages such as AWS S3, Google Cloud Storage, Azure Blob Storage and Sharepoint more easily, removing the need for entering file paths manually.

Improved navigation of cloud storage contents


Workload history

Inspect past workloads in addition to monitoring actively running workloads. Flexibly filter workloads by time ranges and investigate potential issues by the checking logs of all past workloads (for the last 60 days).

Workload history


Project-specific coding environments

Create Python coding environments for specific projects, helping to experiment faster without consulting admins.

Project-specific coding environments


Improved UX for notebooks

Benefit from automatic loading of the associated project and its contents when starting new notebook instance.

Future Notebook enhancements: We’re expanding our code capabilities. Notebooks will be integrated into Workspaces and support configurable resources, including GPUs.

Improved UX for notebooks


GPU support for Workspaces

Leverage GPUs when developing code in the integrated code development environment.

GPU support for Workspaces