Diaflow's Documentation
HomePricingIntegrations
Documentation
Documentation
  • 😎WELCOME TO DIAFLOW
    • Introduction to Generative AI
    • What can you build with Diaflow?
  • πŸ’»USER ACCOUNT
    • Create your user account
    • Delete your user account
    • Log out and log in
    • Change "Personal" & "Workspace" settings
    • Reset user account password
  • πŸš€Getting Started
    • Let's start with the basics
      • How a component works?
      • How a flow works?
      • Creating your first flow
    • Dashboard
      • Services
      • Create a flow from scratch
      • Create a flow from templates
      • View your flows
    • Terminology
  • 🌊Flows
    • Overview
    • Create a flow
    • Delete a flow
    • Manage a flow
    • Publish a flow
    • Unpublish a flow
    • Deployment
    • Component Reference
      • Trigger
        • When inputs are submitted (Apps)
        • Cronjob (Automation)
        • Webhook (Automation)
        • Microsoft Outlook (Automation)
      • Outputs (Apps)
        • Text Output
        • Chart Output
        • Video Output
        • Audio Output
        • Image Output
      • Built in tools
        • Branch
        • Merge (Multiple data source to JSON)
        • Split Data (JSON Formatter)
        • Video to audio
        • Get current date and time
        • Web scraper
        • Document to plain text
        • Retrieve data from spreadsheet (Spreadsheet analyzer)
        • Spreadsheet creator
        • Convert JSON to chart data
        • PDF to image
        • Get weather information
        • HTTP Request
        • Get GEO Location
        • SMTP
        • Loop
      • Built in resources
        • Diaflow Vision
        • Diaflow Vectors
        • Diaflow Drive
        • Diaflow Table
      • Apps
        • Hunter.io
        • Outlook Email
        • Telegram
        • Slack
        • Python
        • YouTube
        • SerpAPI
        • Google Sheet
          • Document-level Operations
          • Sheet-level Operations
          • Data-level Operations
      • Database
        • MySQL
        • Microsoft SQL
        • PostgreSQL
        • Snowflake
      • Private AI/LLM Models
        • OpenAI
          • GPT Variants
          • GPT Vision
          • DALL-E Variants
          • TTS Variants
          • Whisper
        • Anthropic
        • Llama
        • Google Gemini
        • Cohere
        • MistralAI
      • Public AI/LLM Models
        • OpenAI Cloud
        • Perplexity Cloud
        • Deepseek Cloud
        • Anthropic Cloud
        • Replicate
        • Straico
        • OpenRouter
        • Cohere Cloud
        • Google Gemini Cloud
        • MistralAI Cloud
        • ElevenLabs Cloud
      • AI Tools
  • βœ’οΈPRODUCTIVITY TOOLS
    • Tables
    • Drive
    • Vectors
      • Document
      • Article
      • URLs
  • 🏠Workspace
    • History
    • Teams
    • Billing & Subscription
      • Upgrade/Downgrade a subscription
      • Buy credits
      • Credit Usage
      • Cancel a subscription
    • Settings
      • Personnal
      • Workspace
        • Change workspace
        • Workspace settings
        • Custom Domain
        • Delete workspace
      • Change Language
    • Documentation
    • Integrations
    • API keys
  • πŸ“‘Other
    • FAQs
    • Contact Information
Powered by GitBook
On this page
  • Description
  • Inputs
  • Component settings
  • Advanced configurations
  • Outputs

Was this helpful?

  1. Flows
  2. Component Reference
  3. Public AI/LLM Models

ElevenLabs Cloud

Use ElevenLabs's models with your own API Key

Last updated 4 days ago

Was this helpful?

Description

The ElevenLabs node allows you to convert text into lifelike speech using the ElevenLabs Text-to-Speech API. With this node, you can select from various models and voices offered by ElevenLabs to generate high-quality, natural-sounding audio from input text.

This node is ideal for use cases such as:

  • Creating voiceovers for videos or presentations

  • Generating spoken feedback or summaries

  • Building voice-enabled assistants

  • Enhancing accessibility features with audio output

To use this node, you need a valid ElevenLabs API credential. Once configured, you can define the action (currently supports β€œText to Speech” and "Speech to Text"), choose the model and voice, and input the text you want to convert.

The resulting audio file can be used in subsequent steps of your workflow or delivered to users.

Inputs

The ElevenLabs component has the following input connections.

Input Name
Description
Constraints

From data Loaders/ Data source/Vector DB

This input connection represents the context information for the ElevenLabs model

Must originate from a Data Loader/Data Source or VectorDB component.

From Input

This input connection represents the user query for the ElevenLabs model.

Must originate from a component that generates a text string as output such as a Python or Text Input component.

Component settings

Parameter Name
Description

Credentials

You can specify to use your own ElevenLabs credentials

Action

Choose the action: - Text to Speech - Speech to Text

Model

The list of models displayed depends on the customer's credentials.

Voice (With action = Text to Speech)

The list of voices displayed depends on the customer's credentials.

Text (With action = Text to Speech)

This is the input field where you provide the content to be converted into speech.

You can either type plain text directly or refer to the output of previous nodes by using the @ symbol to select dynamic data from earlier steps in your workflow. This allows you to generate speech from content that was extracted, generated, or transformed in previous nodes β€” for example, scraped text, AI-generated summaries, or user input.

Examples:

  • Static text: Hello, welcome to our service.

  • Dynamic text: @web-scraper.output.content (refers to content scraped from a website)

Make sure the referenced data is in plain text format for optimal speech synthesis quality.

Audio file

This field is used to provide the input audio that you want to transcribe into text.

You can either:

  • Upload a direct audio file (in supported formats like .mp3, .wav, etc.), or

  • Dynamically reference audio output from previous nodes in the workflow by using the @ symbol β€” for example, an audio file URL returned by a recording tool, voice assistant, or web scraper.

Example usages:

  • Static file reference: https://example.com/audio/sample.mp3

  • Dynamic reference: @recorder.output.audioFileUrl

Make sure the referenced file is accessible via a valid URL or is passed from a previous node that provides audio in a compatible format.

Advanced configurations

Options
Description

Stability (With action = Text to Speech)

This setting controls how consistent the voice sounds between different runs.

  • Higher values (closer to 1) produce more stable and predictable speech output.

  • Lower values introduce more variation and spontaneity in the generated voice.

Similarity boost (With action = Text to Speech)

This adjusts how strongly the generated voice should try to match the original voice profile.

  • Higher values make the voice more closely resemble the reference voice, which may reduce expressiveness.

  • Lower values allow for more natural variation but may sound less like the reference voice.

Enable catching

This enables local caching of generated audio to improve performance and reduce repeated API calls for identical input.

  • When turned on, previously generated audio will be reused for the same input.

  • You can also configure the Caching time, which determines how long the result stays in cache before being refreshed.

  • Includes an option to Clear cache manually.

Outputs

The ElevenLabs component has the following output connections.

Output Name
Description
Constraints

To Output

This output connection contains the esult of the ElevenLabs component.

Can be connected to any component that accepts a string input.

🌊