Diaflow's Documentation
HomePricingIntegrations
Documentation
Documentation
  • 😎WELCOME TO DIAFLOW
    • Introduction to Generative AI
    • What can you build with Diaflow?
  • 💻USER ACCOUNT
    • Create your user account
    • Delete your user account
    • Log out and log in
    • Change "Personal" & "Workspace" settings
    • Reset user account password
  • 🚀Getting Started
    • Let's start with the basics
      • How a component works?
      • How a flow works?
      • Creating your first flow
    • Dashboard
      • Services
      • Create a flow from scratch
      • Create a flow from templates
      • View your flows
    • Terminology
  • 🌊Flows
    • Overview
    • Create a flow
    • Delete a flow
    • Manage a flow
    • Publish a flow
    • Unpublish a flow
    • Deployment
    • Component Reference
      • Trigger
        • When inputs are submitted (Apps)
        • Cronjob (Automation)
        • Webhook (Automation)
        • Microsoft Outlook (Automation)
      • Outputs (Apps)
        • Text Output
        • Chart Output
        • Video Output
        • Audio Output
        • Image Output
      • Built in tools
        • Branch
        • Merge (Multiple data source to JSON)
        • Split Data (JSON Formatter)
        • Video to audio
        • Get current date and time
        • Web scraper
        • Document to plain text
        • Retrieve data from spreadsheet (Spreadsheet analyzer)
        • Spreadsheet creator
        • Convert JSON to chart data
        • PDF to image
        • Get weather information
        • HTTP Request
        • Get GEO Location
        • SMTP
        • Loop
      • Built in resources
        • Diaflow Vision
        • Diaflow Vectors
        • Diaflow Drive
        • Diaflow Table
      • Apps
        • Hunter.io
        • Outlook Email
        • Telegram
        • Slack
        • Python
        • YouTube
        • SerpAPI
        • Google Sheet
          • Document-level Operations
          • Sheet-level Operations
          • Data-level Operations
      • Database
        • MySQL
        • Microsoft SQL
        • PostgreSQL
        • Snowflake
      • Private AI/LLM Models
        • OpenAI
          • GPT Variants
          • GPT Vision
          • DALL-E Variants
          • TTS Variants
          • Whisper
        • Anthropic
        • Llama
        • Google Gemini
        • Cohere
        • MistralAI
      • Public AI/LLM Models
        • OpenAI Cloud
        • Perplexity Cloud
        • Deepseek Cloud
        • Anthropic Cloud
        • Replicate
        • Straico
        • OpenRouter
        • Cohere Cloud
        • Google Gemini Cloud
        • MistralAI Cloud
        • ElevenLabs Cloud
      • AI Tools
  • ✒️PRODUCTIVITY TOOLS
    • Tables
    • Drive
    • Vectors
      • Document
      • Article
      • URLs
  • 🏠Workspace
    • History
    • Teams
    • Billing & Subscription
      • Upgrade/Downgrade a subscription
      • Buy credits
      • Credit Usage
      • Cancel a subscription
    • Settings
      • Personnal
      • Workspace
        • Change workspace
        • Workspace settings
        • Custom Domain
        • Delete workspace
      • Change Language
    • Documentation
    • Integrations
    • API keys
  • 📑Other
    • FAQs
    • Contact Information
Powered by GitBook
On this page
  • Description
  • Inputs
  • Component settings
  • Advanced configurations
  • Outputs
  • Use case

Was this helpful?

  1. Flows
  2. Component Reference
  3. Private AI/LLM Models
  4. OpenAI

Whisper

Generate ASMR-inducing audio, providing immersive auditory experiences in your flows.

Last updated 1 month ago

Was this helpful?

Description

The OpenAI component allows you to integrate OpenAI Whisper generation into your flows. In particular, the following versions are supported:

  • Whisper

Whisper is an AI model developed by OpenAI. Whisper is a powerful tool for converting spoken language into text, with capabilities that extend to multiple languages and various audio conditions.

The OpenAI component has the identifier of opa-X, where X represents the instance number of the OpenAI component.

Inputs

The OpenAI component has the following input connections.

Input Name
Description
Constraints

From Data Loaders

This input connection represents the context information for the OpenAI model.

Must originate from a Data Loader, Data Source or VectorDB component.

From Audio Input

This input connection represents the user query for the OpenAI model.

Component settings

Parameter Name
Description

Credentials

You can specify to use your own OpenAI credentials or alternatively you can use Diaflow's default credentials.

Model

This parameter specifies the version of OpenAI that the component should use. Available values: - Whisper-1

Endpoint

Available options: - Transcriptions - Translations

Data source

Advanced configurations

Options
Description

Temperature

This parameter is used to control the level of randomness or creativity in the generated ASMR content. A lower temperature value results in more deterministic and conservative outputs, while a higher temperature value encourages more diverse and creative variations in the generated audio. Adjusting the temperature parameter allows users to fine-tune the balance between predictability and novelty in the Whisper-generated ASMR content, catering to individual preferences for relaxation and sensory pleasure.

Response format

Specified the format of the output audio data.

language

For more accurate results, specify the input language from this dropdown box.

Prompt

Describes how you want the OpenAI model to respond. For example, you can specify the role, manner and rules that OpenAI should adhere to. Also mention the component ID to connect the components.

Outputs

The OpenAI component has the following output connections.

Output Name
Description
Constraints

To Output

This output connection contains the audio data result of the OpenAI component

Can be connected to any component that accepts a string input or audio input.

Use case

Here is a simple use case of the Whisper component, where the Whisper component is being used with the whisper model to translate an audio into a text.

🌊