Learn how to create your first voice pipeline for real-time conversational AI. This guide will walk you through logging in, creating a pipeline, and testing it in our playground.

Sign Up and Login

  1. Visit dash.layercode.com
  2. Sign up or log in using email and password, then verify your email.
  3. You’ll be directed to your dashboard where you can manage your pipelines.

Configure Your Voice Pipeline

After logging in for the first time, you’ll be redirected to your first pipeline, created from our recommendedtemplate.

Pipelines can be customized through an intuitive UI with settings for transcription, text-to-speech, and backend (which generates the AI’s response to be spoken). Click the “Edit” button to on any box in the pipeline to configure it.

Feel free to leave all the default settings as is, and skip to testing your pipeline below.

Let’s take a look at the settings available for each stage of the pipeline:

Transcription Settings

Configure how user speech is converted to text. The default transcription provider and model are optimized for low-latency English language transcription. For multi-language support, there are specialized transcription models we support.

SettingDescriptionOptions / Notes
ProviderThe transcription provider used for speech-to-text.See available transcription providers
ModelSelect the transcription model that best fits your needs.See available transcription models
Turn Taking ModeDetermines how the system detects when the user is speaking.Automatic: AI detects when user has finished speaking (using silence detection).
Push to Talk: User controls when they’re speaking by holding down a button.
Read more about turn taking.
Can InterruptWhen Turn Taking Mode is ‘Automatic’: Toggle whether the user can interrupt AI whilst it’s speakingEnable or disable interruption. When disabled, the user can only respond once the AI has finished speaking.

Click “Save Changes” to apply your transcription settings.

Text-to-Speech Settings

Configure the text-to-speech provider and model used to turn text generated by the backend into speech spoken by the AI. The default provider and model are optimized for low-latency English language use cases. For multi-language support, there are specialized text-to-speech models we support.

We recommend experimenting with different providers, models and voices as they all have varying characteristics.

SettingDescriptionOptions / Notes
ProviderThe text-to-speech provider used to generate AI speech.See available text-to-speech providers
ModelSelect the TTS model that matches your quality and speed needs. The default model is often best for English language use cases.See available text-to-speech models
VoiceChoose the voice that best represents your AI.Select from available voices in the chosen provider/model. We recommend experimenting with different voices to find the right one for your use case.

Click “Save Changes” to apply your text-to-speech settings.

Backend Configuration

The Backend receives the transcribed user’s speech, and is responsible for generating the voice AI’s response. Layercode offers a hosted backend or the ability to connect your own backend with a simple webhook.

Get started immediately with Layercode’s optimized backend powered by Gemini Flash 2.0. Our hosted backend provides:

  • Ultra-low latency responses
  • Optimized for real-time conversation
  • Zero backend setup required
SettingDescription
LLM PromptConfigure the personality and behavior of your AI assistant.
Welcome MessageConfigure the message your AI will speak when the conversation first starts. If disabled, the user starts the conversation.
Remember to set up your API keys in your chosen backend environment.

Testing Your Pipeline

Click the “Try it out” button on your pipeline to visit the Playground.

The Playground is a pre-built frontend voice UI for testing out your voice pipeline. If you decide to connect your own backend, this is a great place to test it out. Even if you build your own frontend voice UI, the Playground will still work as a direct way to test your pipeline.

Next Steps

Congratulations! You’ve created your first voice pipeline. Now you can integrate it into your application.

If you are building web or mobile based voice experience, follow our guide below. You can also choose to connect your backend to your pipeline to control the AI’s response (instead of using our hosted backend). This gives you complete control over the AI’s response and allows you to use your own LLM provider and agent libraries.