📱Multimodal journey
From start to finish, set up a multimodal journey in minutes with NLX's Dialog Studio
Checklist
You'll complete the following to successfully launch your multimodal experience:
You may add the following to an existing voice assistant in your workspace, if desired.
Pre-setup: Integrations
Est. time to complete: ~5 minutes
A one-time integration of a Natural Language Processing (NLP) engine must be completed in your workspace:
A one-time integration of a voice-enabled communication channel must be completed in your workspace:
A one-time set up of an Action that sends an SMS:
Be sure your Action has the following properties defined in the Request model schema:
Message (string)
PhoneNumber (string)
URL (string)
Step 1: Create a Journey
Est. time to complete: ~10 minutes
Journeys pair voice prompts from an AI assistant to a digital asset (website, mobile app, etc.) to help guide users through a self-service task. Begin by identifying the elements from your digital asset that need to be mapped to a voice line (pages, buttons, etc.)
Select Journeys in your workspace > Click New journey option > Name your Journey
Click Save
Click + Add step > Enter the bot's voice line in the message field*
Repeat for each step
On the final step of your journey, enable the Action toggle
End: Terminates the phone call after the AI assistant delivers the voice step
Continue: Proceeds from the Continue edge in your multimodal flow (see Step 2)
Click Save
Download your steps to a .csv or .json file using the Download link
🧠 Looking for more? See Add + download steps
Step 2: Create multimodal flow
Est. time to complete: ~5 minutes
As all multimodal journeys begin with a traditional voice experience (IVR), you'll construct an intent workflow that sends the SMS containing a link to your digital asset and initiates your journey (multimodal mode):
Select Intents in your workspace menu > Create a new intent and set the flow to a voice channel you integrated in your workspace
Add training phrases to your AI assistant to match the user's intent, or enable Skip training setting on the intent's Settings tab if your bot only automates this task
After adding a greeting with a Basic node to the Canvas, place an Action node that will send the SMS link (requires an API that sends SMS be previously set up under Actions in your workspace). At a minimum, be sure your Action's Request model has at least two properties 1) Phone number, 2) URL
Within your intent flow, select the Action node > Use the system variable
{system.userId}
for the Phone number field of your payload on the Action node's side panelIn your Action's URL payload field, enter your URL followed by the query parameter
?cid={system.conversationId}
Place and link a Basic node after the Action to indicate a text was successfully sent to the user
From the Basic node, place and link to a Multimodal node > Click Save
🧠 Looking for more? See Construct flows
Step 3: Deploy & install SDK
Est. time to complete: ~10 minutes
You'll need to install NLX's multimodal SDK to each screen of your digital asset so applicable API calls can be made to trigger voice lines where you've defined them:
Select Journeys in workspace menu > Choose your journey > Click Deployment tab of your journey
Choose Review & build > Click Create build
After a successful build, select Deploy from the Production column > Click Create deployment
Select Details link next to the Deployed status > Under Setup instructions, click Open journey configurator
API key: You may auto-generate an API key under the journey's Settings tab, Save, and then enter it in the configurator's field
Conversation ID: Dynamically generated for each conversation session with a user by Dialog Studio, you may parse the ID from the user's URL path. Sample code: https://developer.mozilla.org/en-US/docs/Web/API/URLSearchParams/get#examples
Install code snippet with applicable step IDs (downloaded in Step 1) to each page of your frontend
To avoid CORS errors, make sure to whitelist your URL domain(s) in the journey's Settings.
🧠 Looking for more? See Developer SDK
Step 4: Set up bot & deploy
Est. time to complete: ~5 minutes
Now you'll create the AI assistant users will interface with. This step involves attaching all intent workflows you want your bot to access, defining flows to handle certain behaviors, setting up the voice channel your bot supports, and deploying your conversational AI assistant!
Select Bots from workspace menu > Choose New bot
Enter a descriptive name > Click Save
Click Intents tab of bot > Select + Add intents > Attach the intent containing the multimodal workflow from Step 3
Select Default behaviors, assign an intent to the Welcome behavior (if you intend for your multimodal flow to be the only intent your bot handles, then assign to the Welcome behavior)
Select Channels tab of bot > Expand the voice channel your bot will support (e.g., Amazon Connect, Amazon Chime SDK, etc.). Be sure this is the same channel set on your intent flow(s) attached to the bot, or add the channel to your intent flows > Click + Create channel
Enter required details for voice > Click Create channel > Click Save
Click Deployment tab of bot > Select Create or Review & build
Wait for validation to complete > Select Create build*
When satisfied with a successful build, click Deploy
🧠 Looking for more? See Manage channels
🔧 Need to troubleshoot your build? See how to troubleshoot build errors
Last updated