Skip to main content
Interested in building a real-time intelligence or clinical decision support solution? Look no further than the Corti AI Stream API: Real-time, bidirectional WebSocket Secure (WSS) communication with the Corti AI platform for live transcript generation and clinical fact extraction (FactsR™).

Delivered as a modular API, /stream provides a stateful interaction framework for developers to embed clinical-grade intelligence directly into their healthcare applications, creating safer, leaner, and more trusted AI experiences at the point of care.
Depending on your use case, the stream endpoint can be used for real-time conversational transcript, fact extraction, or both! See the full API specification here.

Using the API

1

Initiate a /stream by creating an Interaction, which will return a web socket URL will be returned along with the interactionId.
2

Connect to the web socket and set your configuration:
ParameterDescription
primaryLanguageSpoken language to be transcribed
diarizationEnable speaker separation (most effective with multichannel audio)
multichannelEnable multichannel audio (be sure to define channels)
participantsAssign speaker roles for audio channels and diarization
modeDefine facts or transcript depending on the desired real-time output
outputLocaleOutput language for extracted facts (required with “mode”:“facts”)
3

Once config is accepted, begin sending audio packets.
4

Receive transcripts every ~3 seconds and facts every ~60 seconds (standard response times can adapted for custom response times).
5

Send the end message to close the audio stream.

Features

Click on the cards to learn more…

Languages

Corti speech recognition is specifically designed for use in the healthcare domain. A tier system has been introduced to categorize functionality and performance that is available per language and endpoint. Languages in the Enhanced and Premier tiers have the utmost functionality and recognition accuracy - they’re the ones recommended for dictation use.

Vocabulary

coming soon
Unparalleled access to and control over the vocabulary used by the speech recognition models will give organizations the utmost control over the dictation experience: Gain visibility into the terminologies the models are trained on and update the vocabulary as needed to optimize for localized or specialized needs, respond to reported issues, and stay ahead of the wave of changes to medical practices and communication.

Please contact us for more information or help.