/plushcap/analysis/assemblyai/integrate-stt-into-llamaindexts

How to integrate spoken audio into LlamaIndex.TS using AssemblyAI

What's this blog post about?

In this tutorial, you will learn how to create an application that can answer questions about an audio file using LlamaIndex.TS and its new integration with AssemblyAI's speech-to-text APIs. LlamaIndex is a flexible data framework for connecting custom data sources to large language models (LLMs). However, LLMs only operate on textual data and do not understand what is said in audio files. With the recent contribution of an integration with AssemblyAI to LlamaIndex.TS, developers can now transcribe audio files to text within LlamaIndex. The tutorial guides you through setting up a TypeScript Node.js project, configuring environment variables for OpenAI and AssemblyAI API keys, using LlamaIndex.TS to build a Q&A application, transcribing an audio file with the AudioTranscriptReader provided by AssemblyAI's integration, and querying the index for answers. Finally, you can explore LeMUR in the Playground using the link provided.

Company
AssemblyAI

Date published
Dec. 6, 2023

Author(s)
Niels Swimberghe

Word count
940

Hacker News points
None found.

Language
English


By Matt Makai. 2021-2024.