Audio Processing Pipelines
Indexify is capable of ingesting and processing audios. You can build pipelines that perform one or more of these functions -
Each of these examples are organized as -
- Extraction Graph/Pipeline Description
- Pipeline Setup Script
- Upload audio files, and retrieve the artifacts of the pipeline such as transcripts, summary, topics, etc.
Transcription Models
Indexify provides the following transcription extractors out of the box -
- OpenAI Whisper
- Apple's Whisper MLX(MacOS)
- Groq Whisper API
- DistilWhisper x PyAnnotate based speaker diarization
You can list all the ASR extractors by running the command indexify-extractor list
There are a ton of other ASR APIs out there, which we haven't integrated with. We would love the community to contribute or you can write a custom extractor to call into any API.
Indexing
The indexing example uses LanceDB by default, but you can swap it out with any other VectorDB that Indexify supports - Qdrant or PgVector.
We are using the embedding model - minilm-l6 in this example, you can list the other available extractor by running the command - indexify-extractor list