Process hours of audio in seconds using efficient chunking, distributed inference, and optimized GPU resources.
asyncio
for concurrent execution.Transcribe
(Entrypoint Chainlet)MacroChunkWorker
(Processing Chainlet)WhisperModel
(Inference Model)micro_chunk_size_sec
→ Balance GPU utilization and inference latency.macro_chunk_size_sec
→ Adjust chunk size for optimal parallelism.WHISPER_URL
in transcribe.py
.