Abstract

Proposed herein is a unified, graphics processing unit (GPU) -efficient architecture for real-time and post-meeting speech intelligence that simultaneously handles transcription, translation, summarization, and action-item extraction without redundant computation. By tightly coupling a shared audio encoder, streaming multimodal caching, blank frame skipping, and persistent storage, the system drastically reduces latency and cost, enabling scalable, low-latency meeting analysis across multiple tasks, even for long-form audio sessions.

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.

Share

COinS