iOS application for finding formants in spoken sounds
-
Updated
May 31, 2025 - Swift
iOS application for finding formants in spoken sounds
Spokestack: give your iOS app a voice interface!
OpenAI API wrapper for Swift
Lightweight Swift library for log-Mel spectrogram extraction with Accelerate & CoreML)
OtosakuStreamingASR-iOS is a real-time speech recognition engine for iOS, built with Swift and Core ML. It uses a fast and lightweight streaming Conformer model optimized for on-device inference. Designed for developers who need efficient audio transcription on mobile.
OtosakuStreamingASR-iOS offers a simple way to integrate real-time speech recognition into your iOS apps. With its efficient on-device processing, you can enhance user experiences without relying on internet connectivity. 🐙✨
An example project showing how we can use Apple Speech to Text cloud service and AWS Machine Learning to process and find meaning in text.
OtosakuFeatureExtractor is a Swift library that simplifies the process of converting raw audio into log-Mel spectrograms for machine learning applications. This tool streamlines audio processing, making it ideal for on-device inference. 🐙📦
Add a description, image, and links to the speech-processing topic page so that developers can more easily learn about it.
To associate your repository with the speech-processing topic, visit your repo's landing page and select "manage topics."