Apple training Siri to detect and adapt to users who stutter using research and podcast catalog


    Apple is investigating ways to enhance its Siri voice assistant for users with atypical speech patterns, the company confirms to the Wall Street Journal. According to the report, Apple is leveraging its podcast library for speech samples that could train Siri to adapt to users who speak with a stutter.

    Here’s the news from the Wall Street Journal report:

    The company is now researching how to automatically detect if someone speaks with a stutter, and has built a bank of 28,000 audio clips from podcasts featuring stuttering to help do so, according to a research paper due to be published by Apple employees this week that was seen by the Wall Street Journal.

    For now, Apple relies on its Hold to Talk feature as a method for interacting with Siri without the voice assistant cutting off users with slower speech patterns than its tuned for, but physically interacting with a device isn’t always convenient.

    Siri can be voice activated on iPhones, iPads, and Macs, and especially HomePod and HomePod mini, using the “Hey Siri” voice command followed by a request. For users who stutter, however, the current version of Siri commonly interprets pauses in speech as the end of a voice command. In turn, this prevents the voice assistant from reaching its full potential for a collection of customers.

    Friend of the site Steve Aquino pointed to the Apple research paper referenced in the WSJ report.

    Here’s the abstract for Apple’s research:

    The ability to automatically detect stuttering events in speech could help speech pathologists track an individual’s fluency over time or help improve speech recognition systems for people with atypical speech patterns. Despite increasing interest in this area, existing public datasets are too small to build generalizable dysfluency detection systems and lack sufficient annotations. In this work, we introduce Stuttering Events in Podcasts (SEP-28k), a dataset containing over 28k clips labeled with five event types including blocks, prolongations, sound repetitions, word repetitions, and interjections. Audio comes from public podcasts largely consisting of people who stutter interviewing other people who stutter. We benchmark a set of acoustic models on SEP-28k and the public FluencyBank dataset and highlight how simply increasing the amount of training data improves relative detection performance by 28% and 24% F1 on each. Annotations from over 32k clips across both datasets will be publicly released.

    The research paper acknowledged that the current approach to tuning Siri for dysfluency is one approach, there remains an opportunity to improve the effort using language models and other methods.

    Lastly, Apple concludes that while its current research focuses on users who stutter, future research should explore other categories like dysarthria that have different characteristics.

    FTC: We use income earning auto affiliate links. More.


    Check out 9to5Mac on YouTube for more Apple news:



    Source link

    Previous articleWe haven’t seen Dell XPS 13 deals this good since Black Friday — up to $670 off
    Next articleXbox Live is experiencing outages, users may experience problems signing in