Apple appears to have emerged as a pioneer pointing in the direction of how AI and machine learning may truly be utilized for the greater good in the face of competitive advances in generative AI models like ChatGPT, prompting anxieties about how AI will affect our near future.
Apple unveiled a number of new improvements for the iPhone and iPad’s cognitive accessibility on May 16. These features are intended to assist those who have speech, hearing, or vision impairments. Apple has always had its “Accessibility” function to aid those with special needs in using a phone, and new upgrades are only anticipated to increase everyone’s access to iPhones and iPads.
Two of the aforementioned characteristics have drawn a lot of interest. Apple’s launch of Personal Voice and Live Speech shows off its capabilities in artificial intelligence and machine learning.
What do the features of personal voice and live speech mean?
By reading aloud a set of example prompts for 15 minutes, Apple will enable iPhone and iPad users to develop their own “personal voice, or a duplicate of their genuine voice.
On their Apple devices, users can prerecord preset words like “I would like a black coffee,” which can be played during real-time Facetime and phone chats.
Additionally, Apple’s Personal Voice and Live Speech integration allows users to type what they want to say and have their software-generated Personal Voice read it aloud to anyone they are chatting with via Facetime or phone calls.
Apple is essentially cloning the user’s voice to be used during real-time calls to make it look as though the user is speaking with “personal voice” and “live speech.
These functions are intended for those who have speech difficulties, such as Amyotrophic Lateral Sclerosis (ALS), or who are about to lose their capacity to speak or have already lost it.
Apple also revealed a host of capabilities for people with limited vision or blindness. The Magnifier mode’s Point and Speak feature enables users to point at an object and hear the text spoken out.
How about safety and confidentiality?
There is always a concern about safety and privacy when it comes to technological breakthroughs like ChatGPT, AI assistance, and voice or image copying. The voice feature on Apple is no different.
According to reports, family members’ voices have been cloned using artificial intelligence to defraud others, particularly the elderly or those with little technological expertise.
Although audio deepfakes are also appearing, video deepfakes have been encountered most frequently.
Consider receiving a call from an unknown number that is actually your close friend or sibling, requesting that you send urgent money into an unidentified account because the person is in danger. In order to prevent the victim of the fraud from double-checking the account or number, urgency is generated.
According to Apple, the privacy of personal voice and live speech will be protected by on-device machine learning.