What makes Smart phones easier to use

 Smartphones have unhurriedly evolved to be more beneficial for people with a range of physical abilities, thanks to tools like screen readers and adjustable text sizes.
With the recent release of Apple’s iOS 16 and Google’s Android 13 software, accessibility features have been introduced or upgraded, including improved live transcription tools and apps that utilize artificial intelligence to identify objects. When enabled, your phone can send you a visual alert when a baby is crying, for example, or a sound alert when you’re approaching a door.
Also, variety of both new and classic accessibility technologies make using phone simpler for everyone. 
Here is tour.





TO BEGIN WITH.
To access all of the tools and features available on an iOS or Android phone, open the Settings app and click Accessibility. 
Take the time to investigate and experiment.


Both Apple and Google offer dedicated Accessibility sections on their websites for further reference, however, keep in mind that your precise features may differ depending on your software version and phone type.

Alternative Navigation.

Swiping and tapping by hand to navigate a phone’s features doesn’t work for everyone, but iOS and Android provide several ways to move through the screens and menus, including quick-tap shortcuts and gestures to perform tasks. These controls (like Apple’s Assistive Touch tools and its Back Tap function, which completes assigned actions when you tap the back of the phone) are in the iOS Touch settings.

Android’s accessibility shortcuts offer similar options. One way to access these is by opening the main Settings icon, selecting System, then Gestures and System Navigation.

Both platforms support navigation through third-party adaptive devices like Bluetooth controllers or by using the camera to recognize facial expressions assigned to actions, like looking to the left to swipe left. These devices and actions can be configured in the iOS Switch Control and Head Tracking settings, or in Google’s Camera Switches and Project Activate apps for Android.

Apple and Google provide several tools for those who can’t see the screen. Apple’s iOS software offers the VoiceOver feature, and Android has a similar tool called TalkBack, which provides audio descriptions of what’s on your screen (like your battery level) as you move your finger around.

Turning on the iOS Voice Control or Android Voice Access option lets you control the phone with spoken commands. Enabling the iOS Spoken Content or Android’s Select to Speak setting has the phone read aloud what’s on the screen — and can be helpful for audio-based proofreading.

Don’t forget a few classic methods of hands-free interaction with your phone. Apple’s Siri and Google Assistant can open apps and perform actions with spoken commands. And Apple’s Dictation feature (in the iOS Keyboard settings) and Google’s Voice Typing function lets you write text by speaking.

Visual Assistance

In their Accessibility settings, iOS and Android include shortcuts to zoom in on sections of the phone screen. But if you’d generally like bigger, bolder text and other display adjustments, open the Settings icon, choose Accessibility, and select Display & Text Size. In Android, go to Settings, then Accessibility, and choose Display Size and Text.

The Magnifier app, Apple’s digital magnifying glass for enlarging objects in the camera’s view, has been upgraded in iOS 16. The app’s new functions are designed to help people who are blind or have low vision use their iPhones to detect doors and people nearby, as well as to identify and describe objects and surroundings.

Magnifier results are spoken aloud or displayed in large type on the iPhone’s screen. The door-and-people detection uses the device’s LiDAR (light detection and ranging) scanner to calculate distance and requires an iPhone 12 or later.

To set up your preferences, open the Magnifier app and select the Settings icon on the lower left corner; if you can’t find the app on your phone, it’s a free download in the App Store. The Magnifier is just one of many vision tools in iOS, and the company’s site has a guide to setting up the app on the iPhone and iPad.

Google’s recently updated Lookout assisted-vision app (a free download in the Play store) can identify currency, text, food labels, objects, and more. Google introduced Lookout in 2018, and it works on Android 6 and later.

Both platforms offer controls to amplify speech around you through your headphones. In iOS, go to the Audio/Visual section for Headphone Accommodations, In Android, visit the Sound Amplifier setting.

With the iOS 16 update, Apple includes Live Captions, a real-time transcription feature that converts audible dialogue around you into text on-screen. Android’s Accessibility toolbox includes the Live Caption setting that automatically captions videos, podcasts, video calls, and other audio media playing on your phone.

Google’s free Live Transcribe & Notification Android app converts nearby speech to on-screen text, and can also provide visual alerts when the phone recognizes sounds like doorbells or smoke alarms. The Sound Recognition tool in the iPhone’s Hearing section of the Accessibility settings does the same. And check your phone’s settings for multisensory notifications, like LED flash alerts or vibrating alerts, so you don’t miss a thing.



1 comment:

Powered by Blogger.