9 min read
9 min read

Google just made that possible to find out. With new updates to Android’s TalkBack feature, blind and low-vision users can now ask Gemini what’s in an image, even when it doesn’t have a caption.
It’s more than just a description. You can follow up with questions like, “What color is it?” or “What kind of object is that?” It’s a huge step for people who’ve had to rely on vague image labels. Now, your phone gives you the full picture literally, with smart AI-powered feedback.

Instead of guessing what’s shown in an app or webpage, you can ask Gemini directly. It’ll scan your whole screen and explain what it sees, then respond to your questions.
For example, if you’re looking at an online store, you might ask, “Are there any discounts?” or “Is this item in stock?” That kind of instant help saves time and cuts confusion. It’s a big win for accessibility, but it’s also a smarter way for anyone to interact with their device more clearly and confidently.

Google’s new Expressive Captions show more than just spoken words; they show how those words are said. If someone says “nooooo,” the captions stretch the letters to match. This makes videos, phone calls, and live events easier to follow emotionally.
Tone matters in conversations, and now the captions give you clues about excitement, sarcasm, hesitation, or surprise. For people who rely on captions, this change adds a layer of communication that was missing before. It turns speech into a more complete experience on screen.

Not everything people need to understand comes from voices. That’s why Google has added labels for background sounds like coughing, laughing, or even someone clearing their throat. These updates help fill in the gaps for people who can’t hear those subtle sounds.
You’ll see “whistling” or “applause” pop up during shows, live streams, or video calls. These tiny
details can help clarify what’s going on in a scene or conversation. It’s all about making sure people don’t miss a thing, even the quiet parts.

Until now, scanned PDFs were like locked boxes, images that looked like text, but weren’t readable. Google has changed that in Chrome by adding Optical Character Recognition, or OCR. This means your computer can now recognize the words inside those scanned files.
Screen readers can speak them out loud, and you can search, copy, or highlight text like normal. It’s a small change with a huge impact, especially for students, workers, and anyone who handles documents online. Reading scanned material just became a whole lot easier.

It’s frustrating when zooming into a website makes everything shift around or overlap. Now, Chrome on Android has a new Page Zoom feature that lets you increase just the text size, without breaking the page layout.
You choose how much to zoom and even decide if you want that zoom level on every site or just a few. Everything else on the page stays in place. This makes reading more comfortable while keeping websites looking clean and usable. It’s great for tired eyes or small screens.

Someone sends you a picture, and now, you don’t have to wonder what it is. TalkBack with Gemini can tell you what’s in that image, whether it’s a sunset, a selfie, or a slice of pizza.
But it doesn’t stop there, you can ask questions about the photo, like, “Is there a dog in this picture?” or “What color is the dress?” That’s a game changer for people using screen readers. Texts, social media posts, and shared images suddenly have a voice of their own, thanks to AI.

Google is working with the College Board to bring accessibility tools into the Bluebook testing app. That means students taking the SAT or AP exams on Chromebooks now have more support built right in.
They’ll be able to use features like ChromeVox screen reader, Dictation, and text simplification tools during the test. It’s not just about legal compliance, it’s about giving every student a fair shot.

For students who need extra help, Google is building on its Face Control and Reading Mode features. Face Control lets users move through screens using just head movements, no touch needed.
Reading Mode makes text easier to follow by changing colors, fonts, and spacing. Together, these tools help kids with physical or learning differences stay engaged and confident. Teachers and students both benefit when the tools match the needs of the learner, not the other way around.

Gemini isn’t limited to just image descriptions anymore. Now it can analyze whatever app or screen you’re on and give you answers in real time. You could ask, “What does this button do?” or “Is this page safe to enter my info?” That means less guessing and fewer mistakes.
It’s especially helpful when navigating unfamiliar apps or websites. Think of it like a tech-savvy assistant guiding you as you go, making smartphones feel a little more human.

Speech recognition doesn’t work the same for everyone. That’s why Google’s Project Euphonia is sharing its research tools on GitHub, letting developers build smarter, more inclusive voice tech.
These tools are designed to better understand people with speech differences or neurological conditions. By giving developers access, Google is helping to improve apps that work for more people. It’s about building tools that listen better, not just louder.

Google is expanding its AI speech recognition in more languages. That’s good news for users who speak languages other than English or who speak with regional accents.
Apps and services will better understand what’s being said and respond more accurately. This creates a smoother experience for a broader group of users. It’s not just helpful, it’s respectful of the way people naturally speak around the world.

Everyone has different comfort levels when it comes to screen size. Now, Chrome lets you save custom zoom settings for each website you visit.
So if you like big text on a news site but smaller font on your dashboard, Chrome remembers your choice. It’s one of those small conveniences that make everyday browsing less of a chore. Once you set it, it just works, no need to adjust every time.

Gemini isn’t just descriptive, it’s helpful. Screen readers powered by Gemini can summarize the content and answer questions, not just read line by line.
That’s a big step forward for people using assistive tech. Instead of slowly going through every item on the screen, users can jump right to what they need. It’s about speed, context, and clarity, all powered by smart AI built into Android.

Google gives you more control over how captions look and act. You can change their size, how long they stay on screen, and even the way they’re styled.
That means you can tailor your captions to match your reading speed and visual needs. From watching fast-paced sports to relaxing with a movie, your captions now keep pace with you. Accessibility is more than just adding a feature; it’s about making it fit you.

For users who can’t see or touch the screen, Gemini and TalkBack work together to provide full access. Ask what’s on the screen, what options are available, or what action a button will take.
This makes it easier to use phones without needing perfect vision or motor skills. Instead of tapping around blindly, users get helpful, spoken feedback on what’s going on. It’s another step toward phones that work for all people, not just some.
Curious what else Google’s been up to? You might want to see why it could be heading for a major shake-up.

These features aren’t just nice-to-have, they’re necessary. Google is proving that accessibility should be baked into design, not added later.
From screen readers to speech recognition and zoom tools, these updates help people interact with their devices on their terms. It’s about fairness, independence, and comfort for users of all kinds. And with more improvements likely on the way, this is just the start of a more inclusive tech future.
Want to see how else Google is looking out for users? Check out how it’s helping prevent accidental purchases.
What do you think of Google’s new accessibility features? Drop a comment below and give us a like if you’re all for more inclusive tech.
Read More From This Brand:
Don’t forget to follow us for more exclusive content right here on MSN.
This slideshow was made with AI assistance and human editing.
This content is exclusive for our subscribers.
Get instant FREE access to ALL of our articles.
Dan Mitchell has been in the computer industry for more than 25 years, getting started with computers at age 7 on an Apple II.
We appreciate you taking the time to share your feedback about this page with us.
Whether it's praise for something good, or ideas to improve something that
isn't quite right, we're excited to hear from you.
Stay up to date on all the latest tech, computing and smarter living. 100% FREE
Unsubscribe at any time. We hate spam too, don't worry.

Lucky you! This thread is empty,
which means you've got dibs on the first comment.
Go for it!