I’m always looking for new functions and ways to use different features in visual assistance applications, and first downloaded the Envision app a few years ago when my friend was raving about its text recognition features at a low vision accessibility event we were both attending. When Envision became free for all users in 2022, I began recommending Envision and Envision AI more often for users looking for a visual assistance app that would work on multiple platforms and support helpful features like file upload, multiple languages, and recognizing faces. Here are my tips for using the Envision app with low vision for a variety of tasks.
Overview of Envision/Envision AI
The Envision app (also known as Envision AI) is a free visual assistance app that uses machine learning and artificial intelligence to recognize text and other types of visual information, including colors, products, people, and more. Envision is designed for blind and low vision users and offers multiple accessibility features, including self-voicing descriptions (no screen reader needed), large font sizes, dark mode, and other accessibility features that make it easier for people to get the information they need.
Unlike other visual assistance apps that are exclusive to one platform or operating system, Envision is available for both Android and iOS devices. On the Apple App Store, Envision is listed under the name Envision AI, but the features between the Android and iOS app are identical.
Related links
- Envision – Perceive Possibility (letsenvision.com)
- Envision AI on the App Store (apple.com)
- Envision – Apps on Google Play
How to use the Envision app
Creating an Envision account
Before using the Envision app for the first time, users are prompted to log in or create a free Envision account in order to use the app. New users will need to connect their Google or Apple account, or provide an email address and password to create an account.
Envision app interface
The default view for the Envision app is a live view of the device’s back camera with a series of buttons on the bottom of the screen that are labeled with various functions/features.
To activate a function, users position their device camera over an object, text, or other area of interest, and select the button of their choice to either read or listen to a visual description/transcribed text. Once the user is finished, they can select the button again or select the “back” button (if available) to return to the default view or choose another option.
Input options for Envision and Envision AI
There are a few different ways to input or upload content for Envision and Envision AI, including:
- Using the device camera in the Envision app. Only the back camera is supported, users cannot switch to the front-facing camera. However, the app does provide audio feedback for positioning items for accuracy
- Sharing an image from the device gallery by selecting the Share button and selecting the option for Envision. This generates an image description and transcribes any text in the image
- Importing an image or file from the Envision app using the Import File function. Saved files can be accessed from the Library function within the app and are stored locally on the device (not uploaded online)
- Using the Scan Text function to scan a copy of a document- single page and multi-page documents are supported
Envision app languages
The Envision App itself is available in over 30 different languages and local translations. However, the Envision app can recognize text in more than 90 different languages, including several indigenous languages, endangered languages, and even dead languages such as Latin and Ancient Greek. A full list of supported languages are linked below.
For text-based features, users can change the reading language for increased accuracy by selecting the Reading Language option at the top of the screen and choosing the language they want to use- this does not change the language of the Envision app itself.
Related links
- Mainstream Technology and Low Vision: Smartphones
- Languages Envision App Recognises – Envision (letsenvision.com)
Envision app functions and features
The Envision app and Envision AI app offer the following functions. Depending on customization settings, some features may be hidden under the More Features menu, though users can customize the display order of functions and rearrange items as needed.
Instant Text
Instant Text reads text that is visible within the view of the device camera. This can include environmental text like signs, flyers, room numbers, or shorter documents that don’t need to be saved for later. Both handwriting and typed text are supported.
Scan Text
Scan Text reads text that is visible in a photo taken by the user within the Envision app. This can include longer single page or multi-page documents like a class reading assignment, letter, or handwritten messages. Before transcribing text, users are prompted to take a picture of the document they want to read, and the Envision app then uses character recognition technology to identify text. Once the text has been identified, users can then read the text in a simplified display, listen to text read out loud, or export text by selecting the Share icon at the bottom of the screen.
Users can adjust the text layout detection in the Settings app.
Library
Library stores files that have been uploaded to the Envision app and are saved locally on the device, not uploaded online. Users can upload files from their device and read them in the Envision app, with the option to save the file to the Library for later access. If the Envision app is deleted, users will lose access to files uploaded to the Library.
Import File
Instead of using the device camera, users can import files from their device or import an image to receive a description or read text. Items can be saved to the Envision library for later use, and can also be translated using the translation option at the top of the screen.
Scan Barcode
Scan Barcode identifies product names and other details, and provides the option to view additional product options on Google.
Describe Scene
Describe Scene takes a photo of what’s in view of the device camera and provides a simple visual description of various objects in the image, as well as people if relevant. Images can also be saved to the device photo gallery.
Explore
Explore uses the live device camera to identify supported objects and people that are within the view of the device camera.
Find Objects
Find Objects can locate common items that are within view of the device camera, prompting users to select the object they are looking for. Some examples of objects include:
- Light switch
- Door handle
- Cat
- Bench
- Bottle
- Laptop
- Table
- Backpack
- Toaster
If Envision detects an object, the user will receive an audio cue that the object is within view of the device camera.
Find People
Find People determines if there are people nearby, or if a specific person is within view. If people are detected, the user will receive an audio cue, and the person’s name will be spoken if they have previously been added to the Envision app.
Teach Envision
Teach Envision provides users with the option to recognize a face by taking five photos of a person’s face and adding their name to the list of recognizable people.
It’s worth noting that this information is stored locally on the device, uploading a user’s face does not make their name and likeness available to anyone who uses the application. So if I trained Envision to recognize my professor, my friend’s Envision app would not automatically recognize my professor unless my friend also used the Teach Envision function for them.
Detect Colors
Detect Colors tells a user what color an object is by using a sensor in the center of the screen. For items that have multiple colors, only the color that is in the center of the device screen will be identified- users will need to move the device around to get additional color information.
Within the Settings menu, users can select between Standard and Descriptive color options. Descriptive offers additional color shade information and color names.
Scan Accessible QR
Scan Accessible QR is designed for scanning QR codes that are specifically designed for accessibility- accessible QR codes have a border around them and can be detected from far away. Some products have accessible QR codes on their packaging, and Envision can identify the names of these products or other information as needed.
Related links
- Reading Handwriting With Assistive Technology
- File Formats For Low Vision and Print Disabilities
- How To Access Images Without Alt Text
- Assistive Technology For Medication Labels
Ways that I use Envision and Envision AI
Here are some examples of ways I have used Envision on my Android phone and Envision AI on my iPad in various environments and contexts
- Recognizing text in another language, with the option for translation or using a simplified reading display. This has been amazing for recognizing uncommon languages!
- Reading a handwritten note from a friend or something left by a family member
- Reading signs or maps, which can include informational or interpretive signage (e.g signs at a museum)
- Browsing a menu or list of options at a restaurant
- Checking store hours posted on a door
- Scanning a long text-only document and reading information in a simplified display
- Spotting potential obstacles before I sit on them- this was especially helpful for spotting a friend’s black cat on a black couch
Related links
- How I Read Research Sources With Assistive Technology
- Reading Handwriting With Assistive Technology
- Simplified Reading Displays and Low Vision
- How I Use My Phone As Assistive Technology In Class
Ways that I don’t use Envision and Envision AI
Envision isn’t the only visual assistance application I use, and there are a few tasks where Envision isn’t particularly helpful and I prefer to use something else. This can include:
- Scanning barcodes for product identification. For some reason, Envision has trouble recognizing several food items in my pantry and personal care products- maybe it’s the packaging? I use Google Lens for identifying items usually.
- Identifying objects via visual search. If I’m in a museum and want to know more about something on display, or want more descriptive information about a plant or other object, Google Lens is again a better option for visual searches. Envision AI specializes in shorter descriptions.
- Identifying multiple objects. While I do use the Find Objects feature if I am looking for something specific (like a cat), I use Google Lookout or Seeing AI if I’m not quite sure what I am looking for or need to find multiple items.
- Scanning text that also includes a lot of images. Since Envision doesn’t recognize images, this can lead to errors with how text is read or mean that text isn’t recognized at all. I scan documents using Microsoft Lens and can use the OCR feature to recognize text if needed.
- Pairing with smart glasses. Envision offers smart glasses that can be used for a variety of visual tasks, but this isn’t something I’m personally interested in.
Related links
- Google Lens Review For Low Vision
- Google Lookout App For Low Vision
- Microsoft Seeing AI And Low Vision
- How I Use Microsoft Lens With Low Vision
More resources for using the Envision app (Envision AI) with low vision
- Users can support feedback to the development team by selecting the Settings menu, and then selecting Give Feedback in the Help section.
- The Envision YouTube channel also has several tutorials and informational videos available, and can be accessed at Envision – YouTube
- Want to learn more about assistive technology for low vision? Check out A to Z of Assistive Technology For Low Vision
- Want to learn more about visual assistance applications? Check out Visual Assistance Apps: Post Round Up