US20150026587A1 - System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer - Google Patents

System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer Download PDF

Info

Publication number
US20150026587A1
US20150026587A1 US14/505,900 US201414505900A US2015026587A1 US 20150026587 A1 US20150026587 A1 US 20150026587A1 US 201414505900 A US201414505900 A US 201414505900A US 2015026587 A1 US2015026587 A1 US 2015026587A1
Authority
US
United States
Prior art keywords
vision test
display
objects
user interface
display settings
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/505,900
Inventor
Arthur Austin Ollivierre
Robert Michael DiNapoli
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US14/505,900 priority Critical patent/US20150026587A1/en
Publication of US20150026587A1 publication Critical patent/US20150026587A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/02Subjective types, i.e. testing apparatus requiring the active assistance of the patient
    • A61B3/028Subjective types, i.e. testing apparatus requiring the active assistance of the patient for testing visual acuity; for determination of refraction, e.g. phoropters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0489Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using dedicated keyboard keys or combinations thereof
    • G06F3/04897Special input arrangements or commands for improving display capability
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/0016Operational features thereof
    • A61B3/0041Operational features thereof characterised by display arrangements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/02Subjective types, i.e. testing apparatus requiring the active assistance of the patient
    • A61B3/028Subjective types, i.e. testing apparatus requiring the active assistance of the patient for testing visual acuity; for determination of refraction, e.g. phoropters
    • A61B3/032Devices for presenting test symbols or characters, e.g. test chart projectors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04806Zoom, i.e. interaction techniques or interactors for controlling the zooming operation

Definitions

  • the disclosed embodiments relate generally to displaying objects in a graphical user interface based on a visual acuity of a viewer.
  • a user must click on the “Start” button, move the cursor to the “Settings” menu item, select “Control Panel”, double click on the “Display Settings” icon, click on the “Settings” tab of the “Display Properties” window, and slide a resolution slider to select a desired resolution level.
  • the user may not be able to navigate through a series of screens or graphical user interfaces to change the display settings.
  • FIG. 1A is a block diagram illustrating a device, according to some embodiments.
  • FIG. 1B illustrates an example method to customize display settings of a display screen of an electronic device (e.g., during a boot-up sequence of the device), according to some embodiments.
  • FIG. 2 is a flowchart of a method for displaying objects in a graphical user interface of a device, according to some embodiments.
  • FIG. 3 is a flowchart of a method for presenting a vision test in a graphical user interface of a device, according to some embodiments.
  • FIG. 4 is a flowchart of a method for displaying vision test objects in a graphical user interface of a device, according to some embodiments.
  • FIG. 5 is a flowchart of a method for identifying display settings corresponding to input received from a user, according to some embodiments.
  • FIG. 6 is a flowchart of a method for displaying objects in a graphical user interface based on the display settings, according to some embodiments.
  • FIG. 7 is a flowchart of a method for generating additional viewing pages to display objects not displayed in a viewable area of a graphical user interface, according to some embodiments.
  • FIG. 8 is a flowchart of a method for restoring display settings at the end of a login session, according to some embodiments.
  • FIG. 9 is a flowchart of a method for retaining display settings between login sessions, according to some embodiments.
  • FIG. 10 illustrates an example vision test, according to some embodiments.
  • FIG. 11 illustrates another example vision test, according to some embodiments.
  • FIG. 12A illustrates objects displayed in an example graphical user interface, according to some embodiments.
  • FIG. 12B illustrates the objects displayed in the example graphical user interface of FIG. 12A after display settings for the graphical user interface have been modified, according to some embodiments.
  • FIG. 12C illustrates the objects displayed in the example graphical user interface of FIG. 12A after display settings for the graphical user interface have been modified, according to some embodiments.
  • FIG. 13 is a block diagram illustrating an example machine for performing the methodologies described herein, according to some embodiments.
  • the example embodiments described herein provide techniques for displaying objects in a graphical user interface based on a visual acuity of a viewer.
  • a vision test is presented in the graphical user interface of the device, where the vision test includes vision test objects used to test a visual acuity level of a user of the device.
  • Input is then received from the user identifying the user's ability to see at least one vision test object in the vision test.
  • display settings corresponding to the input received from the user are identified.
  • Objects are then displayed in the graphical user interface based on the display settings.
  • FIG. 1A is a block diagram illustrating a device 102 , according to some embodiments.
  • the device 102 may include any electronic device coupled to a display device (e.g., a monitor, a television) including, but is not limited to, a desktop computer system, a laptop computer system, a server, a mobile phone, a smart phone, a personal digital assistant (PDA), a gaming console, a portable gaming console, a set top box, a camera, a printer, a television set, or the like.
  • a display device e.g., a monitor, a television
  • PDA personal digital assistant
  • the device 102 includes a vision testing module 104 , a user interface module 106 , and a database 108 .
  • the vision testing module 104 is configured to determine a visual acuity of a user of the device 102 .
  • the visual acuity of the user is the ability of the user of the device 102 to see objects displayed on a display device (e.g., a monitor) of the device 102 .
  • the visual acuity of the user may be determined by an optometrist or an ophthalmologist using a Snellen Test.
  • the Snellen test includes block letters of varying sizes displayed at a predetermined distance (e.g., 20 feet from the user).
  • the ability of a user to read a particular line of letters indicates the visual acuity of the user.
  • other factors may influence the ability of the user to see objects displayed on the display device of the device 102 .
  • the dimensions and the resolution of the display device may affect the ability of the user to see objects displayed on the display device.
  • the user may be able to see objects displayed on a large display at a low resolution because the objects may be large enough to be seen by the user.
  • the user may not be able to see objects displayed on a large display at a high resolution because the individual objects may be too small to be seen by the user.
  • the user may not be able to see objects displayed on a small display at a high resolution because the objects may be too small to be seen by the user.
  • the user may be able to see objects displayed on a small display at a low resolution because the objects may be large enough to be seen by the user.
  • the user interface module 106 is configured to display objects in a graphical user interface of a display device of the device 102 based on display settings.
  • the database 108 stores vision test objects, reference visual acuity levels, reference display settings, predetermined viewing distances of display devices (e.g., based dimensions and resolution of the display devices), scaling factors for objects corresponding to the reference visual acuity levels, and user profiles for users of the device 102 that include a visual acuity level and/or display settings for the user (or a plurality of users) with respect to the device 102 .
  • FIG. 1B illustrates an example method to customize display settings of a display screen of an electronic device (e.g., during a boot-up sequence) of the device, according to some embodiments.
  • the vision testing module 104 is executed and a vision test is presented (operation 122 ) to the user.
  • the vision test is administered each time the device 102 is powered on.
  • the vision testing module 104 is executed and a vision test is administered to the user when requested by the user.
  • Display settings that are identified as a result of the vision test are applied to the device (operation 124 ). In an example embodiment, the display settings are applied to applications running on the device.
  • the method may be performed at any time before, during and after the device boots up.
  • the method allows a user who may have visual impairment to adjust or modify display setting on the devices even though the user is unable to read or adequately perceive images or text displayed on the display screen of the display device.
  • FIG. 2 is a flowchart of a method 200 for displaying objects in a graphical user interface of the device 102 , according to some embodiments.
  • the vision testing module 104 receives (operation 202 ) a request from a user of the device 102 to change display settings of the device 102 based on a vision test.
  • the request may include a voice-activated request, a predetermined key sequence, and/or a predetermined gesture in the graphical user interface of the device 102 .
  • operation 202 may be optional.
  • the vision testing module 104 may always present the vision test to the user of the device 102 if the device 102 is a publically-accessible device (e.g., a kiosk, a public computer system) or on the first login of the user on the device 102 .
  • a publically-accessible device e.g., a kiosk, a public computer system
  • the vision testing module 104 presents (operation 204 ) the vision test in the graphical user interface of the device 102 . Operation 204 is described in more detail by way of example with reference to FIG. 3 .
  • the vision test includes vision test objects used to test a visual acuity level of a user of the device 102 . Attention is now directed to FIGS. 10 and 11 , which illustrate example vision tests, according to some embodiments.
  • FIG. 10 illustrates a vision test 1000 that includes vision test objects 1002 , 1004 , 1006 , and 1008 . As illustrated in FIG. 10 , the vision test objects 1002 , 1004 , 1006 , and 1008 are shown by way of example to include the letter “E” having varying sizes.
  • the user of the device 102 may be asked to select the smallest letter “E” that the user can see. For example, if the user is unable to see letter “E” represented by the vision test object 1002 but is able to see letter “E” represented by the vision test object 1004 , the smallest letter that the user can see would be the letter “E” corresponding to test object 1004 .
  • FIG. 11 illustrates a vision test 1100 that includes vision test objects 1102 , 1104 , and 1106 .
  • the vision test objects 1102 , 1104 , and 1106 include a “Tic-Tac-Toe” pattern including an “X” in the center of the pattern.
  • the user of the device 102 may be asked to select the smallest “Tic-Tac-Toe” pattern that the user can see where the horizontal and vertical lines appear as straight lines.
  • the vision testing module 104 while presenting the vision test to the user of the device 102 , the vision testing module 104 prompts the user for input identifying the user's ability to see at least one vision test object in the vision test. In some embodiments, the vision testing module 104 generates visual prompts (e.g., text). In some embodiments, the vision testing module 104 generates audio prompts (e.g., voice prompts).
  • the vision test includes a sequence of vision test objects that are displayed to the user.
  • the sequence of vision test objects may include vision test objects from different types of vision test techniques (e.g., the vision test 1000 , the vision test 1100 ).
  • the sequence of vision test objects may be used to obtain a more precise level of visual acuity for the user.
  • the vision testing module 104 may present a first vision test object and may ask the user whether the user can see the first vision test object. Based on the user's response, the vision testing module 104 presents an appropriate second vision test object. The vision testing module 104 repeats the process until the level of visual acuity is determined.
  • the vision testing module 104 receives (operation 206 ) input from the user identifying the user's ability to see at least one vision test object in the vision test. For example, the vision testing module 104 may receive input from the user indicating that the smallest letter “E” that the user can see is the letter “E” corresponding to the vision test object 1004 . Similarly, the vision testing module 104 may receive input from the user indicating that smallest “Tic-Tac-Toe” pattern that the user can see where the horizontal and vertical lines appear as straight lines is the “Tic-Tao-Toe” pattern corresponding to the vision test object 1104 . It is however to be noted that this disclosure is not limited to the example vision test illustrated in FIGS. 10 and 11 and that other example embodiments may include other vision tests.
  • the vision testing module 104 identifies (operation 208 ) display settings corresponding to the input received from the user (e.g., from a touch screen on a mobile device, computer or other electronic device including a display screen).
  • the display settings include one or more of a size of an object, a shape of the object, a color of the object, a brightness of the object, a contrast level of the object, and a location of the object. Operation 208 is described in more detail by way of example with reference to FIG. 5 .
  • the user interface module 106 displays (operation 210 ) objects in the graphical user interface based on the display settings.
  • the objects may include icons and/or text.
  • objects are displayed in the graphical user interface using the display settings corresponding to a visual acuity level of a user without requiring the user to navigate through menus. Operation 210 is described in more detail by way of example with reference to FIG. 6 .
  • FIG. 3 is a flowchart of a method for presenting (operation 204 ) a vision test in the graphical user interface of the device 102 , according to some embodiments.
  • the vision testing module 104 determines (operation 302 ) specifications of a display device for the device 102 and a predetermined viewing distance of the display device for the device 102 .
  • the specifications of the display device include physical dimensions of the display device and a resolution of the display device.
  • the user interface module 106 displays (operation 304 ) the vision test objects in the graphical user interface of the device 102 based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102 .
  • the user interface module 106 may scale the letters in the Snellen test to account for the dimensions of the display device, the resolution of the display device, and/or the predetermined viewing distance of the display device.
  • Operation 304 is described in more detail with respect to FIG. 4 , which is a flowchart of a method for displaying (operation 304 ) vision test objects in the graphical user interface of the device 102 , according to some embodiments.
  • the vision testing module 104 calculates (operation 402 ) at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102 .
  • the vision testing module 104 scales (operation 404 ) reference vision test objects using the at least one scaling factor. In some embodiments, each reference vision test object corresponds to a reference visual acuity level.
  • the vision testing module 104 displays (operation 406 ) the scaled reference vision test objects in the graphical user interface of the device 102 .
  • FIG. 5 is a flowchart of a method for identifying (operation 208 ) display settings corresponding to input received from a user, according to some embodiments.
  • the vision testing module 104 identifies (operation 502 ) at least one vision test object selected by the user of the device 102 . For example, the vision testing module 104 may identify that the user selected the vision test object 1002 .
  • the vision testing module 104 determines (operation 504 ) a reference visual acuity level corresponding to the at least one vision test object.
  • a reference visual acuity level may be a standard (or normalized) visual acuity level.
  • the reference visual acuity level may correspond to a Snellen fraction (e.g., 20/40, 20/20).
  • the reference visual acuity level may be independent of the dimensions of the display device, the resolution of the display device, and the predetermined viewing distance of the display device.
  • the user enters the reference visual acuity level of the user. For example, if the user has a visual acuity level of 20/40, the user may enter this visual acuity level into a text field of the vision test. In these embodiments, the operations 502 and 504 are not performed.
  • the vision testing module 104 determines (operation 506 ) reference display settings based on the reference visual acuity level.
  • the reference display settings may be a standard (or normalized) display setting based on the reference visual acuity level and a reference display device.
  • the reference display setting corresponding to a Snellen fraction 20/20 may indicate that a reference display device that is a 23′′ widescreen display device viewed 24′′ away from the user may be set to a resolution of 2048 ⁇ 1152.
  • the reference display setting corresponding to a Snellen fraction 20/40 may indicate that the reference display device (e.g., the 23′′ widescreen display device) viewed 24′′ away from the user may be set to a resolution of 1280 ⁇ 720.
  • the reference display settings are used to scale up or scale down the size of objects displayed on the display device of the device 102 .
  • the mapping of the scale factors e.g., based on the dimensions of the display device, the current resolution of the display device, and the distance of the display device to the user's eye
  • the mapping of the scale factors may be stored in the database 108 .
  • the vision testing module 104 determines (operation 508 ) specifications of a display device for the device 102 and a predetermined viewing distance of the display device for the device 102 .
  • the vision testing module 104 may determine the dimensions and the current resolution of the display device for the device 102 .
  • the vision testing module 104 calculates (510) at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102 .
  • a mobile phone that is 2 inches by 4 inches, has a resolution of 960 ⁇ 640, and is typically viewed at 12 inches from the user's eye may have a scaling factor of 2.
  • the vision testing module 104 scales (operation 512 ) the reference display settings based on the at least one scaling factor to produce the display settings corresponding to the input received from the user. For example, assume that the reference visual acuity level is 20/40 and the display device is the mobile phone described above. Accordingly, the scaling factor is 2 and resolution of the mobile phone is decreased by a factor of 2 (e.g., to increase the size of objects displayed). Note that scale factor accounts for the dimensions of the display device, the resolution of the display device, and the predetermined viewing distance of the display device relative to the reference display device.
  • FIG. 6 is a flowchart of a method for displaying (operation 210 ) objects in the graphical user interface based on the display settings, according to some embodiments.
  • the user interface module 106 scales (operation 602 ) the objects based on the display setting and displays (operation 604 ) at least a subset of the scaled objects in a viewable area of the graphical user interface.
  • the user interface module 106 places (operation 606 ) the scaled objects that are not displayed in the viewable area in a scrollable area of the graphical user interface.
  • FIG. 12A illustrates example objects 1202 - 1224 displayed in an example graphical user interface (GUI) 1200 , according to some embodiments.
  • GUI graphical user interface
  • the user interface module 106 may only display the example objects 1202 , 1204 , 1208 , 1210 , 1214 , and 1216 in a viewable area 1230 of the GUI 1200 , as illustrated in FIG. 12B .
  • the example objects 1206 , 1212 , 1218 , 1220 , 1222 , and 1224 may be placed in the scrollable area 1232 of the GUI 1200 .
  • FIG. 7 is a flowchart of a method 700 for generating additional viewing pages to display objects not displayed in a viewable area of the graphical user interface, according to some embodiments.
  • the user interface module 106 generates (operation 702 ) additional viewing pages to accommodate the scaled objects that are not displayed in the viewable area.
  • the user interface module 106 places (operation 704 ) the scaled objects that are not displayed in the viewable area on the additional viewing pages.
  • FIG. 12C illustrates the example objects 1202 , 1204 , and 1206 after being scaled to increase the size of the objects displayed on a viewing page of the GUI 1200 . Due to the increased size of the objects, only three objects are displayed on the viewing page of the GUI 1200 . The other objects are placed on other viewing pages. Assuming that the unsealed versions of the example objects 1202 - 1224 fit on one viewing page, the user interface module creates three additional viewing pages to accommodate the other objects not displayed.
  • FIG. 8 is a flowchart of a method 800 for restoring display settings at the end of a login session, according to some embodiments.
  • the vision testing module 104 detects (operation 802 ) an end of a login session on the device and restores ( 804 ) the display setting to display settings used prior to presenting the vision test.
  • FIG. 9 is a flowchart of a method 900 for retaining display settings between login sessions, according to some embodiments.
  • the vision testing module 104 detects (operation 902 ) an end of a login session on the device and retains (operation 904 ) the display settings for a future login session on the device.
  • the display setting may be saved in persistent memory of an electronic device.
  • the associated display setting are retrieved from memory and used to configure the display.
  • FIG. 13 depicts a block diagram of a machine in the example form of a computer system 1300 within which may be executed a set of instructions for causing the machine to perform any one or more of the methodologies discussed herein.
  • the machine operates as a standalone device or may be connected (e.g., networked) to other machines.
  • the machine may operate in the capacity of a server or a client machine in a server-client network environment or as a peer machine in a peer-to-peer (or distributed) network environment.
  • the computer system 1300 may include, but is not limited to, a desktop computer system, a laptop computer system, a server, a mobile phone, a smart phone, a personal digital assistant (PDA), a gaming console, a portable gaming console, a set top box, a camera, a printer, a television set, or any other electronic device.
  • a desktop computer system a laptop computer system
  • a server a mobile phone, a smart phone, a personal digital assistant (PDA), a gaming console, a portable gaming console, a set top box, a camera, a printer, a television set, or any other electronic device.
  • PDA personal digital assistant
  • the machine is capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perforin any one or more of the methodologies discussed herein.
  • the example of the computer system 1300 includes a processor 1302 (e.g., a central processing unit (CPU), a graphics processing unit (GPU) or both), and memory 1304 , which communicate with each other via bus 1308 .
  • Memory 1304 includes volatile memory devices (e.g., DRAM, SRAM, DDR RAM, or other volatile solid state memory devices), non-volatile memory devices (e.g., magnetic disk memory devices, optical disk memory devices, flash memory devices, tape drives, or other non-volatile solid state memory devices), or a combination thereof.
  • Memory 1304 may optionally include one or more storage devices remotely located from the computer system 1300 .
  • the computer system 1300 may further include a video display unit 1306 (e.g., a plasma display, a liquid crystal display (LCD) or a cathode ray tube (CRT)).
  • the computer system 1300 also includes input devices 1310 (e.g., keyboard, mouse, trackball, touchscreen display, etc.), output devices 1312 (e.g., speakers), and a network interface device 1316 .
  • the aforementioned components of the computer system 1300 may be located within a single housing or case (e.g., as depicted by the dashed lines in FIG. 13 ). Alternatively, a subset of the components may be located outside of the housing.
  • the video display unit 1306 , the input devices 1310 , and the output devices 1312 may exist outside of the housing, but be coupled to the bus 1308 via external ports or connectors accessible on the outside of the housing.
  • Memory 1304 includes a machine-readable medium 1320 on which is stored one or more sets of data structures and instructions 1322 (e.g., software) embodying or utilized by any one or more of the methodologies or functions described herein.
  • the one or more sets of data structures may store data.
  • a machine-readable medium refers to a storage medium that is readable by a machine (e.g., a computer-readable storage medium).
  • the data structures and instructions 1322 may also reside, completely or at least partially, within memory 1304 and/or within the processor 1302 during execution thereof by computer system 1300 , with memory 1304 and processor 1302 also constituting machine-readable, tangible media.
  • the data structures and instructions 1322 may further be transmitted or received over a network 1350 via network interface device 1316 utilizing any one of a number of well-known transfer protocols (e.g., HyperText Transfer Protocol (HTTP)).
  • Network 1350 can generally include any type of wired or wireless communication channel capable of coupling together computing nodes (e.g., the computer system 1300 ). This includes, but is not limited to, a local area network (LAN), a wide area network (WAN), or a combination of networks. In some embodiments, network 1350 includes the Internet.
  • Modules may constitute either software modules (e.g., code and/or instructions embodied on a machine-readable medium or in a transmission signal) or hardware modules.
  • a hardware module is a tangible unit capable of performing certain operations and may be configured or arranged in a certain manner.
  • one or more computer systems e.g., the computer system 1300
  • one or more hardware modules of a computer system e.g., a processor 1302 or a group of processors
  • software e.g., an application or application portion
  • a hardware module may be implemented mechanically or electronically.
  • a hardware module may comprise dedicated circuitry or logic that is permanently configured (e.g., as a special-purpose processor, such as a field programmable gate array (FPGA) or an application-specific integrated circuit (ASIC)) to perform certain operations.
  • a hardware module may also comprise programmable logic or circuitry (e.g., as encompassed within a processor 1302 or other programmable processor) that is temporarily configured by software to perform certain operations. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.
  • the term “hardware module” should be understood to encompass a tangible entity, be that an entity that is physically constructed, permanently configured (e.g., hardwired) or temporarily configured (e.g., programmed) to operate in a certain manner and/or to perform certain operations described herein.
  • hardware modules are temporarily configured (e.g., programmed)
  • each of the hardware modules need not be configured or instantiated at any one instance in time.
  • the hardware modules comprise a processor 1302 configured using software
  • the processor 1302 may be configured as respective different hardware modules at different times.
  • Software may accordingly configure a processor 1302 , for example, to constitute a particular hardware module at one instance of time and to constitute a different hardware module at a different instance of time.
  • Modules can provide information to, and receive information from, other modules.
  • the described modules may be regarded as being communicatively coupled.
  • communications may be achieved through signal transmission (e.g., over appropriate circuits and buses) that connect the modules.
  • communications between such modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple modules have access.
  • one module may perform an operation and store the output of that operation in a memory device to which it is communicatively coupled.
  • a further module may then, at a later time, access the memory device to retrieve and process the stored output.
  • Modules may also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).
  • processors 1302 may be temporarily configured (e.g., by software, code, and/or instructions stored in a machine-readable medium) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors 1302 may constitute processor-implemented (or computer-implemented) modules that operate to perform one or more operations or functions.
  • the modules referred to herein may, in some example embodiments, comprise processor-implemented (or computer-implemented) modules.
  • the methods described herein may be at least partially processor-implemented (or computer-implemented) and/or processor-executable (or computer-executable). For example, at least some of the operations of a method may be performed by one or more processors 1302 or processor-implemented (or computer-implemented) modules. Similarly, at least some of the operations of a method may be governed by instructions that are stored in a computer readable storage medium and executed by one or more processors 1302 or processor-implemented (or computer-implemented) modules. The performance of certain of the operations may be distributed among the one or more processors 1302 , not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the processors 1302 may be located in a single location (e.g., within a home environment, an office environment or as a server farm), while in other embodiments the processors 1302 may be distributed across a number of locations.

Abstract

A system, a computer readable storage medium including programs, and a computer-implemented method for displaying objects in a graphical user interface of a device are described. A vision test is presented in the graphical user interface of the device, the vision test including vision test objects used to test a visual acuity level of a user of the device. Input is received from the user identifying the user's ability to see at least one vision test object in the vision test. Display settings corresponding to the input received from the user are identified. Objects in the graphical user interface are displayed based on the display settings.

Description

    TECHNICAL FIELD
  • The disclosed embodiments relate generally to displaying objects in a graphical user interface based on a visual acuity of a viewer.
  • BACKGROUND
  • Many users of electronic devices have issues with vision that inhibit the users from being able to see objects displayed on electronic devices. Glasses or contact lenses may allow users that are nearsighted or farsighted to see these objects. However, users may not be wearing their glasses or contact lenses when using electronic devices. Furthermore, some users may have vision issues that are not correctable using glasses or contact lenses. Although some electronic devices allow users to change the display settings of these electronic devices, the users still need to navigate the user interface of the electronic device in order to change the display settings. For example, in the Windows operating system, a user must click on the “Start” button, move the cursor to the “Settings” menu item, select “Control Panel”, double click on the “Display Settings” icon, click on the “Settings” tab of the “Display Properties” window, and slide a resolution slider to select a desired resolution level. During this process, if the user cannot see the text and/or icons, the user may not be able to navigate through a series of screens or graphical user interfaces to change the display settings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The embodiments disclosed in the present disclosure are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings. Like reference numerals refer to corresponding parts throughout the drawings.
  • FIG. 1A is a block diagram illustrating a device, according to some embodiments.
  • FIG. 1B illustrates an example method to customize display settings of a display screen of an electronic device (e.g., during a boot-up sequence of the device), according to some embodiments.
  • FIG. 2 is a flowchart of a method for displaying objects in a graphical user interface of a device, according to some embodiments.
  • FIG. 3 is a flowchart of a method for presenting a vision test in a graphical user interface of a device, according to some embodiments.
  • FIG. 4 is a flowchart of a method for displaying vision test objects in a graphical user interface of a device, according to some embodiments.
  • FIG. 5 is a flowchart of a method for identifying display settings corresponding to input received from a user, according to some embodiments.
  • FIG. 6 is a flowchart of a method for displaying objects in a graphical user interface based on the display settings, according to some embodiments.
  • FIG. 7 is a flowchart of a method for generating additional viewing pages to display objects not displayed in a viewable area of a graphical user interface, according to some embodiments.
  • FIG. 8 is a flowchart of a method for restoring display settings at the end of a login session, according to some embodiments.
  • FIG. 9 is a flowchart of a method for retaining display settings between login sessions, according to some embodiments.
  • FIG. 10 illustrates an example vision test, according to some embodiments.
  • FIG. 11 illustrates another example vision test, according to some embodiments.
  • FIG. 12A illustrates objects displayed in an example graphical user interface, according to some embodiments.
  • FIG. 12B illustrates the objects displayed in the example graphical user interface of FIG. 12A after display settings for the graphical user interface have been modified, according to some embodiments.
  • FIG. 12C illustrates the objects displayed in the example graphical user interface of FIG. 12A after display settings for the graphical user interface have been modified, according to some embodiments.
  • FIG. 13 is a block diagram illustrating an example machine for performing the methodologies described herein, according to some embodiments.
  • DESCRIPTION OF EXAMPLE EMBODIMENTS
  • The description that follows includes illustrative systems, methods, techniques, instruction sequences, and computing machine program products that embody illustrative embodiments. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide an understanding of various embodiments of the inventive subject matter. It will be evident, however, to those skilled in the art that embodiments of the inventive subject matter may be practiced without these specific details. In general, well-known instruction instances, protocols, structures and techniques have not been shown in detail.
  • The example embodiments described herein provide techniques for displaying objects in a graphical user interface based on a visual acuity of a viewer.
  • In some embodiments, a vision test is presented in the graphical user interface of the device, where the vision test includes vision test objects used to test a visual acuity level of a user of the device. Input is then received from the user identifying the user's ability to see at least one vision test object in the vision test. Next, display settings corresponding to the input received from the user are identified. Objects are then displayed in the graphical user interface based on the display settings. These embodiments are described in more detail below.
  • FIG. 1A is a block diagram illustrating a device 102, according to some embodiments. The device 102 may include any electronic device coupled to a display device (e.g., a monitor, a television) including, but is not limited to, a desktop computer system, a laptop computer system, a server, a mobile phone, a smart phone, a personal digital assistant (PDA), a gaming console, a portable gaming console, a set top box, a camera, a printer, a television set, or the like.
  • The device 102 includes a vision testing module 104, a user interface module 106, and a database 108. The vision testing module 104 is configured to determine a visual acuity of a user of the device 102. The visual acuity of the user is the ability of the user of the device 102 to see objects displayed on a display device (e.g., a monitor) of the device 102. The visual acuity of the user may be determined by an optometrist or an ophthalmologist using a Snellen Test. The Snellen test includes block letters of varying sizes displayed at a predetermined distance (e.g., 20 feet from the user). The ability of a user to read a particular line of letters indicates the visual acuity of the user. However, other factors may influence the ability of the user to see objects displayed on the display device of the device 102. For example, the dimensions and the resolution of the display device may affect the ability of the user to see objects displayed on the display device. The user may be able to see objects displayed on a large display at a low resolution because the objects may be large enough to be seen by the user. However, the user may not be able to see objects displayed on a large display at a high resolution because the individual objects may be too small to be seen by the user. Similarly, the user may not be able to see objects displayed on a small display at a high resolution because the objects may be too small to be seen by the user. However, the user may be able to see objects displayed on a small display at a low resolution because the objects may be large enough to be seen by the user.
  • The user interface module 106 is configured to display objects in a graphical user interface of a display device of the device 102 based on display settings. The database 108 stores vision test objects, reference visual acuity levels, reference display settings, predetermined viewing distances of display devices (e.g., based dimensions and resolution of the display devices), scaling factors for objects corresponding to the reference visual acuity levels, and user profiles for users of the device 102 that include a visual acuity level and/or display settings for the user (or a plurality of users) with respect to the device 102.
  • FIG. 1B illustrates an example method to customize display settings of a display screen of an electronic device (e.g., during a boot-up sequence) of the device, according to some embodiments. In some embodiments, after the device 102 is powered on (operation 120), the vision testing module 104 is executed and a vision test is presented (operation 122) to the user. In these embodiments, the vision test is administered each time the device 102 is powered on. In some embodiments, after the device 102 is powered on, the vision testing module 104 is executed and a vision test is administered to the user when requested by the user. Display settings that are identified as a result of the vision test are applied to the device (operation 124). In an example embodiment, the display settings are applied to applications running on the device. It is to be appreciated that the method may be performed at any time before, during and after the device boots up. In an example embodiment, the method allows a user who may have visual impairment to adjust or modify display setting on the devices even though the user is unable to read or adequately perceive images or text displayed on the display screen of the display device.
  • FIG. 2 is a flowchart of a method 200 for displaying objects in a graphical user interface of the device 102, according to some embodiments. In some embodiments, the vision testing module 104 receives (operation 202) a request from a user of the device 102 to change display settings of the device 102 based on a vision test. The request may include a voice-activated request, a predetermined key sequence, and/or a predetermined gesture in the graphical user interface of the device 102. Note that operation 202 may be optional. For example, the vision testing module 104 may always present the vision test to the user of the device 102 if the device 102 is a publically-accessible device (e.g., a kiosk, a public computer system) or on the first login of the user on the device 102.
  • The vision testing module 104 presents (operation 204) the vision test in the graphical user interface of the device 102. Operation 204 is described in more detail by way of example with reference to FIG. 3. In some embodiments, the vision test includes vision test objects used to test a visual acuity level of a user of the device 102. Attention is now directed to FIGS. 10 and 11, which illustrate example vision tests, according to some embodiments. FIG. 10 illustrates a vision test 1000 that includes vision test objects 1002, 1004, 1006, and 1008. As illustrated in FIG. 10, the vision test objects 1002, 1004, 1006, and 1008 are shown by way of example to include the letter “E” having varying sizes. During the vision test, the user of the device 102 may be asked to select the smallest letter “E” that the user can see. For example, if the user is unable to see letter “E” represented by the vision test object 1002 but is able to see letter “E” represented by the vision test object 1004, the smallest letter that the user can see would be the letter “E” corresponding to test object 1004.
  • FIG. 11 illustrates a vision test 1100 that includes vision test objects 1102, 1104, and 1106. As illustrated in FIG. 11, the vision test objects 1102, 1104, and 1106 include a “Tic-Tac-Toe” pattern including an “X” in the center of the pattern. During the vision test, the user of the device 102 may be asked to select the smallest “Tic-Tac-Toe” pattern that the user can see where the horizontal and vertical lines appear as straight lines.
  • In some embodiments, while presenting the vision test to the user of the device 102, the vision testing module 104 prompts the user for input identifying the user's ability to see at least one vision test object in the vision test. In some embodiments, the vision testing module 104 generates visual prompts (e.g., text). In some embodiments, the vision testing module 104 generates audio prompts (e.g., voice prompts).
  • In some embodiments, the vision test includes a sequence of vision test objects that are displayed to the user. The sequence of vision test objects may include vision test objects from different types of vision test techniques (e.g., the vision test 1000, the vision test 1100). Alternatively, or additionally, the sequence of vision test objects may be used to obtain a more precise level of visual acuity for the user. For example, the vision testing module 104 may present a first vision test object and may ask the user whether the user can see the first vision test object. Based on the user's response, the vision testing module 104 presents an appropriate second vision test object. The vision testing module 104 repeats the process until the level of visual acuity is determined.
  • Returning to FIG. 2, the vision testing module 104 receives (operation 206) input from the user identifying the user's ability to see at least one vision test object in the vision test. For example, the vision testing module 104 may receive input from the user indicating that the smallest letter “E” that the user can see is the letter “E” corresponding to the vision test object 1004. Similarly, the vision testing module 104 may receive input from the user indicating that smallest “Tic-Tac-Toe” pattern that the user can see where the horizontal and vertical lines appear as straight lines is the “Tic-Tao-Toe” pattern corresponding to the vision test object 1104. It is however to be noted that this disclosure is not limited to the example vision test illustrated in FIGS. 10 and 11 and that other example embodiments may include other vision tests.
  • The vision testing module 104 identifies (operation 208) display settings corresponding to the input received from the user (e.g., from a touch screen on a mobile device, computer or other electronic device including a display screen). In some embodiments, the display settings include one or more of a size of an object, a shape of the object, a color of the object, a brightness of the object, a contrast level of the object, and a location of the object. Operation 208 is described in more detail by way of example with reference to FIG. 5.
  • The user interface module 106 displays (operation 210) objects in the graphical user interface based on the display settings. The objects may include icons and/or text. Thus, as described above, objects are displayed in the graphical user interface using the display settings corresponding to a visual acuity level of a user without requiring the user to navigate through menus. Operation 210 is described in more detail by way of example with reference to FIG. 6.
  • FIG. 3 is a flowchart of a method for presenting (operation 204) a vision test in the graphical user interface of the device 102, according to some embodiments. The vision testing module 104 determines (operation 302) specifications of a display device for the device 102 and a predetermined viewing distance of the display device for the device 102. In some embodiments, the specifications of the display device include physical dimensions of the display device and a resolution of the display device.
  • The user interface module 106 displays (operation 304) the vision test objects in the graphical user interface of the device 102 based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102. For example, to provide a vision test comparable to the Snellen test, the user interface module 106 may scale the letters in the Snellen test to account for the dimensions of the display device, the resolution of the display device, and/or the predetermined viewing distance of the display device. Operation 304 is described in more detail with respect to FIG. 4, which is a flowchart of a method for displaying (operation 304) vision test objects in the graphical user interface of the device 102, according to some embodiments. The vision testing module 104 calculates (operation 402) at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102. The vision testing module 104 scales (operation 404) reference vision test objects using the at least one scaling factor. In some embodiments, each reference vision test object corresponds to a reference visual acuity level. The vision testing module 104 displays (operation 406) the scaled reference vision test objects in the graphical user interface of the device 102.
  • FIG. 5 is a flowchart of a method for identifying (operation 208) display settings corresponding to input received from a user, according to some embodiments. The vision testing module 104 identifies (operation 502) at least one vision test object selected by the user of the device 102. For example, the vision testing module 104 may identify that the user selected the vision test object 1002.
  • The vision testing module 104 determines (operation 504) a reference visual acuity level corresponding to the at least one vision test object. Note that a reference visual acuity level may be a standard (or normalized) visual acuity level. For example, the reference visual acuity level may correspond to a Snellen fraction (e.g., 20/40, 20/20). The reference visual acuity level may be independent of the dimensions of the display device, the resolution of the display device, and the predetermined viewing distance of the display device.
  • In some embodiments, the user enters the reference visual acuity level of the user. For example, if the user has a visual acuity level of 20/40, the user may enter this visual acuity level into a text field of the vision test. In these embodiments, the operations 502 and 504 are not performed.
  • The vision testing module 104 determines (operation 506) reference display settings based on the reference visual acuity level. The reference display settings may be a standard (or normalized) display setting based on the reference visual acuity level and a reference display device. For example, the reference display setting corresponding to a Snellen fraction 20/20 may indicate that a reference display device that is a 23″ widescreen display device viewed 24″ away from the user may be set to a resolution of 2048×1152. Similarly, the reference display setting corresponding to a Snellen fraction 20/40 may indicate that the reference display device (e.g., the 23″ widescreen display device) viewed 24″ away from the user may be set to a resolution of 1280×720. The reference display settings are used to scale up or scale down the size of objects displayed on the display device of the device 102. The mapping of the scale factors (e.g., based on the dimensions of the display device, the current resolution of the display device, and the distance of the display device to the user's eye) for a particular display device (or class of display devices) to a reference display setting may be stored in the database 108.
  • The vision testing module 104 determines (operation 508) specifications of a display device for the device 102 and a predetermined viewing distance of the display device for the device 102. For example, the vision testing module 104 may determine the dimensions and the current resolution of the display device for the device 102.
  • Since the dimensions, the current resolution, and the predetermined viewing distance (e.g., the typical viewing distance) of the display device for the device 102 may affect the vision test being administered to the user, the vision testing module 104 calculates (510) at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device 102. For example, a mobile phone that is 2 inches by 4 inches, has a resolution of 960×640, and is typically viewed at 12 inches from the user's eye may have a scaling factor of 2.
  • The vision testing module 104 scales (operation 512) the reference display settings based on the at least one scaling factor to produce the display settings corresponding to the input received from the user. For example, assume that the reference visual acuity level is 20/40 and the display device is the mobile phone described above. Accordingly, the scaling factor is 2 and resolution of the mobile phone is decreased by a factor of 2 (e.g., to increase the size of objects displayed). Note that scale factor accounts for the dimensions of the display device, the resolution of the display device, and the predetermined viewing distance of the display device relative to the reference display device.
  • FIG. 6 is a flowchart of a method for displaying (operation 210) objects in the graphical user interface based on the display settings, according to some embodiments. The user interface module 106 scales (operation 602) the objects based on the display setting and displays (operation 604) at least a subset of the scaled objects in a viewable area of the graphical user interface. In some embodiments, the user interface module 106 places (operation 606) the scaled objects that are not displayed in the viewable area in a scrollable area of the graphical user interface. For example, FIG. 12A illustrates example objects 1202-1224 displayed in an example graphical user interface (GUI) 1200, according to some embodiments. After scaling the objects, the user interface module 106 may only display the example objects 1202, 1204, 1208, 1210, 1214, and 1216 in a viewable area 1230 of the GUI 1200, as illustrated in FIG. 12B. The example objects 1206, 1212, 1218, 1220, 1222, and 1224 may be placed in the scrollable area 1232 of the GUI 1200.
  • On mobile devices, the GUI may include “pages” that the user may scroll through to access objects not displayed in the viewable area of the GUI. When objects are scaled to increase their size, the number of pages typically increases because the number of objects that can be displayed on a particular viewing page decreases. FIG. 7 is a flowchart of a method 700 for generating additional viewing pages to display objects not displayed in a viewable area of the graphical user interface, according to some embodiments. The user interface module 106 generates (operation 702) additional viewing pages to accommodate the scaled objects that are not displayed in the viewable area. The user interface module 106 places (operation 704) the scaled objects that are not displayed in the viewable area on the additional viewing pages. FIG. 12C illustrates the example objects 1202, 1204, and 1206 after being scaled to increase the size of the objects displayed on a viewing page of the GUI 1200. Due to the increased size of the objects, only three objects are displayed on the viewing page of the GUI 1200. The other objects are placed on other viewing pages. Assuming that the unsealed versions of the example objects 1202-1224 fit on one viewing page, the user interface module creates three additional viewing pages to accommodate the other objects not displayed.
  • On publically-accessible devices (e.g., public computer systems, kiosks), it may be desirable to restore the display settings to a standard display setting (e.g., a display setting of the device 102 prior to the administration of the vision test). FIG. 8 is a flowchart of a method 800 for restoring display settings at the end of a login session, according to some embodiments. The vision testing module 104 detects (operation 802) an end of a login session on the device and restores (804) the display setting to display settings used prior to presenting the vision test.
  • On personal devices, it may be desirable to retain display settings between login sessions. FIG. 9 is a flowchart of a method 900 for retaining display settings between login sessions, according to some embodiments. The vision testing module 104 detects (operation 902) an end of a login session on the device and retains (operation 904) the display settings for a future login session on the device. For example, the display setting may be saved in persistent memory of an electronic device. Upon subsequently login by the same user, the associated display setting are retrieved from memory and used to configure the display.
  • Example Electronic Device/Machine
  • FIG. 13 depicts a block diagram of a machine in the example form of a computer system 1300 within which may be executed a set of instructions for causing the machine to perform any one or more of the methodologies discussed herein. In alternative embodiments, the machine operates as a standalone device or may be connected (e.g., networked) to other machines. In a networked deployment, the machine may operate in the capacity of a server or a client machine in a server-client network environment or as a peer machine in a peer-to-peer (or distributed) network environment. The computer system 1300 may include, but is not limited to, a desktop computer system, a laptop computer system, a server, a mobile phone, a smart phone, a personal digital assistant (PDA), a gaming console, a portable gaming console, a set top box, a camera, a printer, a television set, or any other electronic device.
  • The machine is capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perforin any one or more of the methodologies discussed herein.
  • The example of the computer system 1300 includes a processor 1302 (e.g., a central processing unit (CPU), a graphics processing unit (GPU) or both), and memory 1304, which communicate with each other via bus 1308. Memory 1304 includes volatile memory devices (e.g., DRAM, SRAM, DDR RAM, or other volatile solid state memory devices), non-volatile memory devices (e.g., magnetic disk memory devices, optical disk memory devices, flash memory devices, tape drives, or other non-volatile solid state memory devices), or a combination thereof. Memory 1304 may optionally include one or more storage devices remotely located from the computer system 1300. The computer system 1300 may further include a video display unit 1306 (e.g., a plasma display, a liquid crystal display (LCD) or a cathode ray tube (CRT)). The computer system 1300 also includes input devices 1310 (e.g., keyboard, mouse, trackball, touchscreen display, etc.), output devices 1312 (e.g., speakers), and a network interface device 1316. The aforementioned components of the computer system 1300 may be located within a single housing or case (e.g., as depicted by the dashed lines in FIG. 13). Alternatively, a subset of the components may be located outside of the housing. For example, the video display unit 1306, the input devices 1310, and the output devices 1312 may exist outside of the housing, but be coupled to the bus 1308 via external ports or connectors accessible on the outside of the housing.
  • Memory 1304 includes a machine-readable medium 1320 on which is stored one or more sets of data structures and instructions 1322 (e.g., software) embodying or utilized by any one or more of the methodologies or functions described herein. The one or more sets of data structures may store data. Note that a machine-readable medium refers to a storage medium that is readable by a machine (e.g., a computer-readable storage medium). The data structures and instructions 1322 may also reside, completely or at least partially, within memory 1304 and/or within the processor 1302 during execution thereof by computer system 1300, with memory 1304 and processor 1302 also constituting machine-readable, tangible media.
  • The data structures and instructions 1322 may further be transmitted or received over a network 1350 via network interface device 1316 utilizing any one of a number of well-known transfer protocols (e.g., HyperText Transfer Protocol (HTTP)). Network 1350 can generally include any type of wired or wireless communication channel capable of coupling together computing nodes (e.g., the computer system 1300). This includes, but is not limited to, a local area network (LAN), a wide area network (WAN), or a combination of networks. In some embodiments, network 1350 includes the Internet.
  • Certain embodiments are described herein as including logic or a number of components, modules, or mechanisms. Modules may constitute either software modules (e.g., code and/or instructions embodied on a machine-readable medium or in a transmission signal) or hardware modules. A hardware module is a tangible unit capable of performing certain operations and may be configured or arranged in a certain manner. In example embodiments, one or more computer systems (e.g., the computer system 1300) or one or more hardware modules of a computer system (e.g., a processor 1302 or a group of processors) may be configured by software (e.g., an application or application portion) as a hardware module that operates to perform certain operations as described herein.
  • In various embodiments, a hardware module may be implemented mechanically or electronically. For example, a hardware module may comprise dedicated circuitry or logic that is permanently configured (e.g., as a special-purpose processor, such as a field programmable gate array (FPGA) or an application-specific integrated circuit (ASIC)) to perform certain operations. A hardware module may also comprise programmable logic or circuitry (e.g., as encompassed within a processor 1302 or other programmable processor) that is temporarily configured by software to perform certain operations. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.
  • Accordingly, the term “hardware module” should be understood to encompass a tangible entity, be that an entity that is physically constructed, permanently configured (e.g., hardwired) or temporarily configured (e.g., programmed) to operate in a certain manner and/or to perform certain operations described herein. Considering embodiments in which hardware modules are temporarily configured (e.g., programmed), each of the hardware modules need not be configured or instantiated at any one instance in time. For example, where the hardware modules comprise a processor 1302 configured using software, the processor 1302 may be configured as respective different hardware modules at different times. Software may accordingly configure a processor 1302, for example, to constitute a particular hardware module at one instance of time and to constitute a different hardware module at a different instance of time.
  • Modules can provide information to, and receive information from, other modules. For example, the described modules may be regarded as being communicatively coupled. Where multiples of such hardware modules exist contemporaneously, communications may be achieved through signal transmission (e.g., over appropriate circuits and buses) that connect the modules. In embodiments in which multiple modules are configured or instantiated at different times, communications between such modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple modules have access. For example, one module may perform an operation and store the output of that operation in a memory device to which it is communicatively coupled. A further module may then, at a later time, access the memory device to retrieve and process the stored output. Modules may also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).
  • The various operations of example methods described herein may be performed, at least partially, by one or more processors 1302 that are temporarily configured (e.g., by software, code, and/or instructions stored in a machine-readable medium) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors 1302 may constitute processor-implemented (or computer-implemented) modules that operate to perform one or more operations or functions. The modules referred to herein may, in some example embodiments, comprise processor-implemented (or computer-implemented) modules.
  • Moreover, the methods described herein may be at least partially processor-implemented (or computer-implemented) and/or processor-executable (or computer-executable). For example, at least some of the operations of a method may be performed by one or more processors 1302 or processor-implemented (or computer-implemented) modules. Similarly, at least some of the operations of a method may be governed by instructions that are stored in a computer readable storage medium and executed by one or more processors 1302 or processor-implemented (or computer-implemented) modules. The performance of certain of the operations may be distributed among the one or more processors 1302, not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the processors 1302 may be located in a single location (e.g., within a home environment, an office environment or as a server farm), while in other embodiments the processors 1302 may be distributed across a number of locations.
  • While the embodiment(s) is (are) described with reference to various implementations and exploitations, it will be understood that these embodiments are illustrative and that the scope of the embodiment(s) is not limited to them. In general, techniques for the embodiments described herein may be implemented with facilities consistent with any hardware system or hardware systems defined herein. Many variations, modifications, additions, and improvements are possible.
  • Plural instances may be provided for components, operations or structures described herein as a single instance. Finally, boundaries between various components, operations, and data stores are somewhat arbitrary, and particular operations are illustrated in the context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within the scope of the embodiment(s). In general, structures and functionality presented as separate components in the example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the embodiment(s).
  • The foregoing description, for purpose of explanation, has been described with reference to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the embodiments to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles and their practical applications, to thereby enable others skilled in the art to best utilize the embodiments and various embodiments with various modifications as are suited to the particular use contemplated.

Claims (21)

1-21. (canceled)
22. A computer-implemented method for displaying objects in a graphical user interface of a device, the method comprising:
initiating, by a vision test module, a vision test including the steps of:
presenting the vision test in the graphical user interface of the device using current display settings, the vision test including vision test objects used to test a visual acuity level of a user of the device, the visual acuity level identifying visual acuity of the user;
receiving input from the user identifying the users ability to see at least one vision test object in the vision test;
identifying, in a database, customized display settings associated with the at least one vision test object;
changing the current display settings to the customized display settings to compensate for the visual impairment; and
saving the customized display settings in persistent memory of the device for subsequent display configuration upon login by the user;
displaying, by the user interface module, further objects generated by applications running on the device with the customized display settings.
23. The computer-implemented method of claim 22, wherein presenting the vision test in the graphical user interface of the device includes:
determining specifications of a display device for the device and a predetermined viewing distance of the display device for the device; and
displaying the vision test objects in the graphical user interface of the device based on the specifications of the display device and the predetermined viewing distance of the display device for the device.
24. The computer-implemented method of claim 23, wherein displaying the vision test objects in the graphical user interface of the device based on the specifications of the display device and the predetermined viewing distance of the display device for the device includes:
calculating at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device;
scaling reference vision test objects using the at least one scaling factor, each reference vision test object corresponding to a reference visual acuity level; and
displaying the scaled reference vision test objects in the graphical user interface of the device.
25. The computer-implemented method of claim 22, wherein identifying the display settings corresponding to the input received from the user includes:
identifying at least one vision test object selected by the user of the device;
determining a reference visual acuity level corresponding to the at least one vision test object;
determining reference display settings based on the reference visual acuity level;
determining specifications of a display device for the device and a predetermined viewing distance of the display device for the device;
calculating at least one scaling factor based on the specifications of the display device and the predetermined viewing distance of the display device for the device; and
scaling the reference display settings based on the at least one scaling factor to produce the display settings corresponding to the input received from the user.
26. The computer-implemented method of claim 22, wherein displaying the objects in the graphical user interface based on the display settings includes:
scaling the objects based on the display settings; and
displaying at least a subset of the scaled objects in a viewable area of the graphical user interface.
27. The computer-implemented method of claim 22, wherein the objects include icons.
28. The computer-implemented method of claim 22, wherein the objects include text.
29. The computer-implemented method of claim 22, further comprising:
detecting an end of a login session on the device; and
restoring the display setting to display settings used prior to presenting the vision test.
30. The computer-implemented method of claim 22, further comprising:
detecting an end of a login session on the device; and
retaining the display settings for a future login session on the device.
31. The computer-implemented method of claim 22, wherein prior to presenting the vision test in the graphical user interface of the device, the method further comprises receiving a request from the user of the device to change display settings of the device based on the vision test.
32. The computer-implemented method of claim 22, wherein the display settings are selected from the group consisting of:
a size of an object;
a shape of the object;
a color of the object;
a brightness of the object;
a contrast level of the object; and
a location of the object.
33. The computer-implemented method of claim 22, wherein the objects are displayed in the graphical user interface using the display settings without requiring the user to navigate through menus.
34. The computer-implemented method of claim 22, wherein the vision test is a Snellen test.
35. A system to display objects in a graphical user interface of a device, the system comprising:
a processor including a vision testing module configured to:
initiate a vision test as used by an optometrist or an ophthalmologist on the device, the vision test including:
presenting the vision test in the graphical user interface of the device using current display settings, the vision test including vision test objects used to test a visual acuity level of a user of the device, the visual acuity level identifying visual acuity of the user;
receiving input from the user identifying the user's ability to see at least one vision test object in the vision test; and
identifying, in a database, customized display settings associated with the at least one vision test object;
changing the current display settings to the customized display settings to compensate for the visual impairment; and
saving the customized display settings in persistent memory of the device for subsequent display configuration upon login by the user;
displaying, by the user interface module, further objects generated by applications running on the device with the customized display settings.
36. The system of claim 35, wherein when presenting the vision test in the graphical user interface of the device, the processor-implemented vision testing module is configured to:
determine specifications of a display device for the device and a predetermined viewing distance of the display device for the device; and
display the vision test objects in the graphical user interface of the device based on the specifications of the display device and the predetermined viewing distance of the display device for the device.
37. The system of claim 35, wherein the vision test is a Snellen test.
38. A non-transitory computer-implemented method for displaying objects in a graphical user interface of a device, the method comprising:
initiating, by a vision test module, a vision test on the device, the vision test including presenting the vision test in the graphical user interface of the device using current display settings, the vision test including vision test objects used to test a visual acuity level, the visual acuity level identifying visual acuity of the user;
receiving input from the user identifying the user's ability to see at least one vision test object in the vision test;
identifying, in a database, customized display settings associated with the at least one vision test object;
changing the current display settings to the customized display settings to compensate for the visual impairment; and
saving the customized display settings in persistent memory of the device for subsequent display configuration upon login by the user;
displaying further objects generated by applications running on the device with the customized display settings.
39. The computer-implemented method of claim 38, wherein the vision test is a Snellen test.
40. The computer-implemented method of claim 38, wherein the database stores scaling factors, the method further comprising scaling the vision test objects using the scaling factors based on reference visual acuity levels.
41. The computer-implemented method of claim 38, wherein customized display settings are saved in a user profile associated with a user of the device.
US14/505,900 2011-04-01 2014-10-03 System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer Abandoned US20150026587A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/505,900 US20150026587A1 (en) 2011-04-01 2014-10-03 System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/078,661 US8881058B2 (en) 2011-04-01 2011-04-01 System and method for displaying objects in a user interface based on a visual acuity of a viewer
US14/505,900 US20150026587A1 (en) 2011-04-01 2014-10-03 System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/078,661 Continuation US8881058B2 (en) 2011-04-01 2011-04-01 System and method for displaying objects in a user interface based on a visual acuity of a viewer

Publications (1)

Publication Number Publication Date
US20150026587A1 true US20150026587A1 (en) 2015-01-22

Family

ID=46929001

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/078,661 Active US8881058B2 (en) 2011-04-01 2011-04-01 System and method for displaying objects in a user interface based on a visual acuity of a viewer
US14/505,900 Abandoned US20150026587A1 (en) 2011-04-01 2014-10-03 System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/078,661 Active US8881058B2 (en) 2011-04-01 2011-04-01 System and method for displaying objects in a user interface based on a visual acuity of a viewer

Country Status (1)

Country Link
US (2) US8881058B2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105664433A (en) * 2016-01-20 2016-06-15 毕宏生 Vision comprehensive training device
US20180376212A1 (en) * 2017-06-23 2018-12-27 Sony Corporation Modifying display region for people with vision impairment
US20190166343A1 (en) * 2017-11-27 2019-05-30 International Business Machines Corporation Optimized computer display rendering for user vision conditions
US10441845B2 (en) 2018-01-06 2019-10-15 Mkh Llc Exercise assembly for squats
US10650702B2 (en) 2017-07-10 2020-05-12 Sony Corporation Modifying display region for people with loss of peripheral vision
US10805676B2 (en) 2017-07-10 2020-10-13 Sony Corporation Modifying display region for people with macular degeneration
US10845954B2 (en) 2017-07-11 2020-11-24 Sony Corporation Presenting audio video display options as list or matrix

Families Citing this family (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7873610B2 (en) 2006-05-26 2011-01-18 Andrew S Poulsen Meta-configuration of profiles
TW201239644A (en) * 2011-03-24 2012-10-01 Hon Hai Prec Ind Co Ltd System and method for dynamically adjusting font size on screen
TW201239869A (en) * 2011-03-24 2012-10-01 Hon Hai Prec Ind Co Ltd System and method for adjusting font size on screen
US8881058B2 (en) * 2011-04-01 2014-11-04 Arthur Austin Ollivierre System and method for displaying objects in a user interface based on a visual acuity of a viewer
US20130127821A1 (en) * 2011-05-11 2013-05-23 Jeffrey Phillip Lewis Method and system for adjusting a display to account for the users' corrective lenses or preferred display settings
CN102999382A (en) * 2011-09-09 2013-03-27 鸿富锦精密工业(深圳)有限公司 Electronic device and switching method thereof
US9314154B2 (en) 2011-10-17 2016-04-19 The Board Of Trustees Of The Leland Stanford Junior University System and method for providing analysis of visual function using a mobile device with display
US9462941B2 (en) 2011-10-17 2016-10-11 The Board Of Trustees Of The Leland Stanford Junior University Metamorphopsia testing and related methods
US20140137054A1 (en) * 2012-11-14 2014-05-15 Ebay Inc. Automatic adjustment of font on a visual display
CN105009055A (en) * 2013-01-31 2015-10-28 惠普发展公司,有限责任合伙企业 Defining a design plan
NL2010302C2 (en) * 2013-02-14 2014-08-18 Optelec Dev B V A system for determining a recommended magnification factor for a magnifier such as a loupe or an electronic magnifier to be used by a person.
US20140282285A1 (en) * 2013-03-14 2014-09-18 Cellco Partnership D/B/A Verizon Wireless Modifying a user interface setting based on a vision ability of a user
FR3005194B1 (en) * 2013-04-25 2016-09-23 Essilor Int METHOD FOR CUSTOMIZING AN IMAGE DISPLAY ELECTRONIC DEVICE
US20140362110A1 (en) * 2013-06-08 2014-12-11 Sony Computer Entertainment Inc. Systems and methods for customizing optical representation of views provided by a head mounted display based on optical prescription of a user
WO2015047205A1 (en) * 2013-09-30 2015-04-02 Echostar Ukraine, L.L.C. Systems, devices and methods for font size selection
US9715863B2 (en) * 2013-09-30 2017-07-25 Microsoft Technology Licensing, Llc Scale factor based on viewing distance
JP2015090569A (en) * 2013-11-06 2015-05-11 ソニー株式会社 Information processing device and information processing method
US10324593B2 (en) 2014-01-28 2019-06-18 International Business Machines Corporation Impairment-adaptive electronic data interaction system
USD766330S1 (en) * 2014-10-10 2016-09-13 King.Com Limited Display screen with icon
US9367129B1 (en) * 2015-02-05 2016-06-14 Wipro Limited Method and system for controlling display of content to user
EP3090913B1 (en) * 2015-05-08 2021-09-29 Continental Automotive GmbH Vehicle control system and method
CA2901477C (en) 2015-08-25 2023-07-18 Evolution Optiks Limited Vision correction system, method and graphical user interface for implementation on electronic devices having a graphical display
CN105380591A (en) * 2015-11-26 2016-03-09 上海斐讯数据通信技术有限公司 Vision detecting device, system and method
CN105708412A (en) * 2016-01-18 2016-06-29 鞍山心态度科技有限公司 Eyesight testing method and device based on television
GB2546517A (en) * 2016-01-21 2017-07-26 Young Susan Eye-testing method
DE102016201250A1 (en) * 2016-01-28 2017-08-03 Conti Temic Microelectronic Gmbh Method and device for determining the range of a sensor for a motor vehicle
CN107198505A (en) * 2017-04-07 2017-09-26 天津市天中依脉科技开发有限公司 Visual function detecting system and method based on smart mobile phone
US10878778B2 (en) * 2017-11-08 2020-12-29 Paypal, Inc. Reactive interface based on eyestrain
US10413172B2 (en) 2017-12-11 2019-09-17 1-800 Contacts, Inc. Digital visual acuity eye examination for remote physician assessment
US11693239B2 (en) 2018-03-09 2023-07-04 Evolution Optiks Limited Vision correction system and method, light field display and light field shaping layer and alignment therefor
CA3021636A1 (en) 2018-10-22 2020-04-22 Evolution Optiks Limited Light field display, adjusted pixel rendering method therefor, and vision correction system and method using same
US11353699B2 (en) 2018-03-09 2022-06-07 Evolution Optiks Limited Vision correction system and method, light field display and light field shaping layer and alignment therefor
EP3542704A1 (en) * 2018-03-22 2019-09-25 Tilak Healthcare Visual testing using mobile devices
CN109431445A (en) * 2018-08-03 2019-03-08 广州视光专业技术服务有限公司 A kind of vision monitoring method, apparatus, terminal device and storage medium
US10636116B1 (en) 2018-10-22 2020-04-28 Evolution Optiks Limited Light field display, adjusted pixel rendering method therefor, and vision correction system and method using same
US11500460B2 (en) 2018-10-22 2022-11-15 Evolution Optiks Limited Light field device, optical aberration compensation or simulation rendering
US10936064B2 (en) 2018-10-22 2021-03-02 Evolution Optiks Limited Light field display, adjusted pixel rendering method therefor, and adjusted vision perception system and method using same addressing astigmatism or similar conditions
US11327563B2 (en) 2018-10-22 2022-05-10 Evolution Optiks Limited Light field vision-based testing device, adjusted pixel rendering method therefor, and online vision-based testing management system and method using same
US11287883B2 (en) 2018-10-22 2022-03-29 Evolution Optiks Limited Light field device, pixel rendering method therefor, and adjusted vision perception system and method using same
US10761604B2 (en) 2018-10-22 2020-09-01 Evolution Optiks Limited Light field vision testing device, adjusted pixel rendering method therefor, and vision testing system and method using same
US10860099B2 (en) 2018-10-22 2020-12-08 Evolution Optiks Limited Light field display, adjusted pixel rendering method therefor, and adjusted vision perception system and method using same addressing astigmatism or similar conditions
US11500461B2 (en) 2019-11-01 2022-11-15 Evolution Optiks Limited Light field vision-based testing device, system and method
US11789531B2 (en) 2019-01-28 2023-10-17 Evolution Optiks Limited Light field vision-based testing device, system and method
US11635617B2 (en) 2019-04-23 2023-04-25 Evolution Optiks Limited Digital display device comprising a complementary light field display or display portion, and vision correction system and method using same
US11902498B2 (en) 2019-08-26 2024-02-13 Evolution Optiks Limited Binocular light field display, adjusted pixel rendering method therefor, and vision correction system and method using same
US11823598B2 (en) 2019-11-01 2023-11-21 Evolution Optiks Limited Light field device, variable perception pixel rendering method therefor, and variable perception system and method using same
US11487361B1 (en) 2019-11-01 2022-11-01 Evolution Optiks Limited Light field device and vision testing system using same
CN111700583B (en) * 2020-05-23 2023-04-18 福建生物工程职业技术学院 Detection method of indoor shared self-service vision detection system

Citations (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4967372A (en) * 1986-05-16 1990-10-30 The United States Of America As Represented By The Department Of Health And Human Services Automatic orientation and interactive addressing of display
US5504853A (en) * 1991-08-24 1996-04-02 International Business Machines Corporation System and method for selecting symbols and displaying their graphics objects in a detail window
US5640176A (en) * 1992-01-24 1997-06-17 Compaq Computer Corporation User interface for easily setting computer speaker volume and power conservation levels
US5724074A (en) * 1995-02-06 1998-03-03 Microsoft Corporation Method and system for graphically programming mobile toys
US5835923A (en) * 1995-03-16 1998-11-10 Hitachi, Ltd. Contents information transmitting/viewing system and method therefor
US6046745A (en) * 1996-03-25 2000-04-04 Hitachi, Ltd. Three-dimensional model making device and its method
US6061666A (en) * 1996-12-17 2000-05-09 Citicorp Development Center Automatic bank teller machine for the blind and visually impaired
US6118442A (en) * 1993-10-20 2000-09-12 Matsushita Electric Industrial Co., Ltd. Graphical-interactive-screen display apparatus and peripheral units
US6169535B1 (en) * 1997-06-30 2001-01-02 Toshiba America Information Systems, Inc. Monitor adjustment control
US6181342B1 (en) * 1998-07-06 2001-01-30 International Business Machines Corp. Computer file directory system displaying visual summaries of visual data in desktop computer documents for quickly identifying document content
US6195082B1 (en) * 1998-03-31 2001-02-27 International Business Machines Corporation Low noise circuit board for trackpoint pointing device
US6233591B1 (en) * 1996-05-06 2001-05-15 Adobe Systems Incorporated Dropping hyperlink onto document object
US20020035560A1 (en) * 1998-06-29 2002-03-21 Masahiro Sone System and method for adaptively configuring a shopping display in response to a recognized customer profile
US20040119715A1 (en) * 2002-12-18 2004-06-24 Microsoft Corporation International font measurement system and method
US20040119714A1 (en) * 2002-12-18 2004-06-24 Microsoft Corporation International automatic font size system and method
US6850245B1 (en) * 1999-04-07 2005-02-01 Fujitsu Limited Display characteristics recognition apparatus, display characteristics recognition program storage medium, computer system, display characteristics adjusting apparatus and display characteristics adjusting program storage medium
US20060280338A1 (en) * 2005-06-08 2006-12-14 Xerox Corporation Systems and methods for the visually impared
US7198369B2 (en) * 2000-12-26 2007-04-03 Yan Chen Method and device for detecting/correcting color vision and their application
US20070150829A1 (en) * 2005-12-22 2007-06-28 Xerox Corporation System and method for dynamic zoom to view documents on small displays
US20070159470A1 (en) * 2006-01-11 2007-07-12 Industrial Technology Research Institute Apparatus for automatically adjusting display parameters relying on visual performance and method for the same
US20070211047A1 (en) * 2006-03-09 2007-09-13 Doan Christopher H Persistent authenticating system and method to map real world object presence into virtual world object awareness
US20070236656A1 (en) * 2006-04-06 2007-10-11 Jeong Young-Min Method of modifying color composition for a color-blind person in a mobile displaying apparatus
US20080034316A1 (en) * 2006-08-01 2008-02-07 Johan Thoresson Scalable scrollbar markers
US20080242950A1 (en) * 2007-03-30 2008-10-02 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational user-health testing
US20100103197A1 (en) * 2008-10-27 2010-04-29 Hong Fu Jin Precision Industry (Shenzhen) Co., Ltd Method for adjusting font size on screen
US20100125561A1 (en) * 2008-04-24 2010-05-20 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US20110093438A1 (en) * 2006-05-26 2011-04-21 Poulsen Andrew S Meta-configuration of profiles
US20110149059A1 (en) * 2009-12-23 2011-06-23 Motorola, Inc. Method and Device for Visual Compensation
US7971155B1 (en) * 2006-10-22 2011-06-28 Hyoungsoo Yoon Dropdown widget
US20110157180A1 (en) * 2009-12-24 2011-06-30 Microsoft Corporation Virtual vision correction for video display
US20110164188A1 (en) * 2009-12-31 2011-07-07 Broadcom Corporation Remote control with integrated position, viewer identification and optical and audio test
US20120016678A1 (en) * 2010-01-18 2012-01-19 Apple Inc. Intelligent Automated Assistant
US20120246678A1 (en) * 2011-03-24 2012-09-27 Tobe Barksdale Distance Dependent Scalable User Interface
US20120254779A1 (en) * 2011-04-01 2012-10-04 Arthur Austin Ollivierre System and method for displaying objects in a user interface based on a visual acuity of a viewer
US20130055164A1 (en) * 2011-08-24 2013-02-28 Sony Ericsson Mobile Communications Ab System and Method for Selecting Objects on a Touch-Sensitive Display of a Mobile Communications Device

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4884199A (en) 1987-03-02 1989-11-28 International Business Macines Corporation User transaction guidance
US7357312B2 (en) 1998-05-29 2008-04-15 Gangi Frank J System for associating identification and personal data for multiple magnetic stripe cards or other sources to facilitate a transaction and related methods
US6386707B1 (en) 1999-11-08 2002-05-14 Russell A. Pellicano Method for evaluating visual acuity over the internet
US7088462B2 (en) 2001-06-29 2006-08-08 International Business Machines Corporation Print manager having a user interface for specifying how documents are directed to print devices
GB2454033A (en) 2007-10-24 2009-04-29 Plastic Logic Ltd Portable paperless electronic printer
US20120250039A1 (en) 2011-04-01 2012-10-04 Arthur Austin Ollivierre System and method for presenting information to a user

Patent Citations (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4967372A (en) * 1986-05-16 1990-10-30 The United States Of America As Represented By The Department Of Health And Human Services Automatic orientation and interactive addressing of display
US5504853A (en) * 1991-08-24 1996-04-02 International Business Machines Corporation System and method for selecting symbols and displaying their graphics objects in a detail window
US5640176A (en) * 1992-01-24 1997-06-17 Compaq Computer Corporation User interface for easily setting computer speaker volume and power conservation levels
US5903254A (en) * 1992-01-24 1999-05-11 Compaq Computer Corporation User interface for easily setting computer speaker volume and power conservation levels
US6118442A (en) * 1993-10-20 2000-09-12 Matsushita Electric Industrial Co., Ltd. Graphical-interactive-screen display apparatus and peripheral units
US5724074A (en) * 1995-02-06 1998-03-03 Microsoft Corporation Method and system for graphically programming mobile toys
US5835923A (en) * 1995-03-16 1998-11-10 Hitachi, Ltd. Contents information transmitting/viewing system and method therefor
US6046745A (en) * 1996-03-25 2000-04-04 Hitachi, Ltd. Three-dimensional model making device and its method
US6233591B1 (en) * 1996-05-06 2001-05-15 Adobe Systems Incorporated Dropping hyperlink onto document object
US6061666A (en) * 1996-12-17 2000-05-09 Citicorp Development Center Automatic bank teller machine for the blind and visually impaired
US6169535B1 (en) * 1997-06-30 2001-01-02 Toshiba America Information Systems, Inc. Monitor adjustment control
US6195082B1 (en) * 1998-03-31 2001-02-27 International Business Machines Corporation Low noise circuit board for trackpoint pointing device
US20020035560A1 (en) * 1998-06-29 2002-03-21 Masahiro Sone System and method for adaptively configuring a shopping display in response to a recognized customer profile
US6181342B1 (en) * 1998-07-06 2001-01-30 International Business Machines Corp. Computer file directory system displaying visual summaries of visual data in desktop computer documents for quickly identifying document content
US6850245B1 (en) * 1999-04-07 2005-02-01 Fujitsu Limited Display characteristics recognition apparatus, display characteristics recognition program storage medium, computer system, display characteristics adjusting apparatus and display characteristics adjusting program storage medium
US7198369B2 (en) * 2000-12-26 2007-04-03 Yan Chen Method and device for detecting/correcting color vision and their application
US20040119714A1 (en) * 2002-12-18 2004-06-24 Microsoft Corporation International automatic font size system and method
US20040119715A1 (en) * 2002-12-18 2004-06-24 Microsoft Corporation International font measurement system and method
US20060280338A1 (en) * 2005-06-08 2006-12-14 Xerox Corporation Systems and methods for the visually impared
US20070150829A1 (en) * 2005-12-22 2007-06-28 Xerox Corporation System and method for dynamic zoom to view documents on small displays
US20070159470A1 (en) * 2006-01-11 2007-07-12 Industrial Technology Research Institute Apparatus for automatically adjusting display parameters relying on visual performance and method for the same
US20070211047A1 (en) * 2006-03-09 2007-09-13 Doan Christopher H Persistent authenticating system and method to map real world object presence into virtual world object awareness
US7843471B2 (en) * 2006-03-09 2010-11-30 International Business Machines Corporation Persistent authenticating mechanism to map real world object presence into virtual world object awareness
US20070236656A1 (en) * 2006-04-06 2007-10-11 Jeong Young-Min Method of modifying color composition for a color-blind person in a mobile displaying apparatus
US8345338B2 (en) * 2006-04-06 2013-01-01 Samsung Electronics Co., Ltd. Method of modifying color composition for a color-blind person in a mobile displaying apparatus
US20110093438A1 (en) * 2006-05-26 2011-04-21 Poulsen Andrew S Meta-configuration of profiles
US20080034316A1 (en) * 2006-08-01 2008-02-07 Johan Thoresson Scalable scrollbar markers
US7971155B1 (en) * 2006-10-22 2011-06-28 Hyoungsoo Yoon Dropdown widget
US20080242950A1 (en) * 2007-03-30 2008-10-02 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational user-health testing
US20100125561A1 (en) * 2008-04-24 2010-05-20 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Computational system and method for memory modification
US20100103197A1 (en) * 2008-10-27 2010-04-29 Hong Fu Jin Precision Industry (Shenzhen) Co., Ltd Method for adjusting font size on screen
US20110149059A1 (en) * 2009-12-23 2011-06-23 Motorola, Inc. Method and Device for Visual Compensation
US20110157180A1 (en) * 2009-12-24 2011-06-30 Microsoft Corporation Virtual vision correction for video display
US20110164188A1 (en) * 2009-12-31 2011-07-07 Broadcom Corporation Remote control with integrated position, viewer identification and optical and audio test
US20120016678A1 (en) * 2010-01-18 2012-01-19 Apple Inc. Intelligent Automated Assistant
US20120246678A1 (en) * 2011-03-24 2012-09-27 Tobe Barksdale Distance Dependent Scalable User Interface
US20120254779A1 (en) * 2011-04-01 2012-10-04 Arthur Austin Ollivierre System and method for displaying objects in a user interface based on a visual acuity of a viewer
US20130055164A1 (en) * 2011-08-24 2013-02-28 Sony Ericsson Mobile Communications Ab System and Method for Selecting Objects on a Touch-Sensitive Display of a Mobile Communications Device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Windows 7 ClearType Text Tuner Makes Text Appear More Sharp And Clear - 7-27-2009 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105664433A (en) * 2016-01-20 2016-06-15 毕宏生 Vision comprehensive training device
US20180376212A1 (en) * 2017-06-23 2018-12-27 Sony Corporation Modifying display region for people with vision impairment
US10650702B2 (en) 2017-07-10 2020-05-12 Sony Corporation Modifying display region for people with loss of peripheral vision
US10805676B2 (en) 2017-07-10 2020-10-13 Sony Corporation Modifying display region for people with macular degeneration
US10845954B2 (en) 2017-07-11 2020-11-24 Sony Corporation Presenting audio video display options as list or matrix
US20190166343A1 (en) * 2017-11-27 2019-05-30 International Business Machines Corporation Optimized computer display rendering for user vision conditions
US10536678B2 (en) * 2017-11-27 2020-01-14 International Business Machines Corporation Optimized computer display rendering for user vision conditions
US11044449B2 (en) * 2017-11-27 2021-06-22 International Business Machines Corporation Optimized computer display rendering for user vision conditions
US10441845B2 (en) 2018-01-06 2019-10-15 Mkh Llc Exercise assembly for squats

Also Published As

Publication number Publication date
US8881058B2 (en) 2014-11-04
US20120254779A1 (en) 2012-10-04

Similar Documents

Publication Publication Date Title
US8881058B2 (en) System and method for displaying objects in a user interface based on a visual acuity of a viewer
US20210168144A1 (en) Interactive user-interface to adjust access privileges
EP3786935B1 (en) Method for controlling frame refresh rate of screen, device and storage medium
US9813461B2 (en) Viewing full screen applications in a sharing session
CN105938699B (en) Electronic device and method for reducing aging
US9727301B2 (en) Gesture-based prioritization of graphical output on remote displays
US8763055B1 (en) Cross-platform video display
CN103513859A (en) Icon display method and icon display device
US8776152B1 (en) Cloud-based cross-platform video display
US8887059B2 (en) Method and apparatus of locally controlling display content of a remote system
US20110145730A1 (en) Utilization of Browser Space
CN105700766B (en) User interface display system and method
WO2012135368A1 (en) System and method for displaying objects in a user interface based on a visual acuity of a viewer
US10564829B2 (en) Optimizing window resize actions for remoted applications
US10547711B2 (en) Using off-screen user interface data during remote sessions
US9699247B2 (en) User experience monitoring for application remoting
US9250782B1 (en) Using split windows for cross-platform document views
US9367931B1 (en) Motion vectors for cross-platform display
CN107562395B (en) Mobile terminal, display data transmission processing method and storage medium
CN106406671A (en) Icon display control method and mobile terminal
CN112099886A (en) Desktop display control method and device of mobile zero terminal
US10997356B1 (en) Computing device user interface having auto-grouping of data columns
US10845983B2 (en) Virtual multi-display
TWI658811B (en) Visual state detection method
US20160292640A1 (en) Compare profiles visual control

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION