US20110018817A1 - Touchpad-enabled remote controller and user interaction methods - Google Patents
Touchpad-enabled remote controller and user interaction methods Download PDFInfo
- Publication number
- US20110018817A1 US20110018817A1 US12/666,951 US66695108A US2011018817A1 US 20110018817 A1 US20110018817 A1 US 20110018817A1 US 66695108 A US66695108 A US 66695108A US 2011018817 A1 US2011018817 A1 US 2011018817A1
- Authority
- US
- United States
- Prior art keywords
- remote control
- control unit
- user
- sensors
- electronic equipment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08C—TRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
- G08C17/00—Arrangements for transmitting signals characterised by the use of a wireless electrical link
- G08C17/02—Arrangements for transmitting signals characterised by the use of a wireless electrical link using a radio link
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/038—Indexing scheme relating to G06F3/038
- G06F2203/0381—Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
-
- G—PHYSICS
- G08—SIGNALLING
- G08C—TRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
- G08C2201/00—Transmission systems of control signals via wireless link
- G08C2201/30—User interface
- G08C2201/32—Remote control based on movements, attitude of remote control device
Definitions
- the present invention takes a fresh look at the remote control of consumer electronic products. It replaces the large number of buttons with a simple handheld remote control unit that includes at least one touchpad together with a rich array of additional sensors, such as acceleration sensors, case perimeter sensors, pressure sensors, RF signal sensors, and the like.
- the remote control system responds to signals from this rich array of sensors using pattern recognition technology that allows the system to control many complex functions within consumer electronic equipment based on the manipulation and gestural movement of the remote control by the user.
- the pattern recognition system is adaptive. It can identify different users by the manner in which the remote control system is utilized and it can adapt so that each user can manipulate the system in his or her own unique way, and still effect control over the various functions of the consumer electronic equipment.
- the remote control unit itself, can also respond to the sensory input “meaning” to after the manner in which the remote control unit behaves. In this way, visual, audible or tactile cues, as well as power conservation strategies, can be controlled and revised based on the sensory input data.
- FIG. 1 a is a perspective view of the remote control unit.
- FIG. 1 b is a plan view of the remote control unit.
- FIG. 1 c is a view of the remote control unit in a portrait orientation.
- FIG. 1 d is a view of the remote control unit in a landscape orientation.
- FIG. 2 is a system block diagram illustrating the remote control system in operation by a user to control a piece of consumer electronic equipment.
- FIG. 3 is a block diagram illustrating an exemplary embodiment of the remote control system, including components associated with the control circuit coupled to the consumer electronic equipment and associated with the remote control unit.
- FIGS. 1 a and 1 b the remote control unit 20 of the remote control system has been illustrated.
- This remote control unit 20 interacts with a control circuit that is coupled to the consumer electronic equipment.
- the control circuit and consumer electronic equipment have not been showed in FIGS. 1 a - 1 d but are shown in subsequent FIGS. 2 and 3 .
- the remote control unit 20 has a touchpad 22 that may include predefined clickable regions, such as the up-down-left-right-okay region 24 , the channel up-down region 26 , the volume up-down region 28 and the mute region 30 .
- predefined clickable regions are merely exemplary of the basic concept that the touch screen can have regions that respond to pressure as a way of signifying that the user has “selected” a particular function. While the basic design of the remote control unit strives to eliminate physical push buttons to a large extent, the remote control unit 20 may still have physical push buttons if desired. Thus, for illustration purposed, four push buttons are shown at 32 , 33 , 34 and 35 . It is also contemplated that the touchpad 22 may be split into two distinct zones with or without a physical divider interposed between the two zones.
- the pre-defined clickable regions may be visually designated on the touchpad surface by either silk screening the region graphics onto the surface of the touchpad 22 , or by using a see-through graphic with backlighting. As will be more fully discussed below, the backlighting can be triggered by the appropriate combination of sensory inputs as recognized by the pattern recognizer also discussed below. It is contemplated that the touchpad surface may not include any pre-defined clickable regions.
- the case of the remote control unit 20 is preferably provided with a series of capacitive sensors, such as sensors 36 around the horizontal side walls of the case perimeter.
- the capacitive sensors 36 are example of case perimeter sensors. Capacitive sensors 36 can also be at other locations, such as on the underside of the case. These sensors detect how the user is holding the remote control. In this regard, different users may grip the remote control in different ways and the capacitive sensors 36 are arranged to be able to discriminate these different ways of holding the remote control.
- the pattern recognition system discussed below, can use this information to recognize these subtle differences.
- the sensors in cooperation with the pattern recognition system enable a user to operate the remote independently of how the remote is being held.
- FIG. 2 illustrates the remote control unit 20 being manipulated by a user 40 to operate a consumer electronic equipment component 48 having a display screen 50 .
- the consumer electronic equipment 48 conventionally has its own electronics that are used to provide the equipment with its normal functionality.
- such functionality includes displaying audio visual material on the display screen 50 .
- This material may include, for example, television programs, pre-recorded content, internet content and the like.
- the associated electronics of the consumer electronic equipment 48 have been illustrated separately at 52 .
- Embedded within the electronics package 52 is a control circuit shown diagrammatically at 60 that defines part of the remote control system. Control circuit 60 is coupled to the consumer electronic equipment 48 and responds to commands sent from the remote control unit 20 to control the operation of the consumer electronic equipment 48 .
- the remote control system is made up of the remote control 20 and the control circuit 60 . Together, these two components implement a sophisticated sensory input detecting and pattern recognizing system that allows the user 40 to control operations of the consumer electronic equipment 48 using a rich variety of finger, hand, wrist, arm and body movements.
- the system may be viewed as effecting a dialogue between the remote control unit 20 and the control circuit 60 , where that dialogue is expressed using a vocabulary and grammar associated with a diverse variety of different sensory inputs, (e.g., from the touchpad 22 , accelerometer 114 , case perimeter, sensor, pressure sensors, RF signal sensors and the like).
- the control system also includes a feedback loop through the user 40 .
- the user 40 has his or her own set of user sensory inputs 62 (sight, sound, touch) and the user 40 manipulates the remote control unit 20 based, in part, on audible and visual information obtained from the consumer electronic equipment 48 , and on visual, audible and tactile information from the remote control unit 20 .
- the remote control system supports a dialogue between remote control unit 20 and control circuit 60 , with a concurrent dialogue between user 40 , the control system and the consumer electronic equipment 48 .
- FIG. 2 thus illustrates that user 40 may receive visual, audible or tactile feedback from remote control 20 and this may be performed concurrently while viewing the display screen 50 .
- the information acquired by user 40 are depicted diagrammatically as user sensory inputs 62 .
- the sensory inputs acquired by the control system has been diagrammatically illustrated at 64 .
- the relationship between the control system sensory inputs 64 and the user sensory inputs 62 is a non-trivial one.
- the user 40 will manipulate the remote control unit 20 , in part, based on what the user 40 is trying to accomplish and also, in part, based on what the user 40 sees on the display screen 50 and what the user 40 also senses audibly, visually or tactilely from the remote control unit 20 and/or consumer electronic equipment 48 .
- the consumer electronic equipment 48 is a television set that has been programmed to block certain channels from being viewed by young children. In order to bypass the parental blocking feature, user 40 must manipulate the remote control unit 20 in a predefined way.
- the parental blocking unlocking feature can be changed each time it is used.
- the adult user must watch what is shown on the display screen 50 in order to learn how to manipulate the control unit to unlock the parental blocking feature.
- the instructions on the display are presented in a form, such as textual instructions, that a young child is not able to read.
- the control of the parental blocking feature relies on a particular manipulation (e.g., flick the wrist three times) that is context-based. A later unlocking operation would be treated as a different context and would potentially have a different gestural command to effect unlocking.
- the example illustrates that the behavior of the remote control system is context-dependent and that the user's sensory perception (e.g., reading the screen, feeling tactile vibrations, hearing particular sounds) will affect how the user's manipulations of the remote control unit 20 are interpreted.
- the user's sensory perception e.g., reading the screen, feeling tactile vibrations, hearing particular sounds
- the control system is able to make sense of a rich and diverse collection of sensory inputs using a pattern recognizer 70 and associated control logic 72 .
- sensory inputs are collected as a temporal sequence from the various sensors within the remote control unit 20 .
- the sensors may include at least one touchpad 22 responsive to manipulation by a user's fingers and at least one additional sensor such as, for example, an acceleration sensor responsive to movement of the remote control unit 20 , case perimeter sensors such as capacitive sensors 36 that discriminate which parts of the case are in contact with the user's body, pressure sensors responsive to pressing forces upon a predetermined region of the touchpad 22 and RF signal sensors responsive to radio frequency signals transmitted from the control circuit 60 .
- the temporal sequence of sensory inputs is fed to the pattern recognizer 70 .
- the pattern recognizer 70 is configured to classify the received sensory input message according to a predetermined recognition scheme to generate message meaning data that are then sent to the control logic 72 .
- the pattern recognizer 70 classifies a received sensory input message based on context information extracted from knowledge of what is being displayed on the display screen 50 of the electronic equipment.
- the control logic 72 is responsive to message meaning data to generate at least one command that is communicated to the control circuit 60 . The operations are described below in more detail.
- the control logic 72 decodes the message meaning data and generates a device control signal.
- the device control signal may be supplied to the remote control unit 20 itself, to effect control over the behavior of the remote control unit 20 (e.g., putting the unit to sleep or waking the unit up) or the device control signal may be sent to and/or used by the control circuit 60 , where it is passed on to the consumer electronic equipment 48 as a command to control the operation of the consumer electronic equipment 48 .
- the pattern recognizer 70 and the control logic 72 may be implemented separately or together and may be deployed in the control circuit 60 , in the remote control 20 , or distributed across both.
- the pattern recognizer 70 employs a trained model that may be adaptively altered or customized to more closely fit each user's style of using the remote control unit 20 .
- the trained model is used to classify the received sensory input message.
- the trained model is adaptable based on interaction with the user 40 .
- the pattern recognizer 70 is preferably provided with an initial set of models that classify certain operations as being mapped onto certain commands or control functions. For example, with reference to FIG. 1 b , an upward sliding motion of the fingertip on channel up-down region 26 might launch a forward channel scanning mode, whereas a single click or finger press upon the upward arrow of the region 26 would simply increment the channel by one.
- This behavior might be classified differently, however, if the remote control unit 20 is positioned in landscape orientation as illustrated in FIG. 1 d .
- the channel up-down region 26 might perform a function entirely unrelated to channel selection.
- the preferred embodiment includes a sensory input mechanism to allow the user 40 to inject a meta command—to let the system know that the user 40 wishes to after the pattern recognition models either for himself or herself, or for all users.
- a rapid back and forth wrist motion might be used to inform the recognition system that the most recent pattern recognition conclusion was wrong and that a different behavior is desired.
- the user 40 has used the remote control unit 20 on a coffee table and then manipulates the channel up-down region 26 , causing the television to begin a channel-scanning mode.
- the user 40 would prefer that the channel scanning mode should not be initiated when the remote control unit 20 is resting on the coffee table (i.e., not being held). To change this behavior, the user 40 would pick up the remote control unit 20 and shake it back and forth in a “no” gesture. This would cause an on-screen prompt to appear on the television display screen 50 , instructing the user 40 how the most recent temporal sequence of sensory inputs can be modified in this context to result in a different device control signal outcome.
- the control system is able to interpret the meaning of user 40 manipulations and gestures that can be quite complex, thereby allowing the user 40 to interact in an intuitive or natural way that can be customized from user to user.
- the pattern recognizer 70 may be based on a statistical model where the control system sensory inputs 64 generate probability scores associated with a plurality of different meanings.
- the pattern recognizer 70 would (a) select the meaning with the highest score, if that score is above a predetermined probability threshold value and/or above the next-most value by a predetermined threshold, or (b) engage the user 40 in a dialogue on-screen to resolve which meaning was intended, if the preceding threshold conditions are not met. The results of such user interaction may then be used to fine tune or adapt the model so that the system learns what behavior is expected for subsequent use.
- FIG. 3 where a detailed description of the remote control unit 20 and control circuit hardware has been illustrated.
- the components associated with the control circuit are shown generally at 60 and the components associated with the remote control unit are shown generally at 20 .
- the consumer electronic equipment is shown at 48 .
- a first processor or CPU 80 is attached to a bus 82 , to which random access memory 84 and programmable nonvolatile random access memory 86 are attached.
- the random access memory 84 and the programmable nonvolatile random access memory 86 are memories for storing control system operation information.
- the first processor includes an input/output (I/O) module 88 that provides an I/O bus 90 to which an RF communication module 92 and consumer electronic product interface 94 are attached.
- the consumer electronic product interface 94 couples to the remaining circuitry of the consumer electronic equipment 48 .
- the radio frequency communication module 92 includes an antenna and is designed to communicate with a corresponding communication module associated with the remote control unit 20 .
- control circuit 60 selectively provides said operation information to the consumer electronic equipment 48 for presentation to the user 40 .
- control circuit 60 provides the control system operation information to the consumer electronic equipment 48 in accordance with said at least one command generated by the control logic 72 .
- the control system operation information may be integrated with different message meaning data.
- the control circuit 60 selectively provides the operation information to the consumer electronic equipment 48 in accordance with the message meaning data generated by the pattern recognizer 70 .
- the remote control unit 20 has a second processor 96 with associated bus 98 , random access memory 99 and nonvolatile programmable random access memory 100 .
- the processor 96 also has an I/O module 102 that supports an I/O bus 104 to which a variety of sensors and other devices may be attached. Attached to the I/O bus 104 is the RF communication module 106 that communicates with its counterpart module 92 of the control circuit 60 .
- the display illumination device 108 is also coupled to the I/O bus 104 so that the backlighting can be switched on and off to render any backlight graphical elements on the touchpad 22 visible or invisible.
- a tactile feedback annunciator/speaker 110 is coupled to the I/O bus.
- the annunciator/speaker 110 may be activated to produce tactile feedback (vibrations) as well as audible tones.
- the above-described operations performed by the pattern recognizer 70 and the control logic 72 are controlled by the second processor 96 , the random access memory 99 , the nonvolatile programmable random access memory 100 , and the like included in the remote control unit 20 .
- the remote control unit 20 includes an assortment of different sensors. These include the touchpad 22 or touchpads 22 , a button pad membrane switch assembly 112 , accelerometer 114 and capacitive sensors 36 .
- the button pad membrane switch assembly 112 may be physically disposed beneath the touchpads 22 so that pressure upon the touchpad 22 will effect a switch state change from off to on. If desired, the button pad membrane switch assembly 112 may employ pressure-sensitive switches that can register a range of pressures, as opposed to a simple on/off binary state.
- the power supply 200 includes a removable battery 202 as well as a power management circuit 204 .
- the power management circuit 204 supplies power to the second processor 96 and to all of the modules within the remote control unit 20 requiring power. Such modules include all of the sensors, display illumination 108 , and annunciator/speaker 110 attached to the I/O bus 104 .
- an RFID tag 206 may be included in the remote control unit circuitry. The RFID tag 206 can be used to help locate the remote control from the control circuit 60 in the event the remote control unit 20 is lost.
- the touchpad 22 is divided into plural separate regions of interaction, such that each region of interaction is interpreted to have a different meaning by the pattern recognizer 70 .
- the touchpad 22 performs the following operations.
- the touchpad sensor can be segmented to provide several different intuitive zones of interaction.
- the touchpad 22 is also clickable by virtue of the button pad membrane switch assembly 112 located beneath or embedded within it.
- the clickable touchpad 22 can register pressure information and react to pressure (both mechanically and electrically) by sending a specific signal while providing sufficient haptic feedback to the user 40 such as through vibrations and sounds via the annunciator/speaker 110 .
- the touchpad 22 allows for the use of at least two contact points simultaneously. (e.g., two finger input) such as one contact point per side of the pad.
- the touchpad 22 can be viewed as divided in two along a medial line (e.g., separating the right and left sides of the touchpad 22 when held in a landscape orientation).
- the touchpad 22 can thus be constructed using two single-position registering touchpads mounted side by side, or one single mufti-touch touchpad with the ability to register with equal precision (two points of contact at the same time).
- the remote control unit 20 may have a complement of physical buttons.
- four buttons 32 - 35 have been illustrated in FIGS. 1 a and 1 b .
- These physical buttons may be implemented using the same button pad membrane switch assembly 112 ( FIG. 3 ) embedded beneath the touchpad 22 .
- the physical buttons like the context-dependent virtual buttons on the touchpad surface, can be backlit to reveal button function names.
- the remote control unit 20 uses its pattern recognition system to interpret the sensory data.
- the remote control unit 20 further includes a case orientation sensor capable of differentiating between portrait orientation and landscape orientation.
- the detection of the orientations of the remote control unit 20 may be performed using the accelerometer 114 , the capacitive sensors 36 , and the like, without the case orientation sensor.
- Included in the sensory data are inputs from the accelerometer 114 or accelerometers 114 and the capacitive sensors 36 placed around the periphery and the bottom of the case.
- the user 40 will naturally turn the remote control unit 20 in his or her hands to best accommodate what he or she is trying to accomplish.
- the pattern recognition system interprets how the user 40 is holding the remote control unit 20 and redefines these zones of interaction so that they will appear to be at the same place, no matter how the remote is oriented.
- the remote control unit 20 can be used with one or two hands, and in both landscape and portrait orientation. This means that the pattern recognition system can detect: whether the user 40 is holding the remote control unit 20 by the right hand or by the left hand; by which hand the user 40 is operating the touchpad 22 ; which part of the remote control unit 20 the user 40 is holding; and the like.
- the regions of interaction are remapped onto different locations within the touchpad 22 based on case orientation. The following describes more details.
- the pattern recognition system can discriminate the difference and will automatically redefine the zones of interaction so that the user 40 can perform the most probably operations in the easiest manner for that user 40 .
- the zones of interaction include, for example, different zones within the touchpad 22 . Different regions of the touchpad 22 may be dedicated to different functions or different user manipulation styles.
- the remote control unit 20 itself can be manipulated into different virtual “zones of interaction” by employing different gestures with the remote in mid-air, such as a quick flick of the wrist to change channels.
- the remote control unit 20 includes power management system.
- the control logic 72 generates at least one command that is communicated to the power management system to after power management system behavior based on sensor input messages or said message meaning data.
- the power management circuit 204 , the power supply 200 , and the removable battery 202 shown in FIG. 3 form the power management system.
- the remote control unit 20 may run on a single AA or AAA battery or batteries for approximately one year.
- the power management system performs: placing of the sensors included in the remote control unit 20 , into a sleep mode; detecting movement of the remote control unit 20 by waking the acceleration sensor up in the sleep mode; determining whether or not the remote control unit 20 is held by the user 40 , by waking the case perimeter sensor up when the movement of the remote control unit 20 is detected; and releasing the sleep mode when the detecting detects that the remote control unit 20 is held by the user 40 .
- the power management system releases the sleep mode, by waking the touchpad 22 up and then waking said RF signal sensor up. The following describes these operations in more detail. With currently available technology, the wireless circuitry associated with RF modules consumes more power than the touch sensors; and the accelerometers 114 and actuators consume less power than the touch sensors.
- the power management circuitry 204 places the wireless circuitry in the sleep mode (or turned off altogether) after a short period of time after the remote control unit 20 is no longer being used (e.g., 30 seconds).
- the touch sensors will then be placed in sleep mode (or turned off) after a somewhat longer period of time (e.g., 2 minutes). This will allow turning on the wireless circuitry again (in case the user 40 touches the surface of the touchpad 22 or picks up the unit within two minutes).
- the accelerometers 114 are put into a low power mode where the circuitry checks the accelerometer status at a much lower rate than the normal accelerometer refresh rate.
- the normal refresh rate might be on the order of 50 Hz whereas in the low power mode the refresh rate might be in the order of 1 Hz, or even 0.1 Hz.
- the power management circuitry 204 would implement a turn on sequence that is essentially the reverse of the turn off sequence, with the accelerometer refresh rate being increased to full rate first, followed by reactivation of the touch sensors and finally by activation of the wireless circuitry.
- the RF modules may periodically be awakened, to check to see if there are any pending messages from the control circuit 60 .
- the remote control unit 20 does not have a dedicated power-on button, as this might be a potential source of user confusion as to whether such button powers on the remote control unit 20 or the television.
- the pattern recognition system is used to handle power-on in an efficient manner.
- the remote control unit 20 turns on when the user 40 first picks it up. For this reason, the system first checks the lower resolution acceleration data to determine if the remote has been moved. If so, the capacitive sensors 36 are next energized to determine if the remote is actually being held (as opposed to simply being inadvertently pushed or moved when resting on the coffee table). If the pattern recognition system determines that the remote control unit 20 is being held, then next the touchpads 22 and finally the wireless circuitry are activated.
- power-on can be triggered by a specific gesture, such as shaking the remote control unit 20 .
- More complex power-on operation can also be utilized, for example, to enforce parental control as discussed above in connection with parental blocking features.
- the pattern recognition system will likewise detect when it is time to turn the remote control unit 20 off by detecting inactivity, or if detecting that the television has been turned off. This latter event would be detectable, for example, by information communicated via the RF modules.
- the control circuit 60 may include a button that will send a remote location message to the remote control unit 20 . More specifically, the control circuit 60 sends a find-me message that causes the remote control unit 20 to perform an attention attracting operation useful in assisting the user 40 to find the remote control unit 20 when it has become misplaced. The following describes more details. The user 40 would push this button if the remote control unit 20 has gotten misplaced. The control circuit 60 would then periodically send a tell-me-where-you-are signal (the find-me message) to the remote via RF.
- the remote control unit's RF module When the remote control unit's RF module next wakes up and finds the wake up signal, it will activate the haptic feedback system (e.g., annunciator/speaker 110 ) causing the unit to make sound and/or vibrate and optionally use the display illumination circuitry 108 to turn the backlighting on.
- the remote control unit 20 and the control circuitry can use RF ranging functionality to measure the distance between the remote control unit 20 and the control circuit 60 . This information has been used to display the distance on the display screen 50 , or even present a picture of the room with highlighted areas identifying where the remote control unit 20 could be.
- the RFID tag 206 may be used, allowing the precise location of the remote control to be displayed on the display screen 50 .
- the remote control system is able to capitalize on its tight coupling with the on-screen information.
- the on-screen information such as instructions on how to deactivate the parental blocking feature, may be stored in the programmable random access memory 86 of the control circuit 60 ( FIG. 3 ) and may then be projected onto the display screen 50 as an overlay upon the presently viewed program.
- the user 40 does not need to look at the remote control unit 20 in order to operate it.
- an overlay image of a keyboard may be presented and the user 40 can navigate to the desired keys by simply manipulating the touch screen while watching a cursor or cursors (one for each finger) on the displayed overlay keyboard.
- the remote control system circuitry can also obtain program guide information and the display overlay can then allow the user 40 to select which programs to view or record by simply manipulating the touch screen.
- the remote control system can use the display screen 50 , with its high resolution graphics capability, to provide an unlimited amount of visual information to the user 40 which would be virtually impossible to provide through a set of dedicated buttons as conventional controllers do.
- the rich collection of diverse sensory inputs allows the user 40 to adopt many different, and even redundant, ways of communicating the user's desires to the system.
- Interpretation of the diverse collection of sensory inputs by the pattern recognizer 70 handles much of the complexity of converting the user's gestural and touch commands into message meaning data that correlate to functions that the consumer electronic equipment 48 can perform.
- the resulting division of labor produces a control system that provides both a very high, visually engaging information content to the user 40 regarding his or her control system choices, with an equally rich collection of gestural and touch commands that the user 40 can employ to get his or her message across to the control system. Compare this to the conventional push button remote control that requires one button, or a sequence of buttons, to be pressed for each desired function, with the added inconvenience that the user 40 must look at the remote control in order to find the desired button to push.
Abstract
The hand held case of the remote control unit includes at least one touchpad, and other sensors, such as acceleration sensors, case perimeter sensors, pressure sensors, RF signal sensors. These sensors provide a rich array of sensory inputs that are classified by a pattern recognizer to generate control commands for both the consumer electronic equipment and the remote control unit itself. A power management system to conserve unit battery power is also responsive to the pattern recognizer to allow intelligent power management control. The control system uses the display of the consumer electronic equipment to provide instructions to the user, and the behavior of the remote control system uses what is displayed on the display as context information for pattern recognition.
Description
- This application claims the benefit of U.S. Patent Applications No. 60/946,804 and Ser. No. 11/977,348 filed on Jun. 28, 2007 and Oct. 24, 2007, respectively. The disclosures of the above applications are incorporated herein by reference.
- Practically all consumer electronic products in use today come with a remote control. In most cases, the remote control has many buttons, each dedicated to the control of one or more specific features of the consumer electronics product. As these products increase in complexity, so does the number of buttons required. At some point, the increased number of buttons renders the remote control mostly useless for a large number of users. Various solutions have been proposed to address such problems. One solution is disclosed in U.S. Pat. No. 6,765,557 to use a touchpad for controlling a home entertainment device such as an Interactive television. Although the use of the touchpad prevents the increase of buttons and thereby reduces the problem of user inconvenience, this is not enough to solve the problem completely.
- The present invention takes a fresh look at the remote control of consumer electronic products. It replaces the large number of buttons with a simple handheld remote control unit that includes at least one touchpad together with a rich array of additional sensors, such as acceleration sensors, case perimeter sensors, pressure sensors, RF signal sensors, and the like. The remote control system responds to signals from this rich array of sensors using pattern recognition technology that allows the system to control many complex functions within consumer electronic equipment based on the manipulation and gestural movement of the remote control by the user. The pattern recognition system is adaptive. It can identify different users by the manner in which the remote control system is utilized and it can adapt so that each user can manipulate the system in his or her own unique way, and still effect control over the various functions of the consumer electronic equipment.
- The remote control unit, itself, can also respond to the sensory input “meaning” to after the manner in which the remote control unit behaves. In this way, visual, audible or tactile cues, as well as power conservation strategies, can be controlled and revised based on the sensory input data.
- For a more complete understanding of the invention, refer to the following description and to the accompanying drawings.
-
FIG. 1 a is a perspective view of the remote control unit. -
FIG. 1 b is a plan view of the remote control unit. -
FIG. 1 c is a view of the remote control unit in a portrait orientation. -
FIG. 1 d is a view of the remote control unit in a landscape orientation. -
FIG. 2 is a system block diagram illustrating the remote control system in operation by a user to control a piece of consumer electronic equipment. -
FIG. 3 is a block diagram illustrating an exemplary embodiment of the remote control system, including components associated with the control circuit coupled to the consumer electronic equipment and associated with the remote control unit. - Referring first to
FIGS. 1 a and 1 b, theremote control unit 20 of the remote control system has been illustrated. Thisremote control unit 20 interacts with a control circuit that is coupled to the consumer electronic equipment. The control circuit and consumer electronic equipment have not been showed inFIGS. 1 a-1 d but are shown in subsequentFIGS. 2 and 3 . - The
remote control unit 20 has atouchpad 22 that may include predefined clickable regions, such as the up-down-left-right-okay region 24, the channel up-downregion 26, the volume up-downregion 28 and themute region 30. It will be understood that these predefined clickable regions are merely exemplary of the basic concept that the touch screen can have regions that respond to pressure as a way of signifying that the user has “selected” a particular function. While the basic design of the remote control unit strives to eliminate physical push buttons to a large extent, theremote control unit 20 may still have physical push buttons if desired. Thus, for illustration purposed, four push buttons are shown at 32, 33, 34 and 35. It is also contemplated that thetouchpad 22 may be split into two distinct zones with or without a physical divider interposed between the two zones. - The pre-defined clickable regions may be visually designated on the touchpad surface by either silk screening the region graphics onto the surface of the
touchpad 22, or by using a see-through graphic with backlighting. As will be more fully discussed below, the backlighting can be triggered by the appropriate combination of sensory inputs as recognized by the pattern recognizer also discussed below. It is contemplated that the touchpad surface may not include any pre-defined clickable regions. - The case of the
remote control unit 20 is preferably provided with a series of capacitive sensors, such assensors 36 around the horizontal side walls of the case perimeter. Thecapacitive sensors 36 are example of case perimeter sensors.Capacitive sensors 36 can also be at other locations, such as on the underside of the case. These sensors detect how the user is holding the remote control. In this regard, different users may grip the remote control in different ways and thecapacitive sensors 36 are arranged to be able to discriminate these different ways of holding the remote control. Although there may be subtle differences in how one user holds the remote control as compared with another, the pattern recognition system, discussed below, can use this information to recognize these subtle differences. Moreover, the sensors in cooperation with the pattern recognition system enable a user to operate the remote independently of how the remote is being held. - Referring now to
FIG. 2 , an overview of the pattern recognition system will be presented.FIG. 2 illustrates theremote control unit 20 being manipulated by auser 40 to operate a consumerelectronic equipment component 48 having adisplay screen 50. The consumerelectronic equipment 48 conventionally has its own electronics that are used to provide the equipment with its normal functionality. In the case of the illustratedcomponent 48 such functionality includes displaying audio visual material on thedisplay screen 50. This material may include, for example, television programs, pre-recorded content, internet content and the like. For illustration purposes, the associated electronics of the consumerelectronic equipment 48 have been illustrated separately at 52. Embedded within theelectronics package 52 is a control circuit shown diagrammatically at 60 that defines part of the remote control system.Control circuit 60 is coupled to the consumerelectronic equipment 48 and responds to commands sent from theremote control unit 20 to control the operation of the consumerelectronic equipment 48. - The remote control system is made up of the
remote control 20 and thecontrol circuit 60. Together, these two components implement a sophisticated sensory input detecting and pattern recognizing system that allows theuser 40 to control operations of the consumerelectronic equipment 48 using a rich variety of finger, hand, wrist, arm and body movements. The system may be viewed as effecting a dialogue between theremote control unit 20 and thecontrol circuit 60, where that dialogue is expressed using a vocabulary and grammar associated with a diverse variety of different sensory inputs, (e.g., from thetouchpad 22,accelerometer 114, case perimeter, sensor, pressure sensors, RF signal sensors and the like). The control system also includes a feedback loop through theuser 40. Theuser 40 has his or her own set of user sensory inputs 62 (sight, sound, touch) and theuser 40 manipulates theremote control unit 20 based, in part, on audible and visual information obtained from the consumerelectronic equipment 48, and on visual, audible and tactile information from theremote control unit 20. Thus, the remote control system supports a dialogue betweenremote control unit 20 andcontrol circuit 60, with a concurrent dialogue betweenuser 40, the control system and the consumerelectronic equipment 48. -
FIG. 2 thus illustrates thatuser 40 may receive visual, audible or tactile feedback fromremote control 20 and this may be performed concurrently while viewing thedisplay screen 50. For illustration purposes, the information acquired byuser 40 are depicted diagrammatically as usersensory inputs 62. Likewise, the sensory inputs acquired by the control system (from a diverse array of different types of sensors) has been diagrammatically illustrated at 64. - The relationship between the control system
sensory inputs 64 and the usersensory inputs 62 is a non-trivial one. Theuser 40 will manipulate theremote control unit 20, in part, based on what theuser 40 is trying to accomplish and also, in part, based on what theuser 40 sees on thedisplay screen 50 and what theuser 40 also senses audibly, visually or tactilely from theremote control unit 20 and/or consumerelectronic equipment 48. To illustrate this point, imagine that the consumerelectronic equipment 48 is a television set that has been programmed to block certain channels from being viewed by young children. In order to bypass the parental blocking feature,user 40 must manipulate theremote control unit 20 in a predefined way. To prevent the child from simply watching the parent and learning the manipulating technique, the parental blocking unlocking feature can be changed each time it is used. The adult user must watch what is shown on thedisplay screen 50 in order to learn how to manipulate the control unit to unlock the parental blocking feature. The instructions on the display are presented in a form, such as textual instructions, that a young child is not able to read. Thus, the control of the parental blocking feature relies on a particular manipulation (e.g., flick the wrist three times) that is context-based. A later unlocking operation would be treated as a different context and would potentially have a different gestural command to effect unlocking. Although this is but one example the example illustrates that the behavior of the remote control system is context-dependent and that the user's sensory perception (e.g., reading the screen, feeling tactile vibrations, hearing particular sounds) will affect how the user's manipulations of theremote control unit 20 are interpreted. - The control system is able to make sense of a rich and diverse collection of sensory inputs using a
pattern recognizer 70 and associatedcontrol logic 72. As theuser 40 manipulates theremote control unit 20, sensory inputs are collected as a temporal sequence from the various sensors within theremote control unit 20. As previously noted, the sensors may include at least onetouchpad 22 responsive to manipulation by a user's fingers and at least one additional sensor such as, for example, an acceleration sensor responsive to movement of theremote control unit 20, case perimeter sensors such ascapacitive sensors 36 that discriminate which parts of the case are in contact with the user's body, pressure sensors responsive to pressing forces upon a predetermined region of thetouchpad 22 and RF signal sensors responsive to radio frequency signals transmitted from thecontrol circuit 60. - The temporal sequence of sensory inputs is fed to the
pattern recognizer 70. Thepattern recognizer 70 is configured to classify the received sensory input message according to a predetermined recognition scheme to generate message meaning data that are then sent to thecontrol logic 72. Thepattern recognizer 70 classifies a received sensory input message based on context information extracted from knowledge of what is being displayed on thedisplay screen 50 of the electronic equipment. Thecontrol logic 72 is responsive to message meaning data to generate at least one command that is communicated to thecontrol circuit 60. The operations are described below in more detail. Thecontrol logic 72 decodes the message meaning data and generates a device control signal. The device control signal may be supplied to theremote control unit 20 itself, to effect control over the behavior of the remote control unit 20 (e.g., putting the unit to sleep or waking the unit up) or the device control signal may be sent to and/or used by thecontrol circuit 60, where it is passed on to the consumerelectronic equipment 48 as a command to control the operation of the consumerelectronic equipment 48. Thepattern recognizer 70 and thecontrol logic 72 may be implemented separately or together and may be deployed in thecontrol circuit 60, in theremote control 20, or distributed across both. - In one embodiment, the
pattern recognizer 70 employs a trained model that may be adaptively altered or customized to more closely fit each user's style of using theremote control unit 20. The trained model is used to classify the received sensory input message. Here, the trained model is adaptable based on interaction with theuser 40. In such trained model embodiment, thepattern recognizer 70 is preferably provided with an initial set of models that classify certain operations as being mapped onto certain commands or control functions. For example, with reference toFIG. 1 b, an upward sliding motion of the fingertip on channel up-down region 26 might launch a forward channel scanning mode, whereas a single click or finger press upon the upward arrow of theregion 26 would simply increment the channel by one. This behavior might be classified differently, however, if theremote control unit 20 is positioned in landscape orientation as illustrated inFIG. 1 d. For example, when in landscape orientation and held by two hands (as determined by the capacitive sensors 36), the channel up-down region 26 might perform a function entirely unrelated to channel selection. - To adapt the model for a particular user, the preferred embodiment includes a sensory input mechanism to allow the
user 40 to inject a meta command—to let the system know that theuser 40 wishes to after the pattern recognition models either for himself or herself, or for all users. For example, a rapid back and forth wrist motion (analogous to shaking one's head in a “no” gesture) might be used to inform the recognition system that the most recent pattern recognition conclusion was wrong and that a different behavior is desired. For example, assume that theuser 40 has used theremote control unit 20 on a coffee table and then manipulates the channel up-down region 26, causing the television to begin a channel-scanning mode. Perhaps theuser 40 would prefer that the channel scanning mode should not be initiated when theremote control unit 20 is resting on the coffee table (i.e., not being held). To change this behavior, theuser 40 would pick up theremote control unit 20 and shake it back and forth in a “no” gesture. This would cause an on-screen prompt to appear on thetelevision display screen 50, instructing theuser 40 how the most recent temporal sequence of sensory inputs can be modified in this context to result in a different device control signal outcome. - Because the
pattern recognizer 70 can respond to a rich variety of different types of sensory inputs, the control system is able to interpret the meaning ofuser 40 manipulations and gestures that can be quite complex, thereby allowing theuser 40 to interact in an intuitive or natural way that can be customized from user to user. In this regard, there may be instances where two or more gestural commands might be very similar and yet might have different meanings and thus might require different commands to be sent to the consumerelectronic equipment 48. To handle this, thepattern recognizer 70 may be based on a statistical model where the control systemsensory inputs 64 generate probability scores associated with a plurality of different meanings. The pattern recognizer 70 would (a) select the meaning with the highest score, if that score is above a predetermined probability threshold value and/or above the next-most value by a predetermined threshold, or (b) engage theuser 40 in a dialogue on-screen to resolve which meaning was intended, if the preceding threshold conditions are not met. The results of such user interaction may then be used to fine tune or adapt the model so that the system learns what behavior is expected for subsequent use. - With the above overview in mind, refer now to
FIG. 3 where a detailed description of theremote control unit 20 and control circuit hardware has been illustrated. InFIG. 3 , the components associated with the control circuit are shown generally at 60 and the components associated with the remote control unit are shown generally at 20. The consumer electronic equipment is shown at 48. - Beginning with the
control circuit 60, a first processor orCPU 80 is attached to abus 82, to whichrandom access memory 84 and programmable nonvolatilerandom access memory 86 are attached. Therandom access memory 84 and the programmable nonvolatilerandom access memory 86 are memories for storing control system operation information. The first processor includes an input/output (I/O)module 88 that provides an I/O bus 90 to which anRF communication module 92 and consumerelectronic product interface 94 are attached. The consumerelectronic product interface 94, in turn, couples to the remaining circuitry of the consumerelectronic equipment 48. The radiofrequency communication module 92 includes an antenna and is designed to communicate with a corresponding communication module associated with theremote control unit 20. With the above structure, thecontrol circuit 60 selectively provides said operation information to the consumerelectronic equipment 48 for presentation to theuser 40. Here, thecontrol circuit 60 provides the control system operation information to the consumerelectronic equipment 48 in accordance with said at least one command generated by thecontrol logic 72. The control system operation information may be integrated with different message meaning data. In this case, thecontrol circuit 60 selectively provides the operation information to the consumerelectronic equipment 48 in accordance with the message meaning data generated by thepattern recognizer 70. - The
remote control unit 20 has asecond processor 96 with associatedbus 98,random access memory 99 and nonvolatile programmablerandom access memory 100. Theprocessor 96 also has an I/O module 102 that supports an I/O bus 104 to which a variety of sensors and other devices may be attached. Attached to the I/O bus 104 is theRF communication module 106 that communicates with itscounterpart module 92 of thecontrol circuit 60. Thedisplay illumination device 108 is also coupled to the I/O bus 104 so that the backlighting can be switched on and off to render any backlight graphical elements on thetouchpad 22 visible or invisible. A tactile feedback annunciator/speaker 110 is coupled to the I/O bus. The annunciator/speaker 110 may be activated to produce tactile feedback (vibrations) as well as audible tones. The above-described operations performed by thepattern recognizer 70 and thecontrol logic 72 are controlled by thesecond processor 96, therandom access memory 99, the nonvolatile programmablerandom access memory 100, and the like included in theremote control unit 20. - As previously discussed, the
remote control unit 20 includes an assortment of different sensors. These include thetouchpad 22 ortouchpads 22, a button padmembrane switch assembly 112,accelerometer 114 andcapacitive sensors 36. The button padmembrane switch assembly 112 may be physically disposed beneath thetouchpads 22 so that pressure upon thetouchpad 22 will effect a switch state change from off to on. If desired, the button padmembrane switch assembly 112 may employ pressure-sensitive switches that can register a range of pressures, as opposed to a simple on/off binary state. - Because the
remote control unit 20 is designed to sit on the coffee table when not in use, abattery power supply 200 is preferred. Thus, thepower supply 200 includes aremovable battery 202 as well as apower management circuit 204. Thepower management circuit 204 supplies power to thesecond processor 96 and to all of the modules within theremote control unit 20 requiring power. Such modules include all of the sensors,display illumination 108, and annunciator/speaker 110 attached to the I/O bus 104. If desired, an RFID tag 206 may be included in the remote control unit circuitry. The RFID tag 206 can be used to help locate the remote control from thecontrol circuit 60 in the event theremote control unit 20 is lost. - The
touchpad 22 is divided into plural separate regions of interaction, such that each region of interaction is interpreted to have a different meaning by thepattern recognizer 70. In more detail, thetouchpad 22 performs the following operations. The touchpad sensor can be segmented to provide several different intuitive zones of interaction. Thetouchpad 22 is also clickable by virtue of the button padmembrane switch assembly 112 located beneath or embedded within it. Theclickable touchpad 22 can register pressure information and react to pressure (both mechanically and electrically) by sending a specific signal while providing sufficient haptic feedback to theuser 40 such as through vibrations and sounds via the annunciator/speaker 110. Thetouchpad 22 allows for the use of at least two contact points simultaneously. (e.g., two finger input) such as one contact point per side of the pad. Thetouchpad 22 can be viewed as divided in two along a medial line (e.g., separating the right and left sides of thetouchpad 22 when held in a landscape orientation). Thetouchpad 22 can thus be constructed using two single-position registering touchpads mounted side by side, or one single mufti-touch touchpad with the ability to register with equal precision (two points of contact at the same time). - <Physical Buttons>
- Although not required in all embodiments, the
remote control unit 20 may have a complement of physical buttons. In this regard, four buttons 32-35 have been illustrated inFIGS. 1 a and 1 b. These physical buttons may be implemented using the same button pad membrane switch assembly 112 (FIG. 3 ) embedded beneath thetouchpad 22. The physical buttons, like the context-dependent virtual buttons on the touchpad surface, can be backlit to reveal button function names. - <Redefining Regions of Interaction>
- To allow for natural operation, the
remote control unit 20 uses its pattern recognition system to interpret the sensory data. Theremote control unit 20 further includes a case orientation sensor capable of differentiating between portrait orientation and landscape orientation. The detection of the orientations of theremote control unit 20 may be performed using theaccelerometer 114, thecapacitive sensors 36, and the like, without the case orientation sensor. Included in the sensory data are inputs from theaccelerometer 114 oraccelerometers 114 and thecapacitive sensors 36 placed around the periphery and the bottom of the case. Theuser 40 will naturally turn theremote control unit 20 in his or her hands to best accommodate what he or she is trying to accomplish. The pattern recognition system interprets how theuser 40 is holding theremote control unit 20 and redefines these zones of interaction so that they will appear to be at the same place, no matter how the remote is oriented. For instance, theremote control unit 20 can be used with one or two hands, and in both landscape and portrait orientation. This means that the pattern recognition system can detect: whether theuser 40 is holding theremote control unit 20 by the right hand or by the left hand; by which hand theuser 40 is operating thetouchpad 22; which part of theremote control unit 20 theuser 40 is holding; and the like. The regions of interaction are remapped onto different locations within thetouchpad 22 based on case orientation. The following describes more details. The pattern recognition system can discriminate the difference and will automatically redefine the zones of interaction so that theuser 40 can perform the most probably operations in the easiest manner for thatuser 40. The zones of interaction include, for example, different zones within thetouchpad 22. Different regions of thetouchpad 22 may be dedicated to different functions or different user manipulation styles. In addition, theremote control unit 20 itself can be manipulated into different virtual “zones of interaction” by employing different gestures with the remote in mid-air, such as a quick flick of the wrist to change channels. - <Power Management>
- The presently preferred embodiment is contemplated for very low power consumption. The
remote control unit 20 includes power management system. Thecontrol logic 72 generates at least one command that is communicated to the power management system to after power management system behavior based on sensor input messages or said message meaning data. Thepower management circuit 204, thepower supply 200, and theremovable battery 202 shown inFIG. 3 form the power management system. For example, theremote control unit 20 may run on a single AA or AAA battery or batteries for approximately one year. The power management system performs: placing of the sensors included in theremote control unit 20, into a sleep mode; detecting movement of theremote control unit 20 by waking the acceleration sensor up in the sleep mode; determining whether or not theremote control unit 20 is held by theuser 40, by waking the case perimeter sensor up when the movement of theremote control unit 20 is detected; and releasing the sleep mode when the detecting detects that theremote control unit 20 is held by theuser 40. The power management system releases the sleep mode, by waking thetouchpad 22 up and then waking said RF signal sensor up. The following describes these operations in more detail. With currently available technology, the wireless circuitry associated with RF modules consumes more power than the touch sensors; and theaccelerometers 114 and actuators consume less power than the touch sensors. For this reason, thepower management circuitry 204 places the wireless circuitry in the sleep mode (or turned off altogether) after a short period of time after theremote control unit 20 is no longer being used (e.g., 30 seconds). The touch sensors will then be placed in sleep mode (or turned off) after a somewhat longer period of time (e.g., 2 minutes). This will allow turning on the wireless circuitry again (in case theuser 40 touches the surface of thetouchpad 22 or picks up the unit within two minutes). Theaccelerometers 114 are put into a low power mode where the circuitry checks the accelerometer status at a much lower rate than the normal accelerometer refresh rate. In this regard the normal refresh rate might be on the order of 50 Hz whereas in the low power mode the refresh rate might be in the order of 1 Hz, or even 0.1 Hz. Thepower management circuitry 204 would implement a turn on sequence that is essentially the reverse of the turn off sequence, with the accelerometer refresh rate being increased to full rate first, followed by reactivation of the touch sensors and finally by activation of the wireless circuitry. In the sleep mode, the RF modules may periodically be awakened, to check to see if there are any pending messages from thecontrol circuit 60. - In the presently preferred embodiment, the
remote control unit 20 does not have a dedicated power-on button, as this might be a potential source of user confusion as to whether such button powers on theremote control unit 20 or the television. - Thus, the pattern recognition system is used to handle power-on in an efficient manner. The
remote control unit 20 turns on when theuser 40 first picks it up. For this reason, the system first checks the lower resolution acceleration data to determine if the remote has been moved. If so, thecapacitive sensors 36 are next energized to determine if the remote is actually being held (as opposed to simply being inadvertently pushed or moved when resting on the coffee table). If the pattern recognition system determines that theremote control unit 20 is being held, then next thetouchpads 22 and finally the wireless circuitry are activated. - Alternatively, power-on can be triggered by a specific gesture, such as shaking the
remote control unit 20. More complex power-on operation can also be utilized, for example, to enforce parental control as discussed above in connection with parental blocking features. - The pattern recognition system will likewise detect when it is time to turn the
remote control unit 20 off by detecting inactivity, or if detecting that the television has been turned off. This latter event would be detectable, for example, by information communicated via the RF modules. - <Remote Finder>
- The
control circuit 60, associated with the consumerelectronic equipment 48, may include a button that will send a remote location message to theremote control unit 20. More specifically, thecontrol circuit 60 sends a find-me message that causes theremote control unit 20 to perform an attention attracting operation useful in assisting theuser 40 to find theremote control unit 20 when it has become misplaced. The following describes more details. Theuser 40 would push this button if theremote control unit 20 has gotten misplaced. Thecontrol circuit 60 would then periodically send a tell-me-where-you-are signal (the find-me message) to the remote via RF. When the remote control unit's RF module next wakes up and finds the wake up signal, it will activate the haptic feedback system (e.g., annunciator/speaker 110) causing the unit to make sound and/or vibrate and optionally use thedisplay illumination circuitry 108 to turn the backlighting on. In addition, if desired, theremote control unit 20 and the control circuitry can use RF ranging functionality to measure the distance between theremote control unit 20 and thecontrol circuit 60. This information has been used to display the distance on thedisplay screen 50, or even present a picture of the room with highlighted areas identifying where theremote control unit 20 could be. Alternatively, the RFID tag 206 may be used, allowing the precise location of the remote control to be displayed on thedisplay screen 50. - <Tight Coupling Between Remote Control System and On-Screen User Interface>
- As illustrated by the previously discussed example regarding parental control, the remote control system is able to capitalize on its tight coupling with the on-screen information. The on-screen information, such as instructions on how to deactivate the parental blocking feature, may be stored in the programmable
random access memory 86 of the control circuit 60 (FIG. 3 ) and may then be projected onto thedisplay screen 50 as an overlay upon the presently viewed program. First, by displaying information to theuser 40 on thedisplay screen 50, theuser 40 does not need to look at theremote control unit 20 in order to operate it. If theuser 40 needs to enter input, such as a spelled word, an overlay image of a keyboard may be presented and theuser 40 can navigate to the desired keys by simply manipulating the touch screen while watching a cursor or cursors (one for each finger) on the displayed overlay keyboard. If desired, the remote control system circuitry can also obtain program guide information and the display overlay can then allow theuser 40 to select which programs to view or record by simply manipulating the touch screen. - One can better understand the effectiveness of the remote control system by considering where the functionality of the system has been placed. By tight integration with the
display screen 50, the remote control system can use thedisplay screen 50, with its high resolution graphics capability, to provide an unlimited amount of visual information to theuser 40 which would be virtually impossible to provide through a set of dedicated buttons as conventional controllers do. The rich collection of diverse sensory inputs allows theuser 40 to adopt many different, and even redundant, ways of communicating the user's desires to the system. Interpretation of the diverse collection of sensory inputs by thepattern recognizer 70 handles much of the complexity of converting the user's gestural and touch commands into message meaning data that correlate to functions that the consumerelectronic equipment 48 can perform. The resulting division of labor produces a control system that provides both a very high, visually engaging information content to theuser 40 regarding his or her control system choices, with an equally rich collection of gestural and touch commands that theuser 40 can employ to get his or her message across to the control system. Compare this to the conventional push button remote control that requires one button, or a sequence of buttons, to be pressed for each desired function, with the added inconvenience that theuser 40 must look at the remote control in order to find the desired button to push. - From the foregoing, it will now be appreciated that the description and broad teachings of the present disclosure can be implemented in a variety of ways. Therefore, while this disclosure has been described with particular examples thereof, the true scope of the invention should not be so limited, as other modifications will become apparent to those of skill in the art upon study of the drawings, specification and following claims.
Claims (22)
1. A remote control system for consumer electronic equipment having a display screen, comprising:
a control circuit coupled to said consumer electronic equipment for controlling operation thereof;
a remote control unit having a handheld case and plurality of sensors disposed within said case, including:
(a) at least one touchpad responsive to manipulation by a user's fingers; and
(b) at least one additional sensor selected from the group consisting of:
(1) acceleration sensors responsive to movement of the remote control unit;
(2) case perimeter sensors that discriminates which parts of the case are in contact with the user's body;
(3) pressure sensors responsive to pressing forces upon a predetermined region of the touchpad; and
(4) RF signal sensors responsive to radio frequency signals transmitted from said control circuit;
said remote control unit being configured to produce a sensory input message containing information extracted from said plurality of sensors;
a pattern recognizer receptive of said sensory input message;
said pattern recognizer being configured to classify the received sensory input message according to a predetermined recognition scheme to generate message meaning data;
said pattern recognizer being further configured to classify the received sensory input message based on context information extracted from knowledge of what is being displayed on the display screen of the electronic equipment;
control logic responsive to said message meaning data to generate at least one command that is communicated to said control circuit.
2. The remote control system of claim 1 ,
wherein said at least one additional sensor comprises at least one acceleration sensor and at least one case perimeter sensor.
3. The remote control system of claim 1 ,
wherein said pattern recognizer includes a trained model operable to classify the received sensory input message.
4. The remote control system of claim 3 ,
wherein said trained model is adaptable based on interaction with the user.
5. The remote control system of claim 1 ,
wherein said remote control unit includes power management system and wherein said control logic generates at least one command that is communicated to said power management system to alter power management system behavior based on sensory input messages said sensory input message.
6. The remote control system of claim 1 ,
wherein said remote control unit includes power management system and wherein said control logic generates at least one command that is communicated to said power management system to alter power management system behavior based on said message meaning data.
7. The remote control system of claim 1 ,
wherein said control circuit includes memory for storing control system operation information and wherein said control circuit selectively provides said operation information to the consumer electronic equipment for presentation to the user.
8. The remote control system of claim 7 ,
wherein said control circuit provides said control system operation information to the consumer electronic equipment in accordance with said at least one command.
9. The remote control system of claim 1 ,
wherein said control circuit includes memory for storing control system operation information that is integrated with different message meaning data; and
wherein said control circuit selectively provides said operation information to the consumer electronic equipment in accordance with the message meaning data generated by said pattern recognizer.
10. The remote control system of claim 1 ,
wherein said at least one additional sensor is an RF signal sensor responsive to radio frequency signals transmitted from said control circuit and wherein said control circuit is configured to send a find-me message that causes the remote control unit to perform an attention attracting operation useful in assisting the user to find the remote control unit when it has become misplaced.
11. The remote control system of claim 1 , further comprising two touchpads disposed in side-by-side relationship each being separately responsive to manipulation by the user's fingers.
12. The remote control system of claim 1 ,
wherein said touchpad is divided into plural separate regions of interaction, such that each region of interaction is interpreted to have a different meaning by said pattern recognizer.
13. The remote system of claim 1 ,
wherein said at least one additional sensor comprises a case orientation sensor capable of differentiating between portrait orientation and landscape orientation.
14. The remote control system of claim 1 ,
wherein said touchpad is divided into plural separate regions of interaction, such that each region of interaction is interpreted to have a different meaning by said pattern recognizer,
wherein said at least one additional sensor comprises a case orientation sensor capable of differentiating between portrait orientation and landscape orientation; and
wherein said regions of interaction are remapped onto different locations within the touchpad based on case orientation.
15. The remote control system of claim 14 ,
wherein said at least one additional sensor further comprises the at least one acceleration sensor and at least one of the case perimeter sensors,
said pattern recognizer redefines a zone of interaction, by interpreting how the user is holding said remote control unit based on information extracted from said at least one additional sensor.
16. The remote control system of claim 15 ,
wherein said pattern recognizer interprets whether the user is holding said remote control unit with one hand or with two hands.
17. A remote control system for consumer electronic equipment having a display screen, comprising:
a control circuit coupled to said consumer electronic equipment for controlling operation thereof;
a remote control unit having a handheld case and plurality of sensors disposed within said case, including:
(a) at least one touchpad responsive to manipulation by a user's fingers;
(b) at least acceleration sensor responsive to movement of the remote control unit;
(c) a case perimeter sensor that discriminates which parts of the case are in contact with the user's body;
(d) at least one pressure sensor responsive to radio frequency signals transmitted from said control circuit;
said remote control unit being configured to produce a sensory input message containing information extracted from said plurality of sensors;
a pattern recognizer receptive of said sensory input message;
said pattern recognizer including a trained model adaptable through interaction with the user and configured to classify the received sensory input message according to a predetermined recognition scheme to generate message meaning data;
said pattern recognizer being further configured to classify the received sensory input message based on context information extracted from knowledge of what is being displayed on the display screen of the electronic equipment;
control logic responsive to said message meaning data to generate at least one command that is communicated to said control circuit;
wherein said remote control unit includes power management system and wherein said control logic generates at least one command that is communicated to said power management system to alter power management system behavior based on said message meaning data; an
wherein said control circuit having memory for storing control system operation information and wherein said control circuit selectively provides said operation information to the consumer electronic equipment for presentation to the user wherein said control circuit selectively provides said operation information to the consumer electronic equipment in accordance with the message meaning data generated by said pattern recognizer.
18. The remote control system of claim 17 ,
wherein said power management system performs:
placing of said plurality of sensors included in said remote control unit, into a sleep mode;
detecting movement of said remote control unit by waking said acceleration sensor up in the sleep mode;
determining whether or not said remote control unit is held by the user, by waking said case perimeter sensor up when the movement of said remote control unit is detected; and
releasing the sleep mode when the detecting detects that said remote control unit is held by the user.
19. The remote control system of claim 18 , wherein said power management system releases the sleep mode, by waking said at least one touchpad up and then waking said RF signal sensor up, when the detecting detects that said remote control unit is held by the user.
20. A method of controlling consumer electronic equipment comprising;
providing a remote control unit having at least one touchpad;
sensing manipulation by a user of said at least one touchpad;
sensing physical movement of said remote control unit;
classifying said sensed manipulation and said sensed physical movement by pattern recognition;
sending at least one control command to the consumer electronic equipment based on the results of said classifying step.
21. A remote control system for consumer electronic equipment having a display screen, comprising:
a control circuit coupled to said consumer electronic equipment for controlling operation thereof;
a remote control unit having a handheld case and plurality of sensors disposed within said case, including:
(a) at least one touchpad responsive to manipulation by a user's fingers; and
(b) at least one additional sensor selected from the group consisting of:
(1) acceleration sensors responsive to movement of the remote control unit;
(2) case perimeter sensors that discriminates which parts of the case are in contact with the user's body;
said remote control unit being configured to produce a sensory input message containing information extracted from said plurality of sensors;
a pattern recognizer receptive of said sensory input message;
said pattern recognizer being configured to classify the received sensory input message according to a predetermined recognition scheme to generate message meaning data;
control logic responsive to said message meaning data to generate at least one command that is communicated to said control circuit,
wherein said at least one additional sensor further comprises the at least one acceleration sensor and at least one of the case perimeter sensors,
said pattern recognizer redefines a zone of interaction, by interpreting how the user is holding said remote control unit based on information extracted from said at least one additional sensor;
wherein said remote control unit includes power management system and wherein said control logic generates at least one command that is communicated to said power management system to alter power management system behavior on said message meaning date; and
wherein said power management system performs:
placing of said plurality of sensors included in said remote control unit, into a sleep mode;
detecting movement of said remote control unit by waking said acceleration sensor up in the sleep mode;
determining whether or not said remote control unit is held by the user, by waking said case perimeter sensor up when the movement of said remote control unit is detected; and
releasing the sleep mode when the detecting detects that said remote control unit is held by the user.
22. The remote control system of claim 21 , further comprising:
(3) pressure sensors responsive to pressing forces upon a predetermined region of the touchpad; and
(4) RF signal sensors responsive to radio frequency signals transmitted from said control circuit; said pattern recognizer being further configured to classify the received sensory input message based on context information extracted from knowledge of what is being displayed on the display screen of the electronic equipment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/666,951 US20110018817A1 (en) | 2007-06-28 | 2008-06-26 | Touchpad-enabled remote controller and user interaction methods |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US94680407P | 2007-06-28 | 2007-06-28 | |
US11/977,348 US7889175B2 (en) | 2007-06-28 | 2007-10-24 | Touchpad-enabled remote controller and user interaction methods |
US12/666,951 US20110018817A1 (en) | 2007-06-28 | 2008-06-26 | Touchpad-enabled remote controller and user interaction methods |
PCT/US2008/068390 WO2009006224A1 (en) | 2007-06-28 | 2008-06-26 | Touchpad-enabled remote controller and user interaction methods |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,348 Continuation US7889175B2 (en) | 2007-06-28 | 2007-10-24 | Touchpad-enabled remote controller and user interaction methods |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110018817A1 true US20110018817A1 (en) | 2011-01-27 |
Family
ID=40159744
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,348 Expired - Fee Related US7889175B2 (en) | 2007-06-28 | 2007-10-24 | Touchpad-enabled remote controller and user interaction methods |
US12/666,951 Abandoned US20110018817A1 (en) | 2007-06-28 | 2008-06-26 | Touchpad-enabled remote controller and user interaction methods |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/977,348 Expired - Fee Related US7889175B2 (en) | 2007-06-28 | 2007-10-24 | Touchpad-enabled remote controller and user interaction methods |
Country Status (4)
Country | Link |
---|---|
US (2) | US7889175B2 (en) |
EP (1) | EP2153305B1 (en) |
JP (1) | JP5473908B2 (en) |
WO (1) | WO2009006224A1 (en) |
Cited By (68)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090153288A1 (en) * | 2007-12-12 | 2009-06-18 | Eric James Hope | Handheld electronic devices with remote control functionality and gesture recognition |
US20100079681A1 (en) * | 2008-09-30 | 2010-04-01 | Echostar Technologies Llc | Systems and methods for graphical control of symbol-based features in a television receiver |
US20110140847A1 (en) * | 2009-12-15 | 2011-06-16 | Echostar Technologies L.L.C. | Audible feedback for input activation of a remote control device |
US20110304648A1 (en) * | 2010-06-15 | 2011-12-15 | Lg Electronics Inc. | Mobile terminal and method for operating the mobile terminal |
US20120025946A1 (en) * | 2010-07-29 | 2012-02-02 | Hon Hai Precision Industry Co., Ltd. | Electronic device with remote function |
US20120032917A1 (en) * | 2009-04-28 | 2012-02-09 | Yoshiyuki Yamaguchi | Touch panel, touch panel manufacturing method, and electronic apparatus |
US8230246B1 (en) * | 2011-03-30 | 2012-07-24 | Google Inc. | Activating a computer device based on the detecting result from a single touch sensor if the battery level is high |
US20120268408A1 (en) * | 2010-01-06 | 2012-10-25 | Huawei Device Co., Ltd. | Method and terminal for displaying picture/interface |
US20120287350A1 (en) * | 2011-05-11 | 2012-11-15 | Toshiba Samsung Storage Technology Korea Corporate | Remote controller, and control method and system using the same |
WO2013106518A1 (en) * | 2012-01-10 | 2013-07-18 | Fanhattan Llc | Touch-enabled remote control |
US8582957B2 (en) | 2008-09-22 | 2013-11-12 | EchoStar Technologies, L.L.C. | Methods and apparatus for visually displaying recording timer information |
US20140114484A1 (en) * | 2012-03-09 | 2014-04-24 | Panasonic Corporation | Appliance control method, appliance control apparatus, control server, and control terminal |
US20140136867A1 (en) * | 2011-06-17 | 2014-05-15 | Sony Corporation | Electronic device, control method of electronic device, and program |
US8763045B2 (en) | 2008-09-30 | 2014-06-24 | Echostar Technologies L.L.C. | Systems and methods for providing customer service features via a graphical user interface in a television receiver |
US8793735B2 (en) | 2008-09-30 | 2014-07-29 | EchoStar Technologies, L.L.C. | Methods and apparatus for providing multiple channel recall on a television receiver |
US8796575B2 (en) | 2012-10-31 | 2014-08-05 | Ford Global Technologies, Llc | Proximity switch assembly having ground layer |
US20140320536A1 (en) * | 2012-01-24 | 2014-10-30 | Google Inc. | Methods and Systems for Determining Orientation of a Display of Content on a Device |
US8878438B2 (en) | 2011-11-04 | 2014-11-04 | Ford Global Technologies, Llc | Lamp and proximity switch assembly and method |
US8922340B2 (en) | 2012-09-11 | 2014-12-30 | Ford Global Technologies, Llc | Proximity switch based door latch release |
US8928336B2 (en) | 2011-06-09 | 2015-01-06 | Ford Global Technologies, Llc | Proximity switch having sensitivity control and method therefor |
US8933708B2 (en) | 2012-04-11 | 2015-01-13 | Ford Global Technologies, Llc | Proximity switch assembly and activation method with exploration mode |
US8975903B2 (en) | 2011-06-09 | 2015-03-10 | Ford Global Technologies, Llc | Proximity switch having learned sensitivity and method therefor |
US8981602B2 (en) | 2012-05-29 | 2015-03-17 | Ford Global Technologies, Llc | Proximity switch assembly having non-switch contact and method |
US8994228B2 (en) | 2011-11-03 | 2015-03-31 | Ford Global Technologies, Llc | Proximity switch having wrong touch feedback |
US9065447B2 (en) | 2012-04-11 | 2015-06-23 | Ford Global Technologies, Llc | Proximity switch assembly and method having adaptive time delay |
US9100614B2 (en) | 2008-10-31 | 2015-08-04 | Echostar Technologies L.L.C. | Graphical interface navigation based on image element proximity |
US9136840B2 (en) | 2012-05-17 | 2015-09-15 | Ford Global Technologies, Llc | Proximity switch assembly having dynamic tuned threshold |
US9143126B2 (en) | 2011-09-22 | 2015-09-22 | Ford Global Technologies, Llc | Proximity switch having lockout control for controlling movable panel |
US9152373B2 (en) | 2011-04-12 | 2015-10-06 | Apple Inc. | Gesture visualization and sharing between electronic devices and remote displays |
US9184745B2 (en) | 2012-04-11 | 2015-11-10 | Ford Global Technologies, Llc | Proximity switch assembly and method of sensing user input based on signal rate of change |
US9197206B2 (en) | 2012-04-11 | 2015-11-24 | Ford Global Technologies, Llc | Proximity switch having differential contact surface |
WO2015182935A1 (en) * | 2014-05-29 | 2015-12-03 | Samsung Electronics Co., Ltd. | Method of controlling display device and remote controller thereof |
US9219472B2 (en) | 2012-04-11 | 2015-12-22 | Ford Global Technologies, Llc | Proximity switch assembly and activation method using rate monitoring |
US9239890B2 (en) | 2011-05-31 | 2016-01-19 | Fanhattan, Inc. | System and method for carousel context switching |
US9287864B2 (en) | 2012-04-11 | 2016-03-15 | Ford Global Technologies, Llc | Proximity switch assembly and calibration method therefor |
US9311204B2 (en) | 2013-03-13 | 2016-04-12 | Ford Global Technologies, Llc | Proximity interface development system having replicator and method |
US9337832B2 (en) | 2012-06-06 | 2016-05-10 | Ford Global Technologies, Llc | Proximity switch and method of adjusting sensitivity therefor |
US20160150472A1 (en) * | 2014-11-21 | 2016-05-26 | Samsung Electronics Co., Ltd. | User terminal for controlling display device and control method thereof |
US9520875B2 (en) | 2012-04-11 | 2016-12-13 | Ford Global Technologies, Llc | Pliable proximity switch assembly and activation method |
US9531379B2 (en) | 2012-04-11 | 2016-12-27 | Ford Global Technologies, Llc | Proximity switch assembly having groove between adjacent proximity sensors |
US20170003818A1 (en) * | 2010-06-21 | 2017-01-05 | Apple Inc. | Portable multi-touch input device |
US9548733B2 (en) | 2015-05-20 | 2017-01-17 | Ford Global Technologies, Llc | Proximity sensor assembly having interleaved electrode configuration |
US9559688B2 (en) | 2012-04-11 | 2017-01-31 | Ford Global Technologies, Llc | Proximity switch assembly having pliable surface and depression |
US9568527B2 (en) | 2012-04-11 | 2017-02-14 | Ford Global Technologies, Llc | Proximity switch assembly and activation method having virtual button mode |
US9607505B2 (en) | 2010-09-22 | 2017-03-28 | Apple Inc. | Closed loop universal remote control |
US9641172B2 (en) | 2012-06-27 | 2017-05-02 | Ford Global Technologies, Llc | Proximity switch assembly having varying size electrode fingers |
US9654103B2 (en) | 2015-03-18 | 2017-05-16 | Ford Global Technologies, Llc | Proximity switch assembly having haptic feedback and method |
US9660644B2 (en) | 2012-04-11 | 2017-05-23 | Ford Global Technologies, Llc | Proximity switch assembly and activation method |
US9716774B2 (en) | 2008-07-10 | 2017-07-25 | Apple Inc. | System and method for syncing a user interface on a server device to a user interface on a client device |
US9778818B2 (en) | 2011-05-31 | 2017-10-03 | Fanhattan, Inc. | System and method for pyramidal navigation |
US9831870B2 (en) | 2012-04-11 | 2017-11-28 | Ford Global Technologies, Llc | Proximity switch assembly and method of tuning same |
US9944237B2 (en) | 2012-04-11 | 2018-04-17 | Ford Global Technologies, Llc | Proximity switch assembly with signal drift rejection and method |
US10004286B2 (en) | 2011-08-08 | 2018-06-26 | Ford Global Technologies, Llc | Glove having conductive ink and method of interacting with proximity sensor |
US10038443B2 (en) | 2014-10-20 | 2018-07-31 | Ford Global Technologies, Llc | Directional proximity switch assembly |
US10112556B2 (en) | 2011-11-03 | 2018-10-30 | Ford Global Technologies, Llc | Proximity switch having wrong touch adaptive learning and method |
US20190172339A1 (en) * | 2017-12-05 | 2019-06-06 | Bose Corporation | Context-sensitive remote control device |
US11403483B2 (en) | 2017-06-20 | 2022-08-02 | Hologic, Inc. | Dynamic self-learning medical image method and system |
US11406332B2 (en) | 2011-03-08 | 2022-08-09 | Hologic, Inc. | System and method for dual energy and/or contrast enhanced breast imaging for screening, diagnosis and biopsy |
US11419565B2 (en) | 2014-02-28 | 2022-08-23 | IIologic, Inc. | System and method for generating and displaying tomosynthesis image slabs |
US11445993B2 (en) | 2017-03-30 | 2022-09-20 | Hologic, Inc. | System and method for targeted object enhancement to generate synthetic breast tissue images |
US11455754B2 (en) | 2017-03-30 | 2022-09-27 | Hologic, Inc. | System and method for synthesizing low-dimensional image data from high-dimensional image data using an object grid enhancement |
US11452486B2 (en) | 2006-02-15 | 2022-09-27 | Hologic, Inc. | Breast biopsy and needle localization using tomosynthesis systems |
US11508340B2 (en) | 2011-11-27 | 2022-11-22 | Hologic, Inc. | System and method for generating a 2D image using mammography and/or tomosynthesis image data |
US11589944B2 (en) | 2013-03-15 | 2023-02-28 | Hologic, Inc. | Tomosynthesis-guided biopsy apparatus and method |
US11663780B2 (en) | 2012-02-13 | 2023-05-30 | Hologic Inc. | System and method for navigating a tomosynthesis stack using synthesized image data |
US11701199B2 (en) | 2009-10-08 | 2023-07-18 | Hologic, Inc. | Needle breast biopsy system and method of use |
US11775156B2 (en) * | 2010-11-26 | 2023-10-03 | Hologic, Inc. | User interface for medical image review workstation |
US11957497B2 (en) | 2022-03-11 | 2024-04-16 | Hologic, Inc | System and method for hierarchical multi-level feature image synthesis and representation |
Families Citing this family (76)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2020651B1 (en) * | 2007-08-01 | 2014-06-25 | fm marketing gmbh | Remote control |
US9489086B1 (en) | 2013-04-29 | 2016-11-08 | Apple Inc. | Finger hover detection for improved typing |
US10126942B2 (en) | 2007-09-19 | 2018-11-13 | Apple Inc. | Systems and methods for detecting a press on a touch-sensitive surface |
US9110590B2 (en) | 2007-09-19 | 2015-08-18 | Typesoft Technologies, Inc. | Dynamically located onscreen keyboard |
US20120075193A1 (en) * | 2007-09-19 | 2012-03-29 | Cleankeys Inc. | Multiplexed numeric keypad and touchpad |
US9454270B2 (en) | 2008-09-19 | 2016-09-27 | Apple Inc. | Systems and methods for detecting a press on a touch-sensitive surface |
US10203873B2 (en) | 2007-09-19 | 2019-02-12 | Apple Inc. | Systems and methods for adaptively presenting a keyboard on a touch-sensitive display |
US20100299710A1 (en) * | 2007-09-20 | 2010-11-25 | Samsung Electronics Co. Ltd. | Method for inputting user command and video apparatus and input apparatus employing the same |
US8823645B2 (en) | 2010-12-28 | 2014-09-02 | Panasonic Corporation | Apparatus for remotely controlling another apparatus and having self-orientating capability |
US8031175B2 (en) * | 2008-04-21 | 2011-10-04 | Panasonic Corporation | Touch sensitive remote control system that detects hand size characteristics of user and adapts mapping to screen display |
KR101451271B1 (en) * | 2007-10-30 | 2014-10-16 | 삼성전자주식회사 | Broadcast receiving apparatus and control method thereof |
KR101536750B1 (en) | 2007-11-08 | 2015-07-15 | 삼성전자주식회사 | Remote controller for setting mode according to state of broadcast receiving apparatus |
DE102007055465A1 (en) * | 2007-11-13 | 2009-05-20 | Trumpf Medizin Systeme Gmbh | Remote control for a device for storing a patient |
EP2071443A3 (en) * | 2007-12-12 | 2012-05-30 | Advanced Digital Broadcast S.A. | Method for controlling value of parameter |
TWI356357B (en) * | 2007-12-24 | 2012-01-11 | Univ Nat Chiao Tung | A method for estimating a body pose |
JP2009213126A (en) * | 2008-02-08 | 2009-09-17 | Rohm Co Ltd | Remote control apparatus and electronic device using electrostatic sensor, and control method of switch |
US20090231138A1 (en) * | 2008-03-14 | 2009-09-17 | Chung Nam Electronics Co. Ltd. | RFID Technology |
US8564547B2 (en) * | 2008-04-15 | 2013-10-22 | Mattel, Inc. | Touch screen remote control device for use with a toy |
CN101598972A (en) * | 2008-06-04 | 2009-12-09 | 鸿富锦精密工业(深圳)有限公司 | Electronic installation and function change method thereof |
US20100052931A1 (en) * | 2008-08-26 | 2010-03-04 | Gm Global Technology Operations, Inc. | Gesture control key fob |
US20100060569A1 (en) * | 2008-09-09 | 2010-03-11 | Lucent Technologies Inc. | Wireless remote control having motion-based control functions and method of manufacture thereof |
US8806229B1 (en) | 2008-09-29 | 2014-08-12 | Cypress Semiconductor Corporation | Power reduction circuits and methods |
KR20100039017A (en) * | 2008-10-07 | 2010-04-15 | 한국전자통신연구원 | Remote control apparatus using menu markup language |
US20100205628A1 (en) * | 2009-02-12 | 2010-08-12 | Davis Bruce L | Media processing methods and arrangements |
KR101520746B1 (en) * | 2009-01-07 | 2015-05-15 | 삼성전자주식회사 | Input device and display device |
US8887052B1 (en) | 2009-01-09 | 2014-11-11 | Google Inc. | Presentation remote control |
US20110019105A1 (en) * | 2009-07-27 | 2011-01-27 | Echostar Technologies L.L.C. | Verification of symbols received through a touchpad of a remote control device in an electronic system to allow access to system functions |
US9055904B2 (en) * | 2009-08-03 | 2015-06-16 | Nike, Inc. | Multi-touch display and input for vision testing and training |
US8432305B2 (en) * | 2009-09-03 | 2013-04-30 | Samsung Electronics Co., Ltd. | Electronic apparatus, control method thereof, remote control apparatus, and control method thereof |
US20110063522A1 (en) * | 2009-09-14 | 2011-03-17 | Jeyhan Karaoguz | System and method for generating television screen pointing information using an external receiver |
US9177484B2 (en) * | 2009-10-01 | 2015-11-03 | Andrew Chan | Apparatus and method of supporting communication and performance among a group of musicians |
US20110080344A1 (en) * | 2009-10-02 | 2011-04-07 | Dedo Interactive Inc. | Blending touch data streams that include touch input data |
EP2323113A1 (en) * | 2009-11-12 | 2011-05-18 | Securitas Direct AB | A method and a system for exchanging information |
US8754746B2 (en) * | 2009-11-16 | 2014-06-17 | Broadcom Corporation | Hand-held gaming device that identifies user based upon input from touch sensitive panel |
US8508482B2 (en) | 2009-11-30 | 2013-08-13 | Neil Van der Byl | Programmable remote control |
KR101670352B1 (en) * | 2009-12-03 | 2016-10-28 | 엘지전자 주식회사 | Mobile Terminal, Electronic Device and Method Of Controlling Electronic Device |
US20110140868A1 (en) * | 2009-12-14 | 2011-06-16 | Securitas Direct Ab | Method and a system for exchanging information |
US20110148803A1 (en) * | 2009-12-23 | 2011-06-23 | Amlogic Co., Ltd. | Remote Controller Having A Touch Panel For Inputting Commands |
JP5454133B2 (en) * | 2009-12-25 | 2014-03-26 | 富士通株式会社 | Detection information correction device, portable device, detection information correction method, and computer program |
JP5258816B2 (en) * | 2010-02-27 | 2013-08-07 | 三菱電機株式会社 | Air conditioner |
GB2481368A (en) * | 2010-06-07 | 2011-12-28 | Electric Mobility Euro Ltd | Furniture controller with automatic disabling of user controls to prevent accidental operation and requiring specific input to reactivate controls |
WO2011158475A1 (en) | 2010-06-16 | 2011-12-22 | パナソニック株式会社 | Information input device, information input method and programme |
US9192065B2 (en) | 2010-08-18 | 2015-11-17 | Mitsubishi Electric Corporation | Electrical device, control method, and program |
US8587416B2 (en) * | 2010-09-02 | 2013-11-19 | Sling Media Pvt Ltd | Locating remote control devices utilizing base unit positioning |
US9430128B2 (en) | 2011-01-06 | 2016-08-30 | Tivo, Inc. | Method and apparatus for controls based on concurrent gestures |
WO2012094479A1 (en) * | 2011-01-06 | 2012-07-12 | Tivo Inc. | Method and apparatus for gesture based controls |
EP2665497A2 (en) | 2011-01-20 | 2013-11-27 | Cleankeys Inc. | Systems and methods for monitoring surface sanitation |
FR2971607B1 (en) * | 2011-02-10 | 2013-08-23 | Parrot | WIRELESS REMOTE CONTROL WITH TOUCH INTERFACE FOR MULTIMEDIA AND HANDS-FREE TELEPHONY EQUIPMENT OF A MOTOR VEHICLE |
US20130298028A1 (en) * | 2011-04-19 | 2013-11-07 | Shahar Porat | Multifunctional input device |
US9081810B1 (en) | 2011-04-29 | 2015-07-14 | Google Inc. | Remote device control using gestures on a touch sensitive device |
US8736564B2 (en) * | 2011-05-25 | 2014-05-27 | Blackberry Limited | Proximity detection between a mobile device and a related object |
US8971572B1 (en) | 2011-08-12 | 2015-03-03 | The Research Foundation For The State University Of New York | Hand pointing estimation for human computer interaction |
US9164779B2 (en) | 2012-02-10 | 2015-10-20 | Nokia Technologies Oy | Apparatus and method for providing for remote user interaction |
US9715823B2 (en) * | 2012-02-23 | 2017-07-25 | Koninklijke Philips N.V. | Remote control device |
US9104260B2 (en) | 2012-04-10 | 2015-08-11 | Typesoft Technologies, Inc. | Systems and methods for detecting a press on a touch-sensitive surface |
US8744200B1 (en) * | 2012-05-04 | 2014-06-03 | Hrl Laboratories, Llc | Hybrid compressive/Nyquist sampling for enhanced sensing |
KR101336289B1 (en) * | 2012-05-08 | 2013-12-03 | 도시바삼성스토리지테크놀러지코리아 주식회사 | Apparatus and method for controlling electronic device |
TWI505553B (en) * | 2012-06-21 | 2015-10-21 | Wistron Neweb Corp | Multi-function radio-frequency device, computer system and method of operating multi-function radio-frequency device |
CN102930705B (en) * | 2012-09-26 | 2014-09-03 | 腾讯科技(深圳)有限公司 | Device control method and device |
KR101473469B1 (en) * | 2012-10-04 | 2014-12-18 | 크루셜텍 (주) | Remote control device and method for controling operation of electronic equipment |
US9710080B2 (en) | 2012-11-20 | 2017-07-18 | Nec Corporation | Portable electronic device including contact sensors, and method for controlling same |
US9524633B2 (en) | 2013-03-14 | 2016-12-20 | Lutron Electronics Co., Inc. | Remote control having a capacitive touch surface and a mechanism for awakening the remote control |
FR3004548B1 (en) * | 2013-04-11 | 2016-08-19 | Bigben Interactive Sa | REMOTE CONTROL WITH RECONFIGURABLE INTERFACE |
JP5659303B1 (en) * | 2013-05-22 | 2015-01-28 | 株式会社東芝 | DEVICE CONTROL METHOD, DEVICE CONTROL DEVICE, PROGRAM, AND ELECTRONIC DEVICE |
US9454251B1 (en) | 2013-06-26 | 2016-09-27 | Google Inc. | Methods, systems, and media for controlling a remote device using a touch screen of a mobile device in a display inhibited state |
US10289302B1 (en) | 2013-09-09 | 2019-05-14 | Apple Inc. | Virtual keyboard animation |
US9687576B2 (en) | 2014-11-14 | 2017-06-27 | Echostar Technologies L.L.C. | Systems and methods for disinfecting a remote control using ultraviolet light |
US9727231B2 (en) | 2014-11-19 | 2017-08-08 | Honda Motor Co., Ltd. | System and method for providing absolute coordinate and zone mapping between a touchpad and a display screen |
US20170371515A1 (en) | 2014-11-19 | 2017-12-28 | Honda Motor Co., Ltd. | System and method for providing absolute and zone coordinate mapping with graphic animations |
US20160182950A1 (en) * | 2014-12-17 | 2016-06-23 | Lenovo (Singapore) Pte. Ltd. | Identification of a user for personalized media content presentation |
JP2017076208A (en) * | 2015-10-14 | 2017-04-20 | オムロンオートモーティブエレクトロニクス株式会社 | Mobile electronic device |
WO2017120855A1 (en) * | 2016-01-14 | 2017-07-20 | 汪宇峰 | Method for collecting data upon determination of motion track and remote control |
KR102497299B1 (en) * | 2016-06-29 | 2023-02-08 | 삼성전자주식회사 | Electronic apparatus and method for controlling the electronic apparatus |
KR20190017280A (en) * | 2017-08-10 | 2019-02-20 | 엘지전자 주식회사 | Mobile terminal and method for controlling of the same |
US11949790B2 (en) * | 2020-10-01 | 2024-04-02 | Bank Of America Corporation | Intelligent tactile resource instrument activation using electroencephalogram signals |
KR102634742B1 (en) * | 2022-11-23 | 2024-02-08 | 주식회사 오성전자 | Remote control that detects direction change |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5602982A (en) * | 1994-09-23 | 1997-02-11 | Kelly Properties, Inc. | Universal automated training and testing software system |
US6908386B2 (en) * | 2002-05-17 | 2005-06-21 | Nintendo Co., Ltd. | Game device changing sound and an image in accordance with a tilt operation |
US7094147B2 (en) * | 2001-08-22 | 2006-08-22 | Nintendo Co., Ltd. | Game system, puzzle game program, and storage medium having program stored therein |
Family Cites Families (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06165267A (en) * | 1992-11-16 | 1994-06-10 | Hitachi Ltd | Remote controller and audio visual equipment using the remote controller |
JP3857504B2 (en) * | 1993-05-21 | 2006-12-13 | ソニー株式会社 | Input device, control system |
US5469194A (en) | 1994-05-13 | 1995-11-21 | Apple Computer, Inc. | Apparatus and method for providing different input device orientations of a computer system |
GB9415627D0 (en) | 1994-08-01 | 1994-09-21 | Marshall James | Verification apparatus |
US5652630A (en) * | 1995-05-31 | 1997-07-29 | International Business Machines Corporation | Video receiver display, three axis remote control, and microcontroller for executing programs |
US5724106A (en) * | 1995-07-17 | 1998-03-03 | Gateway 2000, Inc. | Hand held remote control device with trigger button |
US6211870B1 (en) * | 1997-07-07 | 2001-04-03 | Combi/Mote Corp. | Computer programmable remote control |
US7800592B2 (en) * | 2005-03-04 | 2010-09-21 | Apple Inc. | Hand held electronic device with multiple touch sensing devices |
US6346891B1 (en) | 1998-08-31 | 2002-02-12 | Microsoft Corporation | Remote control system with handling sensor in remote control device |
US6456275B1 (en) | 1998-09-14 | 2002-09-24 | Microsoft Corporation | Proximity sensor in a computer input device |
US6501515B1 (en) | 1998-10-13 | 2002-12-31 | Sony Corporation | Remote control system |
US6407779B1 (en) * | 1999-03-29 | 2002-06-18 | Zilog, Inc. | Method and apparatus for an intuitive universal remote control system |
US6396523B1 (en) * | 1999-07-29 | 2002-05-28 | Interlink Electronics, Inc. | Home entertainment device remote control |
US6429543B1 (en) | 1999-10-01 | 2002-08-06 | Siemens Vdo Automotive Corporation | Innovative switch for remote control applications |
US6765557B1 (en) * | 2000-04-10 | 2004-07-20 | Interlink Electronics, Inc. | Remote control having touch pad to screen mapping |
US20010030667A1 (en) * | 2000-04-10 | 2001-10-18 | Kelts Brett R. | Interactive display interface for information objects |
US20040236699A1 (en) | 2001-07-10 | 2004-11-25 | American Express Travel Related Services Company, Inc. | Method and system for hand geometry recognition biometrics on a fob |
WO2003071410A2 (en) | 2002-02-15 | 2003-08-28 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US11275405B2 (en) | 2005-03-04 | 2022-03-15 | Apple Inc. | Multi-functional hand-held device |
TWM240050U (en) | 2003-04-02 | 2004-08-01 | Elan Microelectronics Corp | Capacitor touch panel with integrated keyboard and handwriting function |
US20050162402A1 (en) | 2004-01-27 | 2005-07-28 | Watanachote Susornpol J. | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback |
FI117308B (en) | 2004-02-06 | 2006-08-31 | Nokia Corp | gesture Control |
US20050185788A1 (en) | 2004-02-23 | 2005-08-25 | Daw Sean P. | Keypad adapted for use in dual orientations |
WO2005109847A2 (en) * | 2004-04-30 | 2005-11-17 | Hillcrest Laboratories, Inc. | Methods and devices for identifying users based on tremor |
TW200539031A (en) | 2004-05-20 | 2005-12-01 | Elan Microelectronics Corp | A capacitor type touch pad with integrated graphic input function |
JP2006013565A (en) * | 2004-06-22 | 2006-01-12 | Sharp Corp | Remote control device |
US7173619B2 (en) * | 2004-07-08 | 2007-02-06 | Microsoft Corporation | Matching digital information flow to a human perception system |
US20060227030A1 (en) | 2005-03-31 | 2006-10-12 | Clifford Michelle A | Accelerometer based control system and method of controlling a device |
US7844301B2 (en) * | 2005-10-14 | 2010-11-30 | Lg Electronics Inc. | Method for displaying multimedia contents and mobile communications terminal capable of implementing the same |
JP4679429B2 (en) | 2006-04-27 | 2011-04-27 | 任天堂株式会社 | Sound output program and sound output device |
-
2007
- 2007-10-24 US US11/977,348 patent/US7889175B2/en not_active Expired - Fee Related
-
2008
- 2008-06-26 US US12/666,951 patent/US20110018817A1/en not_active Abandoned
- 2008-06-26 WO PCT/US2008/068390 patent/WO2009006224A1/en active Search and Examination
- 2008-06-26 EP EP08781028.9A patent/EP2153305B1/en not_active Not-in-force
- 2008-06-26 JP JP2010515130A patent/JP5473908B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5602982A (en) * | 1994-09-23 | 1997-02-11 | Kelly Properties, Inc. | Universal automated training and testing software system |
US7094147B2 (en) * | 2001-08-22 | 2006-08-22 | Nintendo Co., Ltd. | Game system, puzzle game program, and storage medium having program stored therein |
US6908386B2 (en) * | 2002-05-17 | 2005-06-21 | Nintendo Co., Ltd. | Game device changing sound and an image in accordance with a tilt operation |
Cited By (94)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11918389B2 (en) | 2006-02-15 | 2024-03-05 | Hologic, Inc. | Breast biopsy and needle localization using tomosynthesis systems |
US11452486B2 (en) | 2006-02-15 | 2022-09-27 | Hologic, Inc. | Breast biopsy and needle localization using tomosynthesis systems |
US20090153288A1 (en) * | 2007-12-12 | 2009-06-18 | Eric James Hope | Handheld electronic devices with remote control functionality and gesture recognition |
US9767681B2 (en) * | 2007-12-12 | 2017-09-19 | Apple Inc. | Handheld electronic devices with remote control functionality and gesture recognition |
US10825338B2 (en) | 2007-12-12 | 2020-11-03 | Apple Inc. | Handheld electronic devices with remote control functionality and gesture recognition |
US9716774B2 (en) | 2008-07-10 | 2017-07-25 | Apple Inc. | System and method for syncing a user interface on a server device to a user interface on a client device |
US8582957B2 (en) | 2008-09-22 | 2013-11-12 | EchoStar Technologies, L.L.C. | Methods and apparatus for visually displaying recording timer information |
US20100079681A1 (en) * | 2008-09-30 | 2010-04-01 | Echostar Technologies Llc | Systems and methods for graphical control of symbol-based features in a television receiver |
US8763045B2 (en) | 2008-09-30 | 2014-06-24 | Echostar Technologies L.L.C. | Systems and methods for providing customer service features via a graphical user interface in a television receiver |
US8937687B2 (en) * | 2008-09-30 | 2015-01-20 | Echostar Technologies L.L.C. | Systems and methods for graphical control of symbol-based features in a television receiver |
US8793735B2 (en) | 2008-09-30 | 2014-07-29 | EchoStar Technologies, L.L.C. | Methods and apparatus for providing multiple channel recall on a television receiver |
US9100614B2 (en) | 2008-10-31 | 2015-08-04 | Echostar Technologies L.L.C. | Graphical interface navigation based on image element proximity |
US20120032917A1 (en) * | 2009-04-28 | 2012-02-09 | Yoshiyuki Yamaguchi | Touch panel, touch panel manufacturing method, and electronic apparatus |
US11701199B2 (en) | 2009-10-08 | 2023-07-18 | Hologic, Inc. | Needle breast biopsy system and method of use |
US10181260B2 (en) | 2009-12-15 | 2019-01-15 | Echostar Technologies L.L.C. | Audible feedback for input activation of a remote control device |
US8629754B2 (en) * | 2009-12-15 | 2014-01-14 | Echostar Technologies L.L.C. | Audible feedback for input activation of a remote control device |
US20110140847A1 (en) * | 2009-12-15 | 2011-06-16 | Echostar Technologies L.L.C. | Audible feedback for input activation of a remote control device |
US8982073B2 (en) * | 2010-01-06 | 2015-03-17 | Huawei Device Co., Ltd. | Method and terminal for displaying picture/interface |
US20120268408A1 (en) * | 2010-01-06 | 2012-10-25 | Huawei Device Co., Ltd. | Method and terminal for displaying picture/interface |
US8935637B2 (en) * | 2010-06-15 | 2015-01-13 | Lg Electronics Inc. | Mobile terminal and method for operating the mobile terminal |
US20110304648A1 (en) * | 2010-06-15 | 2011-12-15 | Lg Electronics Inc. | Mobile terminal and method for operating the mobile terminal |
US9857912B2 (en) * | 2010-06-21 | 2018-01-02 | Apple Inc. | Portable multi-touch input device |
US20170003818A1 (en) * | 2010-06-21 | 2017-01-05 | Apple Inc. | Portable multi-touch input device |
US10579184B2 (en) | 2010-06-21 | 2020-03-03 | Apple Inc. | Portable multi-touch input device |
US20120025946A1 (en) * | 2010-07-29 | 2012-02-02 | Hon Hai Precision Industry Co., Ltd. | Electronic device with remote function |
US9607505B2 (en) | 2010-09-22 | 2017-03-28 | Apple Inc. | Closed loop universal remote control |
US11775156B2 (en) * | 2010-11-26 | 2023-10-03 | Hologic, Inc. | User interface for medical image review workstation |
US11406332B2 (en) | 2011-03-08 | 2022-08-09 | Hologic, Inc. | System and method for dual energy and/or contrast enhanced breast imaging for screening, diagnosis and biopsy |
US8930734B1 (en) | 2011-03-30 | 2015-01-06 | Google Inc. | Managing power states of a computing device |
US8230246B1 (en) * | 2011-03-30 | 2012-07-24 | Google Inc. | Activating a computer device based on the detecting result from a single touch sensor if the battery level is high |
US9152373B2 (en) | 2011-04-12 | 2015-10-06 | Apple Inc. | Gesture visualization and sharing between electronic devices and remote displays |
US20120287350A1 (en) * | 2011-05-11 | 2012-11-15 | Toshiba Samsung Storage Technology Korea Corporate | Remote controller, and control method and system using the same |
US9239890B2 (en) | 2011-05-31 | 2016-01-19 | Fanhattan, Inc. | System and method for carousel context switching |
US9778818B2 (en) | 2011-05-31 | 2017-10-03 | Fanhattan, Inc. | System and method for pyramidal navigation |
US8975903B2 (en) | 2011-06-09 | 2015-03-10 | Ford Global Technologies, Llc | Proximity switch having learned sensitivity and method therefor |
US8928336B2 (en) | 2011-06-09 | 2015-01-06 | Ford Global Technologies, Llc | Proximity switch having sensitivity control and method therefor |
US20140136867A1 (en) * | 2011-06-17 | 2014-05-15 | Sony Corporation | Electronic device, control method of electronic device, and program |
US9933837B2 (en) | 2011-06-17 | 2018-04-03 | Sony Corporation | Electronic device, control method of electronic device, and program |
US9459686B2 (en) * | 2011-06-17 | 2016-10-04 | Sony Corporation | Electronic device, control method of electronic device, and program |
US10551899B2 (en) | 2011-06-17 | 2020-02-04 | Sony Corporation | Electronic device, control method of electronic device, and program |
US11163353B2 (en) | 2011-06-17 | 2021-11-02 | Sony Corporation | Electronic device, control method of electronic device, and program |
US10004286B2 (en) | 2011-08-08 | 2018-06-26 | Ford Global Technologies, Llc | Glove having conductive ink and method of interacting with proximity sensor |
US10595574B2 (en) | 2011-08-08 | 2020-03-24 | Ford Global Technologies, Llc | Method of interacting with proximity sensor with a glove |
US9143126B2 (en) | 2011-09-22 | 2015-09-22 | Ford Global Technologies, Llc | Proximity switch having lockout control for controlling movable panel |
US10112556B2 (en) | 2011-11-03 | 2018-10-30 | Ford Global Technologies, Llc | Proximity switch having wrong touch adaptive learning and method |
US10501027B2 (en) | 2011-11-03 | 2019-12-10 | Ford Global Technologies, Llc | Proximity switch having wrong touch adaptive learning and method |
US8994228B2 (en) | 2011-11-03 | 2015-03-31 | Ford Global Technologies, Llc | Proximity switch having wrong touch feedback |
US8878438B2 (en) | 2011-11-04 | 2014-11-04 | Ford Global Technologies, Llc | Lamp and proximity switch assembly and method |
US11837197B2 (en) | 2011-11-27 | 2023-12-05 | Hologic, Inc. | System and method for generating a 2D image using mammography and/or tomosynthesis image data |
US11508340B2 (en) | 2011-11-27 | 2022-11-22 | Hologic, Inc. | System and method for generating a 2D image using mammography and/or tomosynthesis image data |
WO2013106518A1 (en) * | 2012-01-10 | 2013-07-18 | Fanhattan Llc | Touch-enabled remote control |
US9146616B2 (en) | 2012-01-10 | 2015-09-29 | Fanhattan Inc. | Touch-enabled remote control |
US20140320536A1 (en) * | 2012-01-24 | 2014-10-30 | Google Inc. | Methods and Systems for Determining Orientation of a Display of Content on a Device |
US11663780B2 (en) | 2012-02-13 | 2023-05-30 | Hologic Inc. | System and method for navigating a tomosynthesis stack using synthesized image data |
US10082775B2 (en) * | 2012-03-09 | 2018-09-25 | Panasonic Intellectual Property Management Co., Ltd. | Appliance control method and apparatus for controlling a target appliance in a building to disguise the number of people in the building |
US20140114484A1 (en) * | 2012-03-09 | 2014-04-24 | Panasonic Corporation | Appliance control method, appliance control apparatus, control server, and control terminal |
US9287864B2 (en) | 2012-04-11 | 2016-03-15 | Ford Global Technologies, Llc | Proximity switch assembly and calibration method therefor |
US9660644B2 (en) | 2012-04-11 | 2017-05-23 | Ford Global Technologies, Llc | Proximity switch assembly and activation method |
US9219472B2 (en) | 2012-04-11 | 2015-12-22 | Ford Global Technologies, Llc | Proximity switch assembly and activation method using rate monitoring |
US9568527B2 (en) | 2012-04-11 | 2017-02-14 | Ford Global Technologies, Llc | Proximity switch assembly and activation method having virtual button mode |
US9831870B2 (en) | 2012-04-11 | 2017-11-28 | Ford Global Technologies, Llc | Proximity switch assembly and method of tuning same |
US8933708B2 (en) | 2012-04-11 | 2015-01-13 | Ford Global Technologies, Llc | Proximity switch assembly and activation method with exploration mode |
US9197206B2 (en) | 2012-04-11 | 2015-11-24 | Ford Global Technologies, Llc | Proximity switch having differential contact surface |
US9944237B2 (en) | 2012-04-11 | 2018-04-17 | Ford Global Technologies, Llc | Proximity switch assembly with signal drift rejection and method |
US9184745B2 (en) | 2012-04-11 | 2015-11-10 | Ford Global Technologies, Llc | Proximity switch assembly and method of sensing user input based on signal rate of change |
US9559688B2 (en) | 2012-04-11 | 2017-01-31 | Ford Global Technologies, Llc | Proximity switch assembly having pliable surface and depression |
US9065447B2 (en) | 2012-04-11 | 2015-06-23 | Ford Global Technologies, Llc | Proximity switch assembly and method having adaptive time delay |
US9531379B2 (en) | 2012-04-11 | 2016-12-27 | Ford Global Technologies, Llc | Proximity switch assembly having groove between adjacent proximity sensors |
US9520875B2 (en) | 2012-04-11 | 2016-12-13 | Ford Global Technologies, Llc | Pliable proximity switch assembly and activation method |
US9136840B2 (en) | 2012-05-17 | 2015-09-15 | Ford Global Technologies, Llc | Proximity switch assembly having dynamic tuned threshold |
US8981602B2 (en) | 2012-05-29 | 2015-03-17 | Ford Global Technologies, Llc | Proximity switch assembly having non-switch contact and method |
US9337832B2 (en) | 2012-06-06 | 2016-05-10 | Ford Global Technologies, Llc | Proximity switch and method of adjusting sensitivity therefor |
US9641172B2 (en) | 2012-06-27 | 2017-05-02 | Ford Global Technologies, Llc | Proximity switch assembly having varying size electrode fingers |
US9447613B2 (en) | 2012-09-11 | 2016-09-20 | Ford Global Technologies, Llc | Proximity switch based door latch release |
US8922340B2 (en) | 2012-09-11 | 2014-12-30 | Ford Global Technologies, Llc | Proximity switch based door latch release |
US8796575B2 (en) | 2012-10-31 | 2014-08-05 | Ford Global Technologies, Llc | Proximity switch assembly having ground layer |
US9311204B2 (en) | 2013-03-13 | 2016-04-12 | Ford Global Technologies, Llc | Proximity interface development system having replicator and method |
US11589944B2 (en) | 2013-03-15 | 2023-02-28 | Hologic, Inc. | Tomosynthesis-guided biopsy apparatus and method |
US11419565B2 (en) | 2014-02-28 | 2022-08-23 | IIologic, Inc. | System and method for generating and displaying tomosynthesis image slabs |
US11801025B2 (en) | 2014-02-28 | 2023-10-31 | Hologic, Inc. | System and method for generating and displaying tomosynthesis image slabs |
US20150350587A1 (en) * | 2014-05-29 | 2015-12-03 | Samsung Electronics Co., Ltd. | Method of controlling display device and remote controller thereof |
WO2015182935A1 (en) * | 2014-05-29 | 2015-12-03 | Samsung Electronics Co., Ltd. | Method of controlling display device and remote controller thereof |
US10038443B2 (en) | 2014-10-20 | 2018-07-31 | Ford Global Technologies, Llc | Directional proximity switch assembly |
US20160150472A1 (en) * | 2014-11-21 | 2016-05-26 | Samsung Electronics Co., Ltd. | User terminal for controlling display device and control method thereof |
US10015739B2 (en) * | 2014-11-21 | 2018-07-03 | Samsung Electronics Co., Ltd. | User terminal for controlling display device and control method thereof |
US9654103B2 (en) | 2015-03-18 | 2017-05-16 | Ford Global Technologies, Llc | Proximity switch assembly having haptic feedback and method |
US9548733B2 (en) | 2015-05-20 | 2017-01-17 | Ford Global Technologies, Llc | Proximity sensor assembly having interleaved electrode configuration |
US11455754B2 (en) | 2017-03-30 | 2022-09-27 | Hologic, Inc. | System and method for synthesizing low-dimensional image data from high-dimensional image data using an object grid enhancement |
US11445993B2 (en) | 2017-03-30 | 2022-09-20 | Hologic, Inc. | System and method for targeted object enhancement to generate synthetic breast tissue images |
US11403483B2 (en) | 2017-06-20 | 2022-08-02 | Hologic, Inc. | Dynamic self-learning medical image method and system |
US11850021B2 (en) | 2017-06-20 | 2023-12-26 | Hologic, Inc. | Dynamic self-learning medical image method and system |
US20190172339A1 (en) * | 2017-12-05 | 2019-06-06 | Bose Corporation | Context-sensitive remote control device |
US10741060B2 (en) * | 2017-12-05 | 2020-08-11 | Bose Corporation | Context-sensitive remote control device |
US11957497B2 (en) | 2022-03-11 | 2024-04-16 | Hologic, Inc | System and method for hierarchical multi-level feature image synthesis and representation |
Also Published As
Publication number | Publication date |
---|---|
EP2153305A1 (en) | 2010-02-17 |
EP2153305B1 (en) | 2015-06-10 |
US7889175B2 (en) | 2011-02-15 |
WO2009006224A1 (en) | 2009-01-08 |
JP2010532143A (en) | 2010-09-30 |
US20090002217A1 (en) | 2009-01-01 |
EP2153305A4 (en) | 2013-05-01 |
JP5473908B2 (en) | 2014-04-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7889175B2 (en) | Touchpad-enabled remote controller and user interaction methods | |
US8456284B2 (en) | Direction and holding-style invariant, symmetric design, and touch- and button-based remote user interaction device | |
US11676482B2 (en) | System and methods for enhanced remote control functionality | |
US8482678B2 (en) | Remote control and gesture-based input device | |
KR101984590B1 (en) | Display device and controlling method thereof | |
US9285867B2 (en) | Device that manages power provided to an object sensor | |
US8619034B2 (en) | Sensor-based display of virtual keyboard image and associated methodology | |
KR101686741B1 (en) | Display device | |
JP5323070B2 (en) | Virtual keypad system | |
US20110304542A1 (en) | Multi purpose remote control with display | |
KR20190002525A (en) | Gadgets for multimedia management of compute devices for people who are blind or visually impaired | |
EP0725331A1 (en) | Information imput/output device using touch panel | |
EP3098796A1 (en) | Remote control apparatus, method of providing vibration feedback thereof, and display system | |
JP2008052584A (en) | Portable electronic device | |
CA3057241A1 (en) | Grid plate | |
TW202240347A (en) | Button, electronic device, and methods of the electronic device | |
KR20030069865A (en) | Remote control device | |
KR20150009314A (en) | Remote control apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |