US20140313421A1 - Apparatus and method for controlling home network devices - Google Patents
Apparatus and method for controlling home network devices Download PDFInfo
- Publication number
- US20140313421A1 US20140313421A1 US14/320,953 US201414320953A US2014313421A1 US 20140313421 A1 US20140313421 A1 US 20140313421A1 US 201414320953 A US201414320953 A US 201414320953A US 2014313421 A1 US2014313421 A1 US 2014313421A1
- Authority
- US
- United States
- Prior art keywords
- unit
- coordinate
- controlled
- image
- devices
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H04N5/4403—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/02—Details
- H04L12/16—Arrangements for providing special services to substations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2816—Controlling appliance services of a home automation network by calling their functionalities
- H04L12/282—Controlling appliance services of a home automation network by calling their functionalities based on user interaction within the home
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4117—Peripherals receiving signals from specially adapted client devices for generating hard copies of the content, e.g. printer, electronic paper
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4131—Peripherals receiving signals from specially adapted client devices home appliance, e.g. lighting, air conditioning system, metering devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/43615—Interfacing a Home Network, e.g. for connecting the client to a plurality of peripherals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/4363—Adapting the video or multiplex stream to a specific local network, e.g. a IEEE 1394 or Bluetooth® network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2823—Reporting information sensed by appliance or service execution status of appliance services in a home automation network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L2012/284—Home automation networks characterised by the type of medium used
- H04L2012/2841—Wireless
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L2012/2847—Home automation networks characterised by the type of home appliance used
- H04L2012/2849—Audio/video appliances
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L2012/2847—Home automation networks characterised by the type of home appliance used
- H04L2012/285—Generic home appliances, e.g. refrigerators
-
- H04N2005/4435—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42226—Reprogrammable remote control devices
Definitions
- Methods and apparatuses consistent with the present invention relate to a home network, and more particularly, to an apparatus and method that can efficiently control devices on a home network.
- a remote controller is usually included with a digital device such as a digital versatile disk (DVD) player, a video cassette recorder (VCR), and a digital television (TV), in order to remotely control a function of the corresponding digital device.
- DVD digital versatile disk
- VCR video cassette recorder
- TV digital television
- the remote controller is usually limited to a specific digital device. Accordingly, a user cannot use one remote controller to remotely control a plurality of digital devices.
- an integrated remote controller capable of remotely controlling two or more digital devices.
- the integrated remote controller also has a problem in that the range of digital devices that can be remotely controlled is limited.
- a user wants to use a specific digital device on a conventional home network, it is difficult to select a digital device connected to the network.
- a list including identification information of a digital device connected to the home network is displayed via a video device such as a digital TV. Then, a user selects a corresponding digital device on the list.
- all digital devices that exist in a home are included in the controlled device list displayed via the video device. If there is a DVD player in every room, a device list including identification information of all DVD players is displayed.
- a user wants to use a DVD player located in living room, it is difficult for the user to distinguish from the list which item corresponds to the DVD player in the living room.
- the present invention provides a user interface using a pointing object in order to easily control a portable broadcast-receiving device.
- an apparatus for controlling home network devices including a display unit that displays a device list on controlled devices arranged in certain spaces, a receiving unit that receives a first image containing the arranged controlled device, the list, and information on light reflected from the arranged controlled devices and the device list, and a control unit that maps the arranged controlled device and a controlled device selected in the list according to the light information detected from the received image.
- an apparatus for controlling home network devices including a receiving unit that receives an image containing a controlled device arranged in a certain space and information on light reflected from the controlled device, a control unit that selects the controlled device, in which the light is reflected, according to the light information detected from the received image, and a display unit that displays a control menu list of selected controlled devices.
- a method of controlling home network devices including displaying a device list on controlled devices arranged in certain spaces, receiving a first image containing the arranged controlled device, the list, and information on light reflected from the arranged controlled device and the list, and mapping the arranged controlled device and a controlled device selected in the list according to the light information detected from the received image.
- a method of controlling home network devices including receiving an image containing a controlled device arranged in a certain space and information on light reflected from the controlled device, selecting the controlled device, in which the light is reflected, according to the light information detected the received image, and displaying a control menu list of selected controlled devices.
- FIG. 1 depicts an apparatus for controlling home network devices according to an exemplary embodiment of the present invention
- FIG. 2 is a block diagram showing a configuration of the home-network-device-controlling apparatus depicted in FIG. 1 ;
- FIGS. 3A to 3F depict a screen displayed in the home-network-device-controlling apparatus depicted in FIG. 1 ;
- FIGS. 4A to 4E depict an image transmitted to the home-network-device-controlling apparatus depicted in FIG. 1 ;
- FIG. 5 illustrates a first mapping table according to an exemplary embodiment of the present invention
- FIG. 6 illustrates a second mapping table according to an exemplary embodiment of the present invention
- FIG. 7 illustrates a third mapping table according to an exemplary embodiment of the present invention.
- FIG. 8 is a flowchart of a home-network-device-controlling method according to an exemplary embodiment of the present invention.
- FIGS. 9 to 11 are flowcharts illustrating in detail operations of the apparatus for controlling a home network device in the method of controlling home network devices depicted in the FIG. 8 .
- These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks.
- the computer program instructions may also be loaded into a computer or other programmable data processing apparatus to cause a series of operational steps to be performed in the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute in the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
- each block of the flowchart illustrations may represent a module, segment, or portion of code, which includes one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of order. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in reverse order, depending upon the functionality involved.
- the wired network technologies can be classified into wired and wireless network technologies.
- the wired network technologies are realized by home phoneline network alliance (PNA), power line control (PLC), Ethernet, IEEE 1394, and universal serial bus (USB).
- the wireless network technologies are realized by infrared data association (IrDA), IEEE 802.11x wireless local area networks (IEEE 802.11), home radio frequency RF (IEEE 802.15), Bluetooth (IEEE 802.15), ultra wide band (UWB), Zigbee, and hiper local area networks (LANs).
- IrDA infrared data association
- IEEE 802.11x wireless local area networks IEEE 802.11
- home radio frequency RF IEEE 802.15
- Bluetooth IEEE 802.15
- UWB ultra wide band
- Zigbee Zigbee
- LANs hiper local area networks
- Home PNA is a wired network technology that can enable a home network with a transfer rate of 10 Mbps using a home phone line; thus, alleviating the need for an additional communication line.
- Home PNA is compatible with IEEE 802.3, Ethernet networks, modems, and integrated services digital networks (ISDNs). Since using higher bandwidth than bandwidth used by phone or other services, home PNA can keep out of jamming on a home phoneline even while the network is used.
- Home PNA enables all computers on the network to share the Internet even though only one computer is connected to the Internet.
- Home PNA can be combined with 10 Mbps Ethernet or 56 Kbps modem, and it allows several users to play an Internet game and chat at the same time via one home phone line.
- PLC wired networks converts a communication signal into a high frequency in the range of 100 KHz through 30 MHz, and then transmits it. After receiving a communication signal, the PLC network filters it using a high frequency filter.
- Ethernet is an LAN technology based on the IEEE 802.3 standard. Ethernet has become widespread it is advantageous in that it, it is safety, and it is cheaper than rival technologies.
- FireWire (IEEE 1394) is a serial bus interface technology that enables up to 63 terminals to be connected; it was originally developed to a hard-disk interface. IEEE 1394 is becoming more popular due to the development of digital and multimedia environments. Since IEEE 1394 supports high transfer rates of 100 Mbps, 200 Mbps, and 400 Mbps according to the mode, it can perform high-speed real-time data transfer, two-way communication, and Plug and Play (PnP).
- USB has a human interface device (HID) characteristic and sufficient interrupt request (IRQ) resources, enabling to be easily connected peripheral devices, and supporting PnP technology.
- HID human interface device
- IRQ interrupt request
- IEEE 802.11x wireless LAN has emerged as an important wireless home network technology as it does not need wiring, can quickly establish an LAN, and can easily change a net structure.
- wireless technology standards such as IEEE 802.11b, 802.11a, 802.11g, 802.11e, and 802.11i.
- HomeRF has become a standard, and it is based on the shared wireless access protocol (SWAP) 1.1 standard.
- SWAP shared wireless access protocol
- the goal of HomeRF is to build a home network using RF instead of infrared communication rays.
- Bluetooth As a low-power and low-cost short range wireless communication technology that enables communication between several devices in a 10 m range, Bluetooth was limited to an applicable range in the first stage, but currently is a PNA technology used between personal communication devices such as mobile phones and personal digital assistants (PDAs), personal computer (PC) peripheral devices such as headsets, keyboards, speakers and printers, and devices connected to a PC by wire.
- PDAs personal digital assistants
- PC personal computer
- UWB was used for special purposes such as military radar and remote-control, but it mainly was released to the public domain by the Federal Communications Commission (FCC) in February 2002.
- FCC Federal Communications Commission
- a sample UWB chip set for a wireless moving-image transmission was announced in the first stage. After 2003, it is expected that a 100 Mbps chip will be developed for a wireless moving image transmission in home.
- a home network device controlling system (hereinafter, referred to as a “control system”) will be described with reference to FIG. 1 .
- FIG. 1 depicts an apparatus for controlling home network devices according to an exemplary embodiment of the present invention.
- the illustrated control system includes controlled devices 400 , 500 , and 600 , which are arranged in a predetermined space, a laser pointer 100 , a video device 300 that controls the controlled devices selected by a user, and a image-capturing device 200 that obtains an image containing information on light reflected from the controlled devices, the video device 300 and the laser pointer 100 .
- the controlled devices 400 , 500 and 600 may be configured to a digital device.
- the digital device refers to a device having a digital circuit that processes digital data, and may be a computer, a printer, a scanner, a pager, a digital camera, a facsimile machine, a digital copier, a PDA, cellular phone, digital phone, digital projector, digital video recorder, digital camcorder, set-top box, digital refrigerator, microwave oven, electric rice-cooker, air cleaner, portable cell phone, VCR, or DVD player.
- the controlled devices are arranged in a predetermined space, and may be connected to the video device via a network, e.g., UPnP, Bluetooth, Home PNA.
- the controlled devices 400 , 500 and 600 which are connected to the video device 300 via the network, provide certain data, e.g., identification information and a control menu of the controlled devices, to the video device 300 .
- the controlled devices receive a control code from the video device 300 , and perform an instruction corresponding to the received control code.
- the control system according to an exemplary embodiment of the present invention includes a first control device 400 , a second controlled device 500 and a third controlled devices 600 .
- the first, second and third controlled devices are a DVD player 400 , a VCR 500 , and a lighting switch 600 , which will be described as an example.
- the video device 300 is connected to the plurality of controlled devices 400 , 500 and 600 via the network, and stores identification and control menu information of the controlled devices and a control instruction for controlling the controlled device.
- the video device 300 transmits the control instruction for controlling the controlled devices selected by a user via the network.
- the video device 300 may be configured as a digital device including a display unit such as a digital TV and a computer monitor.
- the video device 300 is configured as a digital TV, which will be given as an example. A detailed explanation of the video device 300 will be described with reference to FIG. 2 .
- the laser pointer 100 When a power button (not shown) is pressed, the laser pointer 100 emits a light of a specific wavelength, i.e., a laser.
- the center of the laser beamspot 100 is white, and the remaining area is red.
- a user may point to a specific object, e.g., the controlled devices 400 , 500 and 600 located in a certain space or the video device 300 , using the laser pointer 100 , and may visually recognize the area pointed to by the laser pointer 100 .
- a user may select an item of specific objects, e.g., a controlled device list (refer to reference number 3 of FIG. 3D ), displayed on the video device 300 .
- the laser pointer 100 includes a select button (not shown).
- the select button generates a selection event when pressed by a user.
- the selection event refers to an event that can select the controlled devices or the video device 300 pointed to by the laser pointer 100 ; for example, a laser emitted by the laser pointer 100 is turned on and off a certain number of times for a certain amount of time.
- the image-capturing device 200 provides an image including information on the laser (refer to reference number 10 of FIG. 3D ) reflected by the controlled devices 400 , 500 and 600 , the video device 300 and the laser pointer 100 to the video device 300 .
- the image-capturing device 200 is set up at a certain space in the home in order for the controlled devices 400 , 500 and 600 , and the video device 300 to be included within photographic range; for example, it may be set up in a direction facing a wall where the controlled devices and the video device 300 are located, as illustrated in FIG. 1 .
- the image-capturing device 200 is set up as illustrated FIG. 1 , which will be described as an example.
- FIG. 2 is a block diagram showing a configuration of a video device according to an exemplary embodiment of the present invention.
- FIGS. 3A through 3F depict a screen displayed in the video device 300 depicted in FIG. 2 .
- FIGS. 4A through 4E depict an image transmitted to a image-capturing device.
- FIGS. 5 through 7 illustrate first, second, and third mapping tables according to an exemplary embodiment of the present invention.
- the video device 300 illustrated in FIG. 2 includes a transmitting unit 320 , a receiving unit 310 , a display unit 390 , a detecting unit 330 , a compensating unit 340 , a coordinate-system-setting unit 360 , a coordinate-converting unit 370 , a storage unit 380 , and a control unit 350 .
- the receiving unit 310 receives identification and control menu information of the controlled devices from the controlled devices.
- the received identification and control menu information are stored in the following storage unit 380 .
- the receiving unit 310 receives an image via the image-capturing unit 200 , and provides it to the compensating unit 340 .
- the sending unit 320 transmits an instruction for controlling the controlled devices selected by a user to the corresponding controlled devices via a network.
- the display unit 390 displays mesh data having a certain color; for example, it displays red, green and blue mesh data, as illustrated in FIGS. 3A through 3C .
- the display unit 390 enables the detecting unit 330 to detect a display area 30 of the video device 300 , of images provided by the image-capturing device 200 , and enables the compensating unit 340 to compensate the detected display area 30 .
- the display unit 390 may display data having various patterns other than the mesh data. For example, the display unit 390 may display data having a pattern of circle, triangle, or quadrangle, but is not limited thereto. Colors of the mesh data displayed via the display unit 390 are not limited to red, green, and blue; for example, yellow mesh data may be displayed.
- red, green, and blue mesh data is displayed, which is merely exemplary.
- the display unit 390 displays a controlled device list 3 stored in the storage unit 380 , as illustrated in FIG. 3D , or a control menu list 4 of the controlled device selected by a user, e.g., a lighting switch 600 .
- the detecting unit 330 detects the display area 30 of the video device 300 in the image provided by the receiving unit 310 . More particularly, when an image including mesh data is successively received, as illustrated in FIGS. 4A through 4C , the detecting unit 330 detects the display unit 30 of the video device 300 by scanning an area in the received image where the pattern exists. Information on the detected display area 30 is provided to the compensating unit 340 .
- the detecting unit 330 detects areas of the controlled devices 400 , 500 and 600 from the image provided by the receiving unit 310 .
- the detecting unit 330 detects the areas of the controlled devices according to a certain algorithm, e.g., an edge detection algorithm, stored in the storage unit 380 .
- a certain algorithm e.g., an edge detection algorithm
- the detecting unit 330 detects coordinates included in the detected controlled device areas.
- the detecting unit 330 provides information on the detected controlled device areas and the coordinates included in the corresponding areas, i.e., coordinate information of a first coordinate system, to the following control unit 350 .
- the detecting unit 330 detects a coordinate of an area, in which a selection event is generated, from the provided image, and provides coordinate information of the area, in which a selection event is generated, to the following control unit 350 .
- the compensating unit 340 compensates the display area 30 with reference to mesh data displayed via the display unit 390 . More particularly, the compensating unit 340 compares a mesh data pattern displayed via the display unit 390 and a mesh data pattern detected by the detecting unit 330 , thereby compensating a display area 30 distorted from the image.
- a compensation value e.g., a size value or color compensation ratio, used when the compensating unit 340 compensates the image, is stored in the storage unit 380 .
- the compensating unit 340 may compensate an image provided later using the compensation value stored in the storage unit 380 .
- the coordinate-system-setting unit 360 sets a first coordinate system for the image received by the receiving unit 310 .
- the coordinate-system-setting unit 360 sets a second coordinate system corresponding to the display area 30 compensated by the compensating unit 340 .
- the coordinate-system-setting unit 360 sets a second coordinate system that has the upper-left vertex of the compensated display area 30 as the origin, and the second coordinate system is stored in the following storage unit 380 .
- the storage unit 380 stores an algorithm needed to detect the display area 30 from the image provided by the image-capturing device 200 and an algorithm needed to detect the controlled device areas, e.g., the edge-detection algorithm.
- the storage unit 380 stores the first, second, and third mapping tables 50 , 40 and 60 generated by the following control unit 350 .
- the storage unit 380 stores identification information, control menu information, and an instruction code for the corresponding control menu received by the controlled devices 400 , 500 and 600 .
- the storage unit 380 also stores information on a compensation value needed to compensate the display area 30 and the second coordinate system set by the coordinate-system-setting unit 360 .
- the storage unit 360 may be, but is not limited to, at least one of a nonvolatile memory device such as a cache, read-only memory (ROM), programmable ROM (PROM), erasable PROM (EPROM), electric EPROM (EEPROM), or a flash memory, a volatile memory device such as random access memory (RAM), or a storage medium such as hard disk drive (HDD).
- a nonvolatile memory device such as a cache, read-only memory (ROM), programmable ROM (PROM), erasable PROM (EPROM), electric EPROM (EEPROM), or a flash memory
- RAM random access memory
- HDD hard disk drive
- the control unit 350 generates first, second and third mapping tables 50 , 40 and 60 . More particularly, the control unit 350 generates the first mapping table 50 illustrated in FIG. 5 by mapping information, e.g., information on the controlled device areas 51 detected from the image and coordinate values 52 of the first coordinate system included in the corresponding areas, provided by the detecting unit 330 .
- mapping information e.g., information on the controlled device areas 51 detected from the image and coordinate values 52 of the first coordinate system included in the corresponding areas, provided by the detecting unit 330 .
- the control unit 350 When the control device list 3 is displayed via the display unit 390 after the second coordinate system is set by the coordinate-system-setting unit 360 , the control unit 350 generates the second mapping table 40 , illustrated in FIG. 6 , by mapping items 41 displayed in the display area 30 and coordinate values 42 of the second coordinate system included in the area where each item is displayed.
- the control unit 350 generates the third mapping table 60 illustrated in FIG. 7 by mapping coordinate information 62 of the first mapping table 50 and controlled device identification information 61 of the second mapping table 40 .
- the detecting unit 330 detects a coordinate of an area, in which the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to the control unit 350 .
- the control unit 350 determines whether the coordinate value provided by the detecting unit 330 is included in the display area 30 . If the result shows that the coordinate value provided by the detecting unit 330 is not included in the display area 30 , the control unit 350 selects a coordinate group including the coordinate value of the area, in which the selection event is generated, by searching for the first mapping table 50 .
- the detecting unit 330 detects a coordinate, i.e., a coordinate of the first coordinate system of an area where the selection event is generated, and then provides it to the control unit 350 .
- the control unit 350 determines whether the coordinate value provided by the detecting unit 330 is included in the display area 30 . If the result shows that the coordinate value provided in the detecting unit 330 is not included in the display area 30 , the control unit 350 provides the detected coordinate to the coordinate-converting unit 370 .
- the coordinate-converting unit 370 converts the coordinate value of the first coordinate system into the coordinate value of the second coordinate system.
- the converted coordinate value is provided to the control unit 350 .
- control unit 350 selects a controlled device including the converted coordinate value with reference to the second mapping table 40 .
- control unit 350 generates the third mapping table 60 illustrated in FIG. 7 by mapping a coordinate group selected in the first mapping table 50 and the controlled device selected in the second mapping table 40 .
- the control unit 350 selects the controlled device including a coordinate of an area where the selection event is generated, with reference to the third mapping table 60 .
- the control unit 350 enables the display unit 390 to display a control menu corresponding to the selected controlled device.
- FIGS. 8 through 11 A method of controlling home network devices according to an exemplary embodiment of the present invention will be described with reference to FIGS. 8 through 11 .
- FIG. 8 is a flowchart of a method of controlling home network devices according to an exemplary embodiment of the present invention.
- FIGS. 9 to 11 are flowcharts illustrating in detail operations of the apparatus for controlling a home network device of the home network device controlling method depicted in the FIG. 8 .
- the method of controlling home network devices includes setting a coordinate system (S 710 ), mapping a controlled device arranged in a predetermined space and an item corresponding to the controlled device in the controlled device list 3 displayed in the display unit 390 of the video device 300 (S 720 ), and controlling the controlled device.
- the control unit 350 of the video device 300 enables the display unit 390 to display a screen having a predetermined pattern (S 711 ). For example, red, green and blue mesh data is successively displayed as illustrated in FIGS. 3A through 3C . If a predetermined color of mesh data is displayed, an image including the video device 300 displaying the mesh data is received via the receiving unit 310 (S 712 ).
- the detecting unit 330 detects the display area 30 of the video device 300 in the image received via the receiving unit 310 . That is, the detecting unit 330 detects the display area 30 of the video device 300 by detecting a pattern of mesh data.
- the compensating unit 340 compares the mesh data pattern displayed via the display unit 390 and the mesh data pattern detected by the detecting unit 330 , thereby compensating the display area 30 S 713 .
- the compensating unit 340 compensates an image color by comparing a mesh data color displayed via the display unit 390 and a mesh data color detected by the detecting unit 330 .
- the coordinate-system-setting unit 360 sets the second coordinate system on the basis of the display area 30 of the video device 300 (S 714 ). For example, the coordinate-system-setting unit 360 sets a coordinate system that has the upper-left vertex of the compensated display area 30 as the origin ( FIG. 4A ), thus completing coordinate setting for the image.
- the control unit 350 When the second coordinate system for the compensated display area 30 is set, the control unit 350 generates the first mapping table. More particularly, the detecting unit 330 detects an area of the controlled devices 400 , 500 and 600 from the image received via the receiving unit 310 . The detecting unit 330 may detect the area using a predetermined algorithm, e.g., an edge detection algorithm. When the area is detected, the detecting unit 330 provides information on the detected area and coordinates included in the corresponding area to the control unit 350 .
- a predetermined algorithm e.g., an edge detection algorithm
- the control unit 350 generates the first mapping table 50 illustrated in FIG. 5 by creating a table of information on the detected area and coordinates included in the corresponding area.
- the first mapping table 50 may be changed according to the area where the controlled devices exist. That is, if a new controlled device is located within image-capturing range of the image-capturing device 200 , information on an area of the new controlled device is added to the first mapping table 50 . However, unless the existing controlled device is located within image-capturing range of the image-capturing device 200 , information on the corresponding area is deleted in the first mapping table 50 .
- generating the first mapping table 50 may be performed before or after setting a coordinate system (S 710 ).
- the controlled device 350 of the video device 300 enables the controlled device list 3 to be displayed via the display unit 390 (S 721 ).
- the controlled device list 3 may include all the controlled devices 400 , 500 and 600 . That is, a list on all controlled devices 400 , 500 and 600 connected to a network may be displayed via the display unit 390 even though new controlled devices are not included within image-capturing range of the image-capturing device 200 .
- a user selects the controlled devices and the items of the displayed controlled device list 3 using the laser pointer 100 .
- a user may first select a controlled device 400 , 500 or 600 or first select an item of the controlled device list 3 displayed on the video device 300 .
- the former will be described as an example.
- a user points to a controlled device using the laser pointer 100 .
- a user points the lighting switch 600 as illustrated in FIG. 4D .
- the user presses a select button (not shown) included in the laser pointer 100 .
- the select button of the laser pointer 100 is pressed, a light emitted by the laser pointer 100 is turned on and off a predetermined number of times for a predetermined amount of time. That is, a selection event is generated.
- the detecting unit 330 of the video device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to the control device 350 (S 722 ).
- the control device 350 determines whether the detected coordinate is included in the display area 30 (S 723 ). As a result, when the detect coordinate is not included in the display area 30 (“NO” in S 723 ), the control device 350 searches the first mapping table 50 stored in the storage unit 380 . Then, the control device 350 selects a coordinate group including the coordinate in the area where the selection event is generated (S 724 ). If a coordinate of the area where the selection event is generated in FIG. 4D is (X 21 , Y 21 ), the control unit 350 selects a coordinate group corresponding to the third area in the first mapping table 50 .
- a user points to an item of the controlled device list displayed via the display unit 390 using the laser pointer 100 .
- a user points to a switch item as illustrated in FIG. 3D .
- the select button of the laser pointer 100 is pressed, a selection event is generated through that a light emitted by the laser pointer 100 which is turned on and off a predetermined number of times for a predetermined amount of time.
- the detecting unit 330 of the video device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to the control device 350 (S 722 ).
- the control device 350 determines whether the detected coordinate is included in the display area 30 (S 723 ). As a result, when the detected coordinate is included in the display area 30 (“YES” in S 723 ), the control device 350 provides the detected coordinate to the coordinate-converting unit 370 .
- the coordinate-converting unit 370 converts the coordinate value provided by the detecting unit 330 into a coordinate value corresponding to the second coordinate system, and then provides it to the control unit 350 (S 726 ).
- control unit 350 selects controlled devices 400 , 500 and 600 including the converted coordinate with reference to the second mapping table 40 (S 727 ). If a coordinate value converted by the coordinate-converting unit 370 is (X 21 , Y 21 ), the control unit 350 selects a controlled device including (X 21 , Y 21 ) in the second mapping table 40 , i.e., the lighting switch 600 .
- control unit 350 maps a coordinate group selected in the first mapping table 50 , i.e., a coordinate group corresponding to the third area, and a controlled device selected in the second mapping table 40 , i.e., the lighting switch 600 (S 728 ).
- a user allows the third mapping table 60 as illustrated in FIG. 7 to be generated by repeating the described steps as many times as the number of the controlled devices photographed by the image-capturing device 300 .
- mapping of the controlled devices arranged in a space and the items of the controlled device list 3 displayed on the video device 300 is completed.
- a user points to the controlled devices 400 , 500 and 600 using the laser pointer 100 , and presses the select button. For example, a user points to the lighting switch 600 , and then presses the select button.
- the select button is pressed, a light emitted by the laser pointer 100 is turned on and off a predetermined number of times for a predetermined amount of time. That is, a selection event is generated.
- the detecting unit 330 of the video device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system from the image received via the receiving unit 310 , and provides it to the control device 350 (S 731 ).
- the control device 350 selects the controlled devices 400 , 500 and 600 including a coordinate value provided by the detecting unit 330 with reference to the third mapping table 60 (S 732 ).
- the control device 350 enables the control menu of the selected controlled devices 400 , 500 and 600 to be displayed via the display unit 390 (S 733 ). If a coordinate value provided by the detecting unit 330 is (X 23 , Y 23 ), the control unit 350 selects the lighting switch 600 in the third mapping table 60 . As a result, the control menu list 4 corresponding to the lighting switch is displayed via the display unit 390 as illustrated in FIG. 3E .
- a user points to a desired menu item using the laser pointer 100 , and then selects a pointed menu by pressing the select button. That is, a user generates a selection event. For example, a user points a switch off menu that turns off the lighting, and then generates a selection event by pressing the select button as illustrated in FIG. 3F .
- a select button (not shown) included in the laser pointer 100 .
- a selection event is generated, through that, a laser released by the laser pointer 100 is turned on and off a predetermined number of times for a predetermined time.
- the detecting unit 330 detects a coordinate of an area where the selection event is generated from the received image. Since the detected coordinate is included in the display unit 390 , the coordinate is converted into a coordinate value of the second coordinate system by the coordinate-converting unit 370 . The converted coordinate value is provided to the control unit 350 .
- control unit 350 recognizes a control menu including the converted coordinate value, e.g., a switch off menu, by searching for a table (not shown) including information on a coordinate corresponding to the control menu list 4 of the lighting switch 600 . Then, the control unit 350 transmits an instruction corresponding to the recognized control menu, e.g., a switch off instruction via the transmitting unit 320 to the lighting switch 600 with reference to the information stored in the storage unit 380 .
- a control menu including the converted coordinate value e.g., a switch off menu
- the apparatus for controlling home network devices and method according to the present invention may produce one or more effects described below.
- a plurality of controlled devices can be easily controlled without a plurality of remote controllers.
- information on a position of a controlled device can be provided to a user when a controlled device list is displayed.
Abstract
Description
- This application is a continuation application of U.S. application Ser. No. 11/601,639, filed Nov. 20, 2006, which claims priority from Korean Patent Application No. 10-2005-0129189 filed on Dec. 24, 2005 in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
- 1. Field of the Invention
- Methods and apparatuses consistent with the present invention relate to a home network, and more particularly, to an apparatus and method that can efficiently control devices on a home network.
- 2. Description of the Related Art
- Recently, products aimed at a home network market are being developed, while the use of various in home digital devices is becoming widespread.
- A remote controller is usually included with a digital device such as a digital versatile disk (DVD) player, a video cassette recorder (VCR), and a digital television (TV), in order to remotely control a function of the corresponding digital device. However, the remote controller is usually limited to a specific digital device. Accordingly, a user cannot use one remote controller to remotely control a plurality of digital devices.
- To solve this problem, an integrated remote controller capable of remotely controlling two or more digital devices has been developed. However, the integrated remote controller also has a problem in that the range of digital devices that can be remotely controlled is limited.
- Also, if several taskpads are being operated in a video device such as a digital TV, it is difficult to select a specific taskpad via the conventional remote controller.
- Similarly, if a user wants to use a specific digital device on a conventional home network, it is difficult to select a digital device connected to the network. Specifically, in the conventional home network, a list including identification information of a digital device connected to the home network is displayed via a video device such as a digital TV. Then, a user selects a corresponding digital device on the list. However, all digital devices that exist in a home are included in the controlled device list displayed via the video device. If there is a DVD player in every room, a device list including identification information of all DVD players is displayed. Here, if a user wants to use a DVD player located in living room, it is difficult for the user to distinguish from the list which item corresponds to the DVD player in the living room.
- Several proposals, for example, Korean Unexamined Patent No. 2002-0047419: “Pointing Device Using Laser and Camera,” have been suggested. However, the above-describe problem still remains to be solved.
- Accordingly, an apparatus and method that can efficiently control a plurality of digital devices and easily perform an operation in a predetermined video device is needed.
- The present invention provides a user interface using a pointing object in order to easily control a portable broadcast-receiving device.
- According to an aspect of the present invention, there is provided an apparatus for controlling home network devices, the apparatus including a display unit that displays a device list on controlled devices arranged in certain spaces, a receiving unit that receives a first image containing the arranged controlled device, the list, and information on light reflected from the arranged controlled devices and the device list, and a control unit that maps the arranged controlled device and a controlled device selected in the list according to the light information detected from the received image.
- According to another aspect of the present invention, there is provided an apparatus for controlling home network devices, the apparatus including a receiving unit that receives an image containing a controlled device arranged in a certain space and information on light reflected from the controlled device, a control unit that selects the controlled device, in which the light is reflected, according to the light information detected from the received image, and a display unit that displays a control menu list of selected controlled devices.
- According to further aspect of the present invention, there is provided a method of controlling home network devices, the method including displaying a device list on controlled devices arranged in certain spaces, receiving a first image containing the arranged controlled device, the list, and information on light reflected from the arranged controlled device and the list, and mapping the arranged controlled device and a controlled device selected in the list according to the light information detected from the received image.
- According to still another aspect of the present invention, there is provided a method of controlling home network devices, the method including receiving an image containing a controlled device arranged in a certain space and information on light reflected from the controlled device, selecting the controlled device, in which the light is reflected, according to the light information detected the received image, and displaying a control menu list of selected controlled devices.
- The above and other aspects of the present invention will become apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings, in which:
-
FIG. 1 depicts an apparatus for controlling home network devices according to an exemplary embodiment of the present invention; -
FIG. 2 is a block diagram showing a configuration of the home-network-device-controlling apparatus depicted inFIG. 1 ; -
FIGS. 3A to 3F depict a screen displayed in the home-network-device-controlling apparatus depicted inFIG. 1 ; -
FIGS. 4A to 4E depict an image transmitted to the home-network-device-controlling apparatus depicted inFIG. 1 ; -
FIG. 5 illustrates a first mapping table according to an exemplary embodiment of the present invention; -
FIG. 6 illustrates a second mapping table according to an exemplary embodiment of the present invention; -
FIG. 7 illustrates a third mapping table according to an exemplary embodiment of the present invention; -
FIG. 8 is a flowchart of a home-network-device-controlling method according to an exemplary embodiment of the present invention; -
FIGS. 9 to 11 are flowcharts illustrating in detail operations of the apparatus for controlling a home network device in the method of controlling home network devices depicted in theFIG. 8 . - Exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings.
- Advantages and features of the present invention and methods of accomplishing the same may be understood more readily by reference to the following detailed description of the exemplary embodiments and the accompanying drawings. The present invention may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete and will fully convey the concept of the invention to those skilled in the art, and the present invention will only be defined by the appended claims. Like reference numerals refer to like elements throughout the specification.
- The present invention is described hereinafter with reference to flowchart illustrations of user interfaces, methods, and computer program products according to embodiments of the invention. It should be understood that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented by computer program instructions. These computer program instructions can be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart block or blocks. These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks. The computer program instructions may also be loaded into a computer or other programmable data processing apparatus to cause a series of operational steps to be performed in the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute in the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
- And each block of the flowchart illustrations may represent a module, segment, or portion of code, which includes one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of order. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in reverse order, depending upon the functionality involved.
- First, home network technologies will be briefly described before explaining the present invention.
- Home network technologies can be classified into wired and wireless network technologies. The wired network technologies are realized by home phoneline network alliance (PNA), power line control (PLC), Ethernet, IEEE 1394, and universal serial bus (USB). The wireless network technologies are realized by infrared data association (IrDA), IEEE 802.11x wireless local area networks (IEEE 802.11), home radio frequency RF (IEEE 802.15), Bluetooth (IEEE 802.15), ultra wide band (UWB), Zigbee, and hiper local area networks (LANs).
- Home PNA is a wired network technology that can enable a home network with a transfer rate of 10 Mbps using a home phone line; thus, alleviating the need for an additional communication line. Home PNA is compatible with IEEE 802.3, Ethernet networks, modems, and integrated services digital networks (ISDNs). Since using higher bandwidth than bandwidth used by phone or other services, home PNA can keep out of jamming on a home phoneline even while the network is used. Home PNA enables all computers on the network to share the Internet even though only one computer is connected to the Internet. Home PNA can be combined with 10 Mbps Ethernet or 56 Kbps modem, and it allows several users to play an Internet game and chat at the same time via one home phone line.
- PLC wired networks converts a communication signal into a high frequency in the range of 100 KHz through 30 MHz, and then transmits it. After receiving a communication signal, the PLC network filters it using a high frequency filter.
- Ethernet is an LAN technology based on the IEEE 802.3 standard. Ethernet has become widespread it is advantageous in that it, it is safety, and it is cheaper than rival technologies.
- FireWire (IEEE 1394) is a serial bus interface technology that enables up to 63 terminals to be connected; it was originally developed to a hard-disk interface. IEEE 1394 is becoming more popular due to the development of digital and multimedia environments. Since IEEE 1394 supports high transfer rates of 100 Mbps, 200 Mbps, and 400 Mbps according to the mode, it can perform high-speed real-time data transfer, two-way communication, and Plug and Play (PnP).
- USB has a human interface device (HID) characteristic and sufficient interrupt request (IRQ) resources, enabling to be easily connected peripheral devices, and supporting PnP technology.
- However, IEEE 802.11x wireless LAN has emerged as an important wireless home network technology as it does not need wiring, can quickly establish an LAN, and can easily change a net structure. There are several wireless technology standards such as IEEE 802.11b, 802.11a, 802.11g, 802.11e, and 802.11i.
- HomeRF has become a standard, and it is based on the shared wireless access protocol (SWAP) 1.1 standard. The goal of HomeRF is to build a home network using RF instead of infrared communication rays.
- As a low-power and low-cost short range wireless communication technology that enables communication between several devices in a 10 m range, Bluetooth was limited to an applicable range in the first stage, but currently is a PNA technology used between personal communication devices such as mobile phones and personal digital assistants (PDAs), personal computer (PC) peripheral devices such as headsets, keyboards, speakers and printers, and devices connected to a PC by wire.
- UWB was used for special purposes such as military radar and remote-control, but it mainly was released to the public domain by the Federal Communications Commission (FCC) in February 2002. A sample UWB chip set for a wireless moving-image transmission was announced in the first stage. After 2003, it is expected that a 100 Mbps chip will be developed for a wireless moving image transmission in home.
- Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings.
- A home network device controlling system (hereinafter, referred to as a “control system”) will be described with reference to
FIG. 1 . -
FIG. 1 depicts an apparatus for controlling home network devices according to an exemplary embodiment of the present invention. - The illustrated control system includes controlled
devices laser pointer 100, avideo device 300 that controls the controlled devices selected by a user, and a image-capturingdevice 200 that obtains an image containing information on light reflected from the controlled devices, thevideo device 300 and thelaser pointer 100. - As a home network device, the controlled
devices - The controlled devices are arranged in a predetermined space, and may be connected to the video device via a network, e.g., UPnP, Bluetooth, Home PNA. The controlled
devices video device 300 via the network, provide certain data, e.g., identification information and a control menu of the controlled devices, to thevideo device 300. The controlled devices receive a control code from thevideo device 300, and perform an instruction corresponding to the received control code. The control system according to an exemplary embodiment of the present invention includes afirst control device 400, a second controlleddevice 500 and a third controlleddevices 600. The first, second and third controlled devices are aDVD player 400, aVCR 500, and alighting switch 600, which will be described as an example. - The
video device 300 is connected to the plurality of controlleddevices video device 300 transmits the control instruction for controlling the controlled devices selected by a user via the network. Thevideo device 300 may be configured as a digital device including a display unit such as a digital TV and a computer monitor. Thevideo device 300 is configured as a digital TV, which will be given as an example. A detailed explanation of thevideo device 300 will be described with reference toFIG. 2 . - When a power button (not shown) is pressed, the
laser pointer 100 emits a light of a specific wavelength, i.e., a laser. The center of thelaser beamspot 100 is white, and the remaining area is red. A user may point to a specific object, e.g., the controlleddevices video device 300, using thelaser pointer 100, and may visually recognize the area pointed to by thelaser pointer 100. - A user may select an item of specific objects, e.g., a controlled device list (refer to
reference number 3 ofFIG. 3D ), displayed on thevideo device 300. Preferably, but not necessarily, thelaser pointer 100 includes a select button (not shown). The select button generates a selection event when pressed by a user. The selection event refers to an event that can select the controlled devices or thevideo device 300 pointed to by thelaser pointer 100; for example, a laser emitted by thelaser pointer 100 is turned on and off a certain number of times for a certain amount of time. - The image-capturing
device 200 provides an image including information on the laser (refer toreference number 10 ofFIG. 3D ) reflected by the controlleddevices video device 300 and thelaser pointer 100 to thevideo device 300. The image-capturingdevice 200 is set up at a certain space in the home in order for the controlleddevices video device 300 to be included within photographic range; for example, it may be set up in a direction facing a wall where the controlled devices and thevideo device 300 are located, as illustrated inFIG. 1 . The image-capturingdevice 200 is set up as illustratedFIG. 1 , which will be described as an example. - A video device according to an exemplary embodiment of the present invention will be described with reference to
FIGS. 2 through 7 .FIG. 2 is a block diagram showing a configuration of a video device according to an exemplary embodiment of the present invention.FIGS. 3A through 3F depict a screen displayed in thevideo device 300 depicted inFIG. 2 .FIGS. 4A through 4E depict an image transmitted to a image-capturing device.FIGS. 5 through 7 illustrate first, second, and third mapping tables according to an exemplary embodiment of the present invention. - The
video device 300 illustrated inFIG. 2 includes a transmittingunit 320, a receiving unit 310, adisplay unit 390, a detectingunit 330, a compensating unit 340, a coordinate-system-settingunit 360, a coordinate-convertingunit 370, astorage unit 380, and acontrol unit 350. - The receiving unit 310 receives identification and control menu information of the controlled devices from the controlled devices. The received identification and control menu information are stored in the following
storage unit 380. The receiving unit 310 receives an image via the image-capturingunit 200, and provides it to the compensating unit 340. - The sending
unit 320 transmits an instruction for controlling the controlled devices selected by a user to the corresponding controlled devices via a network. - The
display unit 390 displays mesh data having a certain color; for example, it displays red, green and blue mesh data, as illustrated inFIGS. 3A through 3C . Thedisplay unit 390 enables the detectingunit 330 to detect adisplay area 30 of thevideo device 300, of images provided by the image-capturingdevice 200, and enables the compensating unit 340 to compensate the detecteddisplay area 30. Thedisplay unit 390 may display data having various patterns other than the mesh data. For example, thedisplay unit 390 may display data having a pattern of circle, triangle, or quadrangle, but is not limited thereto. Colors of the mesh data displayed via thedisplay unit 390 are not limited to red, green, and blue; for example, yellow mesh data may be displayed. Hereinafter, red, green, and blue mesh data is displayed, which is merely exemplary. - The
display unit 390 displays a controlleddevice list 3 stored in thestorage unit 380, as illustrated inFIG. 3D , or a control menu list 4 of the controlled device selected by a user, e.g., alighting switch 600. - The detecting
unit 330 detects thedisplay area 30 of thevideo device 300 in the image provided by the receiving unit 310. More particularly, when an image including mesh data is successively received, as illustrated inFIGS. 4A through 4C , the detectingunit 330 detects thedisplay unit 30 of thevideo device 300 by scanning an area in the received image where the pattern exists. Information on the detecteddisplay area 30 is provided to the compensating unit 340. - The detecting
unit 330 detects areas of the controlleddevices unit 330 detects the areas of the controlled devices according to a certain algorithm, e.g., an edge detection algorithm, stored in thestorage unit 380. When the controlled device areas are detected, the detectingunit 330 detects coordinates included in the detected controlled device areas. Then, the detectingunit 330 provides information on the detected controlled device areas and the coordinates included in the corresponding areas, i.e., coordinate information of a first coordinate system, to the followingcontrol unit 350. - The detecting
unit 330 detects a coordinate of an area, in which a selection event is generated, from the provided image, and provides coordinate information of the area, in which a selection event is generated, to the followingcontrol unit 350. - The compensating unit 340 compensates the
display area 30 with reference to mesh data displayed via thedisplay unit 390. More particularly, the compensating unit 340 compares a mesh data pattern displayed via thedisplay unit 390 and a mesh data pattern detected by the detectingunit 330, thereby compensating adisplay area 30 distorted from the image. A compensation value, e.g., a size value or color compensation ratio, used when the compensating unit 340 compensates the image, is stored in thestorage unit 380. The compensating unit 340 may compensate an image provided later using the compensation value stored in thestorage unit 380. - The coordinate-system-setting
unit 360 sets a first coordinate system for the image received by the receiving unit 310. The coordinate-system-settingunit 360 sets a second coordinate system corresponding to thedisplay area 30 compensated by the compensating unit 340. For example, the coordinate-system-settingunit 360 sets a second coordinate system that has the upper-left vertex of the compensateddisplay area 30 as the origin, and the second coordinate system is stored in the followingstorage unit 380. - The
storage unit 380 stores an algorithm needed to detect thedisplay area 30 from the image provided by the image-capturingdevice 200 and an algorithm needed to detect the controlled device areas, e.g., the edge-detection algorithm. Thestorage unit 380 stores the first, second, and third mapping tables 50, 40 and 60 generated by the followingcontrol unit 350. Thestorage unit 380 stores identification information, control menu information, and an instruction code for the corresponding control menu received by the controlleddevices storage unit 380 also stores information on a compensation value needed to compensate thedisplay area 30 and the second coordinate system set by the coordinate-system-settingunit 360. Thestorage unit 360 may be, but is not limited to, at least one of a nonvolatile memory device such as a cache, read-only memory (ROM), programmable ROM (PROM), erasable PROM (EPROM), electric EPROM (EEPROM), or a flash memory, a volatile memory device such as random access memory (RAM), or a storage medium such as hard disk drive (HDD). - The
control unit 350 generates first, second and third mapping tables 50, 40 and 60. More particularly, thecontrol unit 350 generates the first mapping table 50 illustrated inFIG. 5 by mapping information, e.g., information on the controlleddevice areas 51 detected from the image and coordinatevalues 52 of the first coordinate system included in the corresponding areas, provided by the detectingunit 330. - When the
control device list 3 is displayed via thedisplay unit 390 after the second coordinate system is set by the coordinate-system-settingunit 360, thecontrol unit 350 generates the second mapping table 40, illustrated inFIG. 6 , by mapping items 41 displayed in thedisplay area 30 and coordinatevalues 42 of the second coordinate system included in the area where each item is displayed. - The
control unit 350 generates the third mapping table 60 illustrated inFIG. 7 by mapping coordinateinformation 62 of the first mapping table 50 and controlled device identification information 61 of the second mapping table 40. - If a selection event is generated when a controlled device is pointed to by the
laser pointer 100, the detectingunit 330 detects a coordinate of an area, in which the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to thecontrol unit 350. - The
control unit 350 determines whether the coordinate value provided by the detectingunit 330 is included in thedisplay area 30. If the result shows that the coordinate value provided by the detectingunit 330 is not included in thedisplay area 30, thecontrol unit 350 selects a coordinate group including the coordinate value of the area, in which the selection event is generated, by searching for the first mapping table 50. - If a selection event is generated when a certain item is pointed to in the
control device list 3 displayed thedisplay unit 390 by thelaser pointer 100, the detectingunit 330 detects a coordinate, i.e., a coordinate of the first coordinate system of an area where the selection event is generated, and then provides it to thecontrol unit 350. - The
control unit 350 determines whether the coordinate value provided by the detectingunit 330 is included in thedisplay area 30. If the result shows that the coordinate value provided in the detectingunit 330 is not included in thedisplay area 30, thecontrol unit 350 provides the detected coordinate to the coordinate-convertingunit 370. - The coordinate-converting
unit 370 converts the coordinate value of the first coordinate system into the coordinate value of the second coordinate system. The converted coordinate value is provided to thecontrol unit 350. - If the converted coordinate value is provided, the
control unit 350 selects a controlled device including the converted coordinate value with reference to the second mapping table 40. - Then, the
control unit 350 generates the third mapping table 60 illustrated inFIG. 7 by mapping a coordinate group selected in the first mapping table 50 and the controlled device selected in the second mapping table 40. - After the third mapping table 60 is generated as described above, if a selection event is generated when the light emitted by the
laser pointer 100 is located on a controlled device, thecontrol unit 350 selects the controlled device including a coordinate of an area where the selection event is generated, with reference to the third mapping table 60. Thecontrol unit 350 enables thedisplay unit 390 to display a control menu corresponding to the selected controlled device. - A method of controlling home network devices according to an exemplary embodiment of the present invention will be described with reference to
FIGS. 8 through 11 . -
FIG. 8 is a flowchart of a method of controlling home network devices according to an exemplary embodiment of the present invention.FIGS. 9 to 11 are flowcharts illustrating in detail operations of the apparatus for controlling a home network device of the home network device controlling method depicted in theFIG. 8 . - Referring to
FIG. 8 , the method of controlling home network devices includes setting a coordinate system (S710), mapping a controlled device arranged in a predetermined space and an item corresponding to the controlled device in the controlleddevice list 3 displayed in thedisplay unit 390 of the video device 300 (S720), and controlling the controlled device. - The setting of a coordinate system (S710) will be described with reference to
FIG. 9 in the following. - The
control unit 350 of thevideo device 300 enables thedisplay unit 390 to display a screen having a predetermined pattern (S711). For example, red, green and blue mesh data is successively displayed as illustrated inFIGS. 3A through 3C . If a predetermined color of mesh data is displayed, an image including thevideo device 300 displaying the mesh data is received via the receiving unit 310 (S712). - Then, the detecting
unit 330 detects thedisplay area 30 of thevideo device 300 in the image received via the receiving unit 310. That is, the detectingunit 330 detects thedisplay area 30 of thevideo device 300 by detecting a pattern of mesh data. - When the
display area 30 of thevideo device 300 is detected by the detectingunit 330, the compensating unit 340 compares the mesh data pattern displayed via thedisplay unit 390 and the mesh data pattern detected by the detectingunit 330, thereby compensating thedisplay area 30 S713. The compensating unit 340 compensates an image color by comparing a mesh data color displayed via thedisplay unit 390 and a mesh data color detected by the detectingunit 330. - When the image color and the
display area 30 is compensated, the coordinate-system-settingunit 360 sets the second coordinate system on the basis of thedisplay area 30 of the video device 300 (S714). For example, the coordinate-system-settingunit 360 sets a coordinate system that has the upper-left vertex of the compensateddisplay area 30 as the origin (FIG. 4A ), thus completing coordinate setting for the image. - When the second coordinate system for the compensated
display area 30 is set, thecontrol unit 350 generates the first mapping table. More particularly, the detectingunit 330 detects an area of the controlleddevices unit 330 may detect the area using a predetermined algorithm, e.g., an edge detection algorithm. When the area is detected, the detectingunit 330 provides information on the detected area and coordinates included in the corresponding area to thecontrol unit 350. - Then, the
control unit 350 generates the first mapping table 50 illustrated inFIG. 5 by creating a table of information on the detected area and coordinates included in the corresponding area. Here, the first mapping table 50 may be changed according to the area where the controlled devices exist. That is, if a new controlled device is located within image-capturing range of the image-capturingdevice 200, information on an area of the new controlled device is added to the first mapping table 50. However, unless the existing controlled device is located within image-capturing range of the image-capturingdevice 200, information on the corresponding area is deleted in the first mapping table 50. As described above, generating the first mapping table 50 may be performed before or after setting a coordinate system (S710). - Mapping a controlled device arranged in a predetermined space and an item of the controlled
device list 3 displayed in thedisplay unit 390 of the video device 300 (S720) will be described with reference toFIG. 10 . - When coordinate setting for the
display area 30 is completed, the controlleddevice 350 of thevideo device 300 enables the controlleddevice list 3 to be displayed via the display unit 390 (S721). The controlleddevice list 3 may include all the controlleddevices devices display unit 390 even though new controlled devices are not included within image-capturing range of the image-capturingdevice 200. - A user selects the controlled devices and the items of the displayed controlled
device list 3 using thelaser pointer 100. A user may first select a controlleddevice device list 3 displayed on thevideo device 300. Hereinafter, the former will be described as an example. - A user points to a controlled device using the
laser pointer 100. For example, a user points thelighting switch 600 as illustrated inFIG. 4D . Then, the user presses a select button (not shown) included in thelaser pointer 100. When the select button of thelaser pointer 100 is pressed, a light emitted by thelaser pointer 100 is turned on and off a predetermined number of times for a predetermined amount of time. That is, a selection event is generated. - The detecting
unit 330 of thevideo device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to the control device 350 (S722). - The
control device 350 determines whether the detected coordinate is included in the display area 30 (S723). As a result, when the detect coordinate is not included in the display area 30 (“NO” in S723), thecontrol device 350 searches the first mapping table 50 stored in thestorage unit 380. Then, thecontrol device 350 selects a coordinate group including the coordinate in the area where the selection event is generated (S724). If a coordinate of the area where the selection event is generated inFIG. 4D is (X21, Y21), thecontrol unit 350 selects a coordinate group corresponding to the third area in the first mapping table 50. - A user points to an item of the controlled device list displayed via the
display unit 390 using thelaser pointer 100. For example, a user points to a switch item as illustrated inFIG. 3D . Then, a user presses a select button (not shown) included in thelaser pointer 100. When the select button of thelaser pointer 100 is pressed, a selection event is generated through that a light emitted by thelaser pointer 100 which is turned on and off a predetermined number of times for a predetermined amount of time. - The detecting
unit 330 of thevideo device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system, and provides it to the control device 350 (S722). - The
control device 350 determines whether the detected coordinate is included in the display area 30 (S723). As a result, when the detected coordinate is included in the display area 30 (“YES” in S723), thecontrol device 350 provides the detected coordinate to the coordinate-convertingunit 370. - The coordinate-converting
unit 370 converts the coordinate value provided by the detectingunit 330 into a coordinate value corresponding to the second coordinate system, and then provides it to the control unit 350 (S726). - Then, the
control unit 350 selects controlleddevices unit 370 is (X21, Y21), thecontrol unit 350 selects a controlled device including (X21, Y21) in the second mapping table 40, i.e., thelighting switch 600. - Then, the
control unit 350 maps a coordinate group selected in the first mapping table 50, i.e., a coordinate group corresponding to the third area, and a controlled device selected in the second mapping table 40, i.e., the lighting switch 600 (S728). - A user allows the third mapping table 60 as illustrated in
FIG. 7 to be generated by repeating the described steps as many times as the number of the controlled devices photographed by the image-capturingdevice 300. As a result, mapping of the controlled devices arranged in a space and the items of the controlleddevice list 3 displayed on thevideo device 300 is completed. - The controlling of controlled devices (S730) will be described with reference to
FIG. 11 . - A user points to the controlled
devices laser pointer 100, and presses the select button. For example, a user points to thelighting switch 600, and then presses the select button. When the select button is pressed, a light emitted by thelaser pointer 100 is turned on and off a predetermined number of times for a predetermined amount of time. That is, a selection event is generated. - The detecting
unit 330 of thevideo device 300 detects a coordinate of an area where the selection event is generated, i.e., a coordinate of the first coordinate system from the image received via the receiving unit 310, and provides it to the control device 350 (S731). - The
control device 350 selects the controlleddevices unit 330 with reference to the third mapping table 60 (S732). Thecontrol device 350 enables the control menu of the selected controlleddevices unit 330 is (X23, Y23), thecontrol unit 350 selects thelighting switch 600 in the third mapping table 60. As a result, the control menu list 4 corresponding to the lighting switch is displayed via thedisplay unit 390 as illustrated inFIG. 3E . - When the control menu list 4 corresponding to the
lighting switch 600 is displayed, a user points to a desired menu item using thelaser pointer 100, and then selects a pointed menu by pressing the select button. That is, a user generates a selection event. For example, a user points a switch off menu that turns off the lighting, and then generates a selection event by pressing the select button as illustrated inFIG. 3F . - Then, a user presses a select button (not shown) included in the
laser pointer 100. When the select button of thelaser pointer 100 is pressed, a selection event is generated, through that, a laser released by thelaser pointer 100 is turned on and off a predetermined number of times for a predetermined time. - The detecting
unit 330 detects a coordinate of an area where the selection event is generated from the received image. Since the detected coordinate is included in thedisplay unit 390, the coordinate is converted into a coordinate value of the second coordinate system by the coordinate-convertingunit 370. The converted coordinate value is provided to thecontrol unit 350. - Then, the
control unit 350 recognizes a control menu including the converted coordinate value, e.g., a switch off menu, by searching for a table (not shown) including information on a coordinate corresponding to the control menu list 4 of thelighting switch 600. Then, thecontrol unit 350 transmits an instruction corresponding to the recognized control menu, e.g., a switch off instruction via the transmittingunit 320 to thelighting switch 600 with reference to the information stored in thestorage unit 380. - As described above, the apparatus for controlling home network devices and method according to the present invention may produce one or more effects described below.
- First, a plurality of controlled devices can be easily controlled without a plurality of remote controllers.
- Second, since a laser pointer is used, several taskpads displayed in the video device can be easily selected.
- Third, information on a position of a controlled device can be provided to a user when a controlled device list is displayed.
- Although the exemplary embodiments of the present invention have been disclosed for illustrative purposes, those skilled in the art will appreciate that various modifications, additions and substitutions are possible, without departing from the scope and spirit of the invention as disclosed in the accompanying claims.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/320,953 US9270917B2 (en) | 2005-12-24 | 2014-07-01 | Apparatus and method for controlling home network devices |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2005-0129189 | 2005-12-24 | ||
KR1020050129189A KR100800998B1 (en) | 2005-12-24 | 2005-12-24 | Apparatus and method for home network device controlling |
US11/601,639 US8797464B2 (en) | 2005-12-24 | 2006-11-20 | Apparatus and method for controlling home network devices |
US14/320,953 US9270917B2 (en) | 2005-12-24 | 2014-07-01 | Apparatus and method for controlling home network devices |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/601,639 Continuation US8797464B2 (en) | 2005-12-24 | 2006-11-20 | Apparatus and method for controlling home network devices |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140313421A1 true US20140313421A1 (en) | 2014-10-23 |
US9270917B2 US9270917B2 (en) | 2016-02-23 |
Family
ID=37887752
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/601,639 Expired - Fee Related US8797464B2 (en) | 2005-12-24 | 2006-11-20 | Apparatus and method for controlling home network devices |
US14/320,953 Expired - Fee Related US9270917B2 (en) | 2005-12-24 | 2014-07-01 | Apparatus and method for controlling home network devices |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/601,639 Expired - Fee Related US8797464B2 (en) | 2005-12-24 | 2006-11-20 | Apparatus and method for controlling home network devices |
Country Status (5)
Country | Link |
---|---|
US (2) | US8797464B2 (en) |
EP (2) | EP1802042A1 (en) |
JP (1) | JP4664272B2 (en) |
KR (1) | KR100800998B1 (en) |
CN (1) | CN100502478C (en) |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8260896B2 (en) | 2007-02-02 | 2012-09-04 | Mwa Intelligence, Inc. | Monitoring business machines using a mesh network on field nodes |
US8107946B2 (en) * | 2007-02-22 | 2012-01-31 | Control4 Corporation | System and method for using a wired network to send response messages in an automation system |
US20090052899A1 (en) * | 2007-08-22 | 2009-02-26 | Motorola, Inc. | Method and apparatus for controlled device selection by a portable electronic device |
KR101362221B1 (en) * | 2007-10-18 | 2014-02-12 | 삼성전자주식회사 | Universal remote control apparatus, universal remote controller system, and method thereof based on batch instruction |
JP2009218711A (en) * | 2008-03-07 | 2009-09-24 | Canon Inc | Information processor, image processor, control method of information processor, control method of image processor and program |
CN101600211B (en) * | 2008-06-06 | 2013-07-17 | 纬创资通股份有限公司 | Wireless communication system for shareable system resource and relative method thereof |
KR101613319B1 (en) | 2008-08-14 | 2016-04-19 | 삼성전자주식회사 | Method and system for inputting service in home network using universal plug and play |
JP5427592B2 (en) * | 2009-12-24 | 2014-02-26 | 株式会社バッファロー | Device control system, terminal device and device |
CN102255947A (en) * | 2011-05-16 | 2011-11-23 | 深圳创维数字技术股份有限公司 | STB (Set Top Box) and man-machine interaction method for STB |
CN102393714B (en) * | 2011-11-02 | 2013-07-24 | 华南理工大学 | Own coding household electrical appliance control system based on ZigBee and realization method thereof |
CN103456150A (en) * | 2012-05-30 | 2013-12-18 | 华硕电脑股份有限公司 | Remote control system and remote control method thereof |
KR101467173B1 (en) | 2013-02-04 | 2014-12-01 | 주식회사 케이티 | Method and Apparatus of resource management of M2M network |
KR101999231B1 (en) | 2013-02-27 | 2019-07-11 | 주식회사 케이티 | Control Unit for Vehicle Components And Mobile Terminal for Vehicle Control |
KR101687340B1 (en) * | 2013-09-12 | 2016-12-16 | 주식회사 케이티 | Method for setting home network operating environment and apparatus therefor |
KR101593115B1 (en) | 2013-10-15 | 2016-02-11 | 주식회사 케이티 | Method for monitoring legacy device status in home network system and home network system |
KR102299980B1 (en) * | 2014-10-07 | 2021-09-09 | 삼성전자주식회사 | Method and electronic device for selecting and controlling a home network device |
US20180113209A1 (en) * | 2016-10-21 | 2018-04-26 | Waymo Llc | Radar generated occupancy grid for autonomous vehicle perception and planning |
KR102407026B1 (en) | 2017-08-31 | 2022-06-08 | 한화테크윈 주식회사 | Home automation system using real-time indoor image |
CN112859636A (en) * | 2021-02-03 | 2021-05-28 | 深圳绿米联创科技有限公司 | Intelligent household control method and device, household control equipment and readable storage medium |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5515486A (en) * | 1994-12-16 | 1996-05-07 | International Business Machines Corporation | Method, apparatus and memory for directing a computer system to display a multi-axis rotatable, polyhedral-shape panel container having front panels for displaying objects |
US5554980A (en) * | 1993-03-12 | 1996-09-10 | Mitsubishi Denki Kabushiki Kaisha | Remote control system |
US20010030668A1 (en) * | 2000-01-10 | 2001-10-18 | Gamze Erten | Method and system for interacting with a display |
US20050068466A1 (en) * | 2003-09-26 | 2005-03-31 | Waters Richard C. | Self-correcting rear projection television |
US20050155052A1 (en) * | 2002-10-04 | 2005-07-14 | Barbara Ostrowska | Parental control for a networked multiroom system |
US20050235329A1 (en) * | 2004-04-19 | 2005-10-20 | Broadcom Corporation | Systems and methods for integrated control within a home entertainment system |
US20060125968A1 (en) * | 2004-12-10 | 2006-06-15 | Seiko Epson Corporation | Control system, apparatus compatible with the system, and remote controller |
US20060248462A1 (en) * | 2005-04-29 | 2006-11-02 | Microsoft Corporation | Remote control of on-screen interactions |
US8089455B1 (en) * | 2006-11-28 | 2012-01-03 | Wieder James W | Remote control with a single control button |
US8508472B1 (en) * | 2006-11-28 | 2013-08-13 | James W. Wieder | Wearable remote control with a single control button |
Family Cites Families (58)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5355149A (en) * | 1992-05-27 | 1994-10-11 | Spacelabs Medical, Inc. | Scanning system for touch screen keyboards |
US5565929A (en) * | 1992-10-13 | 1996-10-15 | Sony Corporation | Audio-visual control apparatus for determining a connection of appliances and controlling functions of appliances |
US5488427A (en) * | 1993-04-16 | 1996-01-30 | Matsushita Electric Industrial Co., Ltd. | Television system including television set, and accessory devices controlled by a single remote control device |
JP3186390B2 (en) * | 1993-12-27 | 2001-07-11 | ソニー株式会社 | Electronic equipment control system and electronic equipment control device |
DE69532662T2 (en) * | 1994-06-09 | 2005-02-10 | Corporation For National Research Initiatives | Hinweisvorrichtungsschittstelle |
US5737440A (en) * | 1994-07-27 | 1998-04-07 | Kunkler; Todd M. | Method of detecting a mark on a oraphic icon |
US5995155A (en) * | 1995-07-17 | 1999-11-30 | Gateway 2000, Inc. | Database navigation system for a home entertainment system |
US5889506A (en) * | 1996-10-25 | 1999-03-30 | Matsushita Electric Industrial Co., Ltd. | Video user's environment |
JP3037161B2 (en) * | 1996-11-08 | 2000-04-24 | 日本電気アイシーマイコンシステム株式会社 | Graphic image display device and graphic image display method |
JP3680886B2 (en) * | 1997-01-29 | 2005-08-10 | 株式会社エクォス・リサーチ | Remote control device |
US5956025A (en) * | 1997-06-09 | 1999-09-21 | Philips Electronics North America Corporation | Remote with 3D organized GUI for a home entertainment system |
WO1998059283A2 (en) * | 1997-06-25 | 1998-12-30 | Samsung Electronics Co., Ltd. | Improved home network, browser based, command and control |
US6127941A (en) * | 1998-02-03 | 2000-10-03 | Sony Corporation | Remote control device with a graphical user interface |
US6501515B1 (en) * | 1998-10-13 | 2002-12-31 | Sony Corporation | Remote control system |
US6532592B1 (en) * | 1998-11-09 | 2003-03-11 | Sony Corporation | Bi-directional remote control unit and method of using the same |
JP2000197158A (en) * | 1998-12-28 | 2000-07-14 | Sanyo Electric Co Ltd | Audio video control system |
US6976267B1 (en) * | 1999-04-09 | 2005-12-13 | Sony Corporation | Method and apparatus for controlling connections between devices |
US6993134B1 (en) * | 1999-08-09 | 2006-01-31 | Koninklijke Philips Electronics N.V. | Key exchange via a portable remote control device |
US6346933B1 (en) | 1999-09-21 | 2002-02-12 | Seiko Epson Corporation | Interactive display presentation system |
US6618076B1 (en) | 1999-12-23 | 2003-09-09 | Justsystem Corporation | Method and apparatus for calibrating projector-camera system |
JP4310674B2 (en) * | 2000-01-11 | 2009-08-12 | ソニー株式会社 | Method for selecting device connected to electronic apparatus and electronic apparatus equipped with this selection method |
JP3975045B2 (en) * | 2000-01-24 | 2007-09-12 | パナソニック コミュニケーションズ株式会社 | Network control device and remote display device |
US6717528B1 (en) * | 2000-02-17 | 2004-04-06 | International Business Machines Corporation | System and method for dynamic feedback projection from a hand-held pointing device |
JP2001236181A (en) | 2000-02-22 | 2001-08-31 | Fuji Electric Co Ltd | Pointing device |
US6791467B1 (en) * | 2000-03-23 | 2004-09-14 | Flextronics Semiconductor, Inc. | Adaptive remote controller |
US7337217B2 (en) * | 2000-07-21 | 2008-02-26 | Samsung Electronics Co., Ltd. | Architecture for home network on world wide web |
US6704000B2 (en) * | 2000-11-15 | 2004-03-09 | Blue Iris Technologies | Method for remote computer operation via a wireless optical device |
KR100359216B1 (en) | 2000-12-13 | 2002-11-04 | 제일모직주식회사 | Slurry for polishing metal layer of semiconductor device |
US20020167617A1 (en) * | 2001-05-11 | 2002-11-14 | Vornsand Steven J. | Closed loop television control system |
JP3811025B2 (en) | 2001-07-03 | 2006-08-16 | 株式会社日立製作所 | Network system |
US6885362B2 (en) * | 2001-07-12 | 2005-04-26 | Nokia Corporation | System and method for accessing ubiquitous resources in an intelligent environment |
US7773852B2 (en) * | 2001-10-12 | 2010-08-10 | Funai Electric Co., Ltd. | Video processing apparatus, video processing method and DVD video apparatus |
JP2003283865A (en) | 2002-03-26 | 2003-10-03 | Matsushita Electric Ind Co Ltd | Apparatus controller |
TW577227B (en) * | 2002-04-23 | 2004-02-21 | Ind Tech Res Inst | Method and apparatus for removing background of visual content |
US20030222849A1 (en) | 2002-05-31 | 2003-12-04 | Starkweather Gary K. | Laser-based user input device for electronic projection displays |
JP2004135051A (en) * | 2002-10-10 | 2004-04-30 | Sony Corp | Information processing system, apparatus and method for providing service, apparatus and method for processing information, recording medium, and program |
CN101101520B (en) | 2002-11-20 | 2012-11-14 | 皇家飞利浦电子股份有限公司 | User interface system based on pointing device |
KR100514191B1 (en) * | 2003-01-23 | 2005-09-13 | 삼성전자주식회사 | remote controller and set-top-box for it |
KR200311058Y1 (en) | 2003-01-30 | 2003-04-21 | 김희동 | Apparatus for controlling electronic devices including laser reception member |
US7256772B2 (en) * | 2003-04-08 | 2007-08-14 | Smart Technologies, Inc. | Auto-aligning touch system and method |
KR100606760B1 (en) * | 2003-07-07 | 2006-07-31 | 엘지전자 주식회사 | Home Network System According to User Preengagement and Control Method of The Same |
US7268802B2 (en) * | 2003-08-20 | 2007-09-11 | Hewlett-Packard Development Company, L.P. | Photography system with remote control subject designation and digital framing |
WO2005052863A2 (en) * | 2003-11-28 | 2005-06-09 | Bracco Imaging S.P.A. | Method and system for distinguishing surfaces in 3d data sets ('dividing voxels') |
KR100996224B1 (en) * | 2003-12-04 | 2010-11-24 | 삼성전자주식회사 | Display device, and method for driving thereof |
US20050198663A1 (en) * | 2003-12-18 | 2005-09-08 | Samsung Electronics Co., Ltd. | User interface method and system for navigation in networked devices |
JP2005227896A (en) * | 2004-02-10 | 2005-08-25 | Pentax Corp | Display selection device and display selection method for electronic device |
JP3882929B2 (en) * | 2004-03-29 | 2007-02-21 | セイコーエプソン株式会社 | Image processing system, projector, and image processing method |
CN100426208C (en) * | 2004-06-07 | 2008-10-15 | 皇家飞利浦电子股份有限公司 | Spatial interaction system |
US7137707B2 (en) * | 2004-07-01 | 2006-11-21 | Mitsubishi Electric Research Laboratories, Inc | Projector-camera system with laser pointers |
JP4617894B2 (en) * | 2005-01-18 | 2011-01-26 | 船井電機株式会社 | Input switching device and television device |
US7728911B2 (en) * | 2005-01-26 | 2010-06-01 | Hewlett-Packard Development Company, L.P. | Positively indicating to user whether signal has been detected on each video input |
JP4259486B2 (en) * | 2005-03-08 | 2009-04-30 | セイコーエプソン株式会社 | Video source search support method, video display apparatus and projector using the search support method |
US7843865B2 (en) * | 2005-12-08 | 2010-11-30 | Intel Corporation | System and method for controlling home network devices using multicast enabled remote controls |
KR100694162B1 (en) * | 2006-02-13 | 2007-03-12 | 삼성전자주식회사 | Method and apparatus for providing location information of device in home network |
US7814441B2 (en) * | 2006-05-09 | 2010-10-12 | Inus Technology, Inc. | System and method for identifying original design intents using 3D scan data |
TW200803482A (en) * | 2006-06-01 | 2008-01-01 | Micro Nits Co Ltd | Image processing method of indicator input system |
KR100753537B1 (en) * | 2006-06-09 | 2007-08-30 | 주식회사 아이너스기술 | Method for reverse modeling using mesh data with feature |
KR101446090B1 (en) * | 2008-05-27 | 2014-10-02 | 삼성전자주식회사 | Control device and method for providing user interface thereof |
-
2005
- 2005-12-24 KR KR1020050129189A patent/KR100800998B1/en not_active IP Right Cessation
-
2006
- 2006-11-20 US US11/601,639 patent/US8797464B2/en not_active Expired - Fee Related
- 2006-12-11 EP EP06125783A patent/EP1802042A1/en not_active Ceased
- 2006-12-11 EP EP14187952.8A patent/EP2840740A1/en not_active Withdrawn
- 2006-12-20 JP JP2006342260A patent/JP4664272B2/en not_active Expired - Fee Related
- 2006-12-22 CN CNB2006101699269A patent/CN100502478C/en not_active Expired - Fee Related
-
2014
- 2014-07-01 US US14/320,953 patent/US9270917B2/en not_active Expired - Fee Related
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5554980A (en) * | 1993-03-12 | 1996-09-10 | Mitsubishi Denki Kabushiki Kaisha | Remote control system |
US5515486A (en) * | 1994-12-16 | 1996-05-07 | International Business Machines Corporation | Method, apparatus and memory for directing a computer system to display a multi-axis rotatable, polyhedral-shape panel container having front panels for displaying objects |
US20010030668A1 (en) * | 2000-01-10 | 2001-10-18 | Gamze Erten | Method and system for interacting with a display |
US20050155052A1 (en) * | 2002-10-04 | 2005-07-14 | Barbara Ostrowska | Parental control for a networked multiroom system |
US20050068466A1 (en) * | 2003-09-26 | 2005-03-31 | Waters Richard C. | Self-correcting rear projection television |
US20050235329A1 (en) * | 2004-04-19 | 2005-10-20 | Broadcom Corporation | Systems and methods for integrated control within a home entertainment system |
US20060125968A1 (en) * | 2004-12-10 | 2006-06-15 | Seiko Epson Corporation | Control system, apparatus compatible with the system, and remote controller |
US20060248462A1 (en) * | 2005-04-29 | 2006-11-02 | Microsoft Corporation | Remote control of on-screen interactions |
US8089455B1 (en) * | 2006-11-28 | 2012-01-03 | Wieder James W | Remote control with a single control button |
US8508472B1 (en) * | 2006-11-28 | 2013-08-13 | James W. Wieder | Wearable remote control with a single control button |
Also Published As
Publication number | Publication date |
---|---|
US9270917B2 (en) | 2016-02-23 |
US8797464B2 (en) | 2014-08-05 |
KR20070067582A (en) | 2007-06-28 |
CN100502478C (en) | 2009-06-17 |
CN1988626A (en) | 2007-06-27 |
EP1802042A1 (en) | 2007-06-27 |
JP4664272B2 (en) | 2011-04-06 |
KR100800998B1 (en) | 2008-02-11 |
US20070165787A1 (en) | 2007-07-19 |
JP2007174662A (en) | 2007-07-05 |
EP2840740A1 (en) | 2015-02-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9270917B2 (en) | Apparatus and method for controlling home network devices | |
KR101442441B1 (en) | Intelligent remote control system | |
US6967695B2 (en) | Remote control network system for remotely controlling a plurality of electric apparatus such as home appliances | |
CN107683465B (en) | Control terminal device, audio system, and audio system control method | |
US8550337B2 (en) | Systems and methods for programming a remote control device | |
US8380242B2 (en) | Methods and systems for audio and video communication | |
KR100534611B1 (en) | Apparatus and method for remote controlling household electric appliances using an wireless terminal | |
US9355557B2 (en) | Universal remote controller and remote control method thereof | |
CN103765826B (en) | Communicator, communication system, communication means and TV receiving system | |
JP2006197248A (en) | Radio communication system, data transmitting device, and radio communication method | |
CN105847952A (en) | Multi-screen linkage control method and multi-screen linkage control device | |
US7589792B2 (en) | Broadcasting signal receiving system including viewing improvement control part | |
TW201503639A (en) | Wireless control system and wireless network expansion method applied thereto | |
US20100325554A1 (en) | Photo-Based Network Control Device and Related Method | |
KR100763175B1 (en) | Method for recognition of user's life pattern and Home network system using by method for recognition of user's life pattern | |
US7995101B2 (en) | Image processing system, image supply apparatus, image receiving apparatus, lighting apparatus and controlling method therefor | |
JP2004235960A (en) | Controller and control method, recording medium, and program | |
US20240146996A1 (en) | Display device and control method therefor | |
US20230335035A1 (en) | Display device and power-off control method thereof | |
JP2003333640A (en) | Radio communication device and radio communication method | |
CN117461318A (en) | Display apparatus and control method thereof | |
KR20110012548A (en) | Operation method for display device | |
JP2008079144A (en) | Remote controller and television receiver |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KIM, JOO-HYUN;KO, YOUNG-GOO;REEL/FRAME:033220/0970 Effective date: 20070108 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20200223 |