US20030118974A1 - Video indexing based on viewers' behavior and emotion feedback - Google Patents
Video indexing based on viewers' behavior and emotion feedback Download PDFInfo
- Publication number
- US20030118974A1 US20030118974A1 US10/023,810 US2381001A US2003118974A1 US 20030118974 A1 US20030118974 A1 US 20030118974A1 US 2381001 A US2381001 A US 2381001A US 2003118974 A1 US2003118974 A1 US 2003118974A1
- Authority
- US
- United States
- Prior art keywords
- behavior
- emotion
- viewers
- video
- categories
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B23/00—Models for scientific, medical, or mathematical purposes, e.g. full-sized devices for demonstration purposes
- G09B23/28—Models for scientific, medical, or mathematical purposes, e.g. full-sized devices for demonstration purposes for medicine
Definitions
- the technical field relates to video imaging system, and, in particular, to video indexing system.
- a video camera typically records both video and audio to generate a video sequence, which can be stored in a secondary storage, such as a hard disk or a CD-ROM.
- video sequences typically have varied content or great length. Since a user normally cannot write down what is on a video sequence or where on a video sequence particular scenes, movies, events are recorded, the user may have to sit and view an entire video sequence to remember what was recorded or to retrieve a particular scene. Video indexing allows a user to have easy access to different sections of the video sequence so that the user do not need to fast forward through the whole video sequence.
- automatic video indexing typically generates different key frames, with the need for extensive post-processing, which involves automatic search for shot changes, scene changes, and ultimately, frames that may serve as key-frames.
- automatic video indexing may or may not help a user find a particular video event within a recording.
- a method for video indexing based on viewers' behavior and emotion feedback includes monitoring one or more viewers during a multimedia sequence play back and detecting the one or more viewers' behavior and emotion feedback. The method further includes categorizing the one or more viewers' behavior and emotion feedback as a plurality of categories of behavior and emotion segments, and indexing the multimedia sequence using the plurality of categories of behavior and emotion segments. As a result, the viewers or other users can selectively view the multimedia sequence using the plurality of categories of behavior and emotion segments as video indices.
- the method for video indexing based on viewers' behavior and emotion feedback is transparent to the user, and may generate custom indexing to a particular user, affording a user-friendly hierarchy for video browsing.
- FIG. 1A illustrates an exemplary emotion detection system that can be connected with a network and a remote storage
- FIG. 1B illustrates how the emotion detection system monitors viewers' feedback
- FIG. 2 illustrates an exemplary hardware components of a computer that may be used to in connection with the exemplary method for video indexing based on viewers' behavior and emotion feedback;
- FIG. 3 illustrates an exemplary chart of behavior and emotion feedback
- FIGS. 4 - 6 illustrate examples for video indexing based on viewers' behavior and emotion feedback
- FIG. 7 is a flow chart illustrating a method for video indexing based on viewers' behavior and emotion feedback.
- a method and a corresponding apparatus for video indexing based on viewers' behavior and emotion feedback automatically index a multimedia sequence, such as a video sequence, using viewers' casual feedback while watching the multimedia sequence.
- the method and apparatus use a system, such as an emotion detection system that includes one or more multimedia acquisition devices, such as video cameras, to detect viewers' behavior and emotion feedback, and to produce as video indices various categories of behavior and emotion segments. In other words, human reactions are detected and indexed into video bitstream.
- Each video camera may include various sensors.
- the viewers or other users, i.e., user may later selectively view the multimedia sequence by browsing through the video indices of various categories of emotion.
- the video indices are similar to chapters within digital video disc (DVD) movies that enable a user to jump directly to a particular chapter without having to fast forward.
- the method and apparatus for video indexing is transparent to the user, and may generate custom indexing to a particular user, affording a user-friendly hierarchy for video browsing.
- the emotion detection system is known in the art, which may be able to detect any type of emotion, some reliable and some not so reliable. Many different categories may be generated to represent different emotions, such as “amused”, “angry”, “astonished”, “curious”, “patient”, “happy”, “proud”, “scared”, and “worried”.
- neuroimaging as a means for validating whether a stimulus such as advertisement, communication, or product evokes a certain mental response such as emotion, preference, or memory, or to predict the consequences of the stimulus on later behavior such as consumption or purchasing.
- Cowie et al. disclose a hybrid psychological and linguistic emotion analysis system capable of using information from faces and voices to recognize people's emotions.
- Video indexing is described, for example, in “Content-Based Browsing of Video Sequences” by Arman et al., ACM multimedia, pages 97-103, 1994; and “Content Based Video Indexing and Retrieval” by Smoliar et al., IEEE multimedia, pages 62-72, 1994, which are incorporated herein by reference.
- Arman et al. disclose a novel methodology to represent the contents of a video sequence. The methodology uses a content-based browsing system that forms an abstraction to represent each shot of the sequence by using a representative frame, and allows a user to easily navigate the frame, i.e., rapidly view a video sequence in order to find a particular point within the sequence.
- Smoliar et al. disclose a method for content-based video indexing and retrieval. The method includes parsing the video stream into generic clips, indexing the video clips when inserted into a database, and retrieval and browsing the database through queries based on text and/or visual examples.
- the method and apparatus for video indexing based on viewers' behavior and emotion feedback can be used in combination with other types of video indexing, such as traditional scene change or shot change indexing, or manual indexing using high resolution still images.
- Video indexing using high resolution still images is described, for example, in U.S. patent application, entitled “Video Indexing Using High Resolution Still Images,” filed on the same day herewith, which is incorporated herein by reference.
- Different types of video indexing may be used in the same video sequence.
- the user can add manual indexing, for example, within two minutes of a “happy” scene, using a still image of a person falling down, causing laughter among people.
- the emotion detection system 100 may detect and categorize different viewers' behavior and emotion feedback every time the same video sequence is watched, and gradually generate more information regarding the video sequence. Therefore, the emotion detection system 100 can iteratively generate more accurate indices based on different viewers' feedback. For example, if every time a video sequence is play back, different viewers laugh while watching a certain segment of the video sequence, the emotion detection system 100 may accurately categorize that segment as “funny”, and index the segment accordingly.
- FIG. 1A illustrates an exemplary emotion detection system 100 that can be connected with a network 160 and a remote storage 150 .
- the exemplary emotion detection system 100 includes one or more multimedia acquisition devices, such as video cameras 130 , and a computer 140 .
- the video cameras 130 point to and monitor one or more viewers during a multimedia sequence, such as a video sequence, play back.
- Each video camera 130 includes an image sensor 110 for capturing facial expression and physical movement of the viewers.
- Each video camera 130 may also include an audio sensor 120 , such as a microphone, for capturing audio signal in the viewers' voice.
- other types of sensor 125 such as electric or temperature sensors, may be used by the emotion detection system 100 for video indexing.
- electrodes may be placed on human head to acquire electrical responses.
- the emotion detection system 100 may include different combinational use of the sensors.
- the emotion detection system 100 uses the computer 140 to categorize different behavior and emotion segments of a scene, and automatically index the video sequence according to various categories of behavior and emotion segments.
- the indexed video sequence may be transmitted to the network 160 and/or the remote storage 150 to be stored.
- the remote storage 150 may be a hard disk, a CD-ROM, a DVD RW, or a server connected to the network 160 .
- FIG. 1B illustrates how the emotion detection system monitors viewers' feedback.
- viewers 180 watch a video sequence using, for example, a digital video player 170 playing on a television set.
- the video camera 130 is connected to the digital video player 170 and the computer 140 .
- the video camera 130 monitors the viewers' behavior and emotion feedback during the video sequence play back.
- the computer 140 detects and categorizes viewers' behavior and emotion feedback, and indexes the video sequence using various categories of behavior and emotion segments as video indices.
- the emotion detection system 100 combines facial expression of the viewers, such as a smile, and audio signal in the viewers' voice, such as a laughter, to categorize the video segment as, for example, “funny”.
- the “funny” segment may then be used to index the video sequence.
- one of the viewers 180 is snoring, that section of the video sequence can be categorized and indexed as, for example, “non-interesting”.
- one of the viewers 180 is shouting, that section of the video sequence can be categorized and indexed as, for example, “interesting”.
- Physical movement may be detected and used as video indices as well.
- Detecting viewers 180 leaving a room may use face detection algorithms.
- that section of the video sequence may be categorized and indexed as “interesting”, especially when people start laughing or commenting on the scene.
- a certain emotion may belong to multiple categories in the emotion detection system 100 .
- a particular segment may be “happy” and “interesting”, so that video indexing using either “happy” or “interesting” emotions may contain the same segment.
- one of the viewers 180 or another user can easily browse through specific categories of behavior and emotion segments within a stored video sequence by jumping to the particular segments directly. For example, a user may watch only “happy” moments within a video sequence.
- different video indices may be generated by different users accordingly to the users' preference. For example, one user may watch “funny” segments of a video sequence, whereas another user may watch “astonished” segments of the same video sequence.
- FIG. 2 illustrates exemplary hardware components of a computer 140 that may be used to in connection with the exemplary method for video indexing based on viewers' behavior and emotion feedback.
- the computer 140 has a connection with a network 160 , such as the Internet or other type of computer or telephone networks, for sending recorded multimedia sequence and video indices to friends and family by, for example, email.
- the computer 140 typically includes a memory 202 , a secondary storage device 212 , a processor 214 , an input device 216 , a display device 210 , and an output device 208 .
- the memory 202 may include random access memory (RAM) or similar types of memory.
- the secondary storage device 212 may include a hard disk drive, floppy disk drive, CD-ROM drive, DVD RW, or other types of non-volatile data storage.
- the secondary storage device 212 may correspond with various databases or other resources.
- the processor 214 may execute applications or other information stored in the memory 202 , the secondary storage 212 , or received from the Internet or other network 160 .
- the input device 216 may include any device for entering data into the computer 140 , such as a keyboard, key pad, cursor-control device, touch-screen (possibly with a stylus), or microphone.
- the display device 210 may include any type of device for presenting visual image, such as, for example, a computer monitor, flat-screen display, or display panel.
- the output device 208 may include any type of device for presenting data in hard copy format, such as a printer, and other types of output devices including speakers or any device for providing data in audio form.
- the computer 140 can possibly include multiple input devices, output devices, and display devices.
- the computer 140 is depicted with various components, one skilled in the art will appreciate that this computer can contain additional or different components.
- aspects of an implementation consistent with the present invention are described as being stored in memory, one skilled in the art will appreciate that these aspects can also be stored on or read from other types of computer program products or computer-readable media, such as secondary storage devices, including hard disks, floppy disks, CD-ROM, or DVD RW; a carrier wave from the Internet or other network; or other forms of RAM or ROM.
- the computer-readable media may include instructions for controlling the computer 140 to perform a particular method.
- FIG. 3 illustrates an exemplary chart of behavior and emotion feedback.
- a viewer's feedback may be very active 301 or very passive 303 , very positive 307 or very negative 305 .
- various of behavior and emotion feedback are possible, for example, surprise 311 , joy 313 , anticipation 321 , acceptance 323 , sadness 331 , disgust 333 , anger 341 , and fear 343 .
- the indexed multimedia sequence may be downloaded to the same or another computer 140 either by transmitting over wireless channels, through a network connection, or through a wired connection, such as universal serial bus (USB) or Firewire (IEEE 1394).
- the computer 140 may read the local storage of the video camera 130 by directly connecting to a reader of the computer 140 .
- the indexed multimedia sequence may be played back on any video display device, such as a computer 140 .
- the video display device may display the video indices as labeled icons for easy indexing, for example, “happy #1”, “funny #2”, or “interesting #5”.
- a particular video index, for example, “happy #1” may be played back by clicking on an associated icon as displayed on the display device 210 .
- the user may browse through the multimedia sequence using the various categories of behavior and emotion segments as video indices. For example, when viewing the recorded multimedia sequence using the computer 140 , a user may double click on one of the icons, for example, “funny”, and start viewing all “funny” segments within the multimedia sequence.
- the indexed multimedia sequence may also be saved on a server connected to the network 160 , to be retrieved by other users.
- the indexed multimedia sequence may be transmitted to other users through the network 160 by, for example, e-mail.
- a friend or a family member who receives the multimedia sequence and the video indices may then easily browse through particular behavior and emotion segments of the multimedia sequence.
- FIGS. 4 - 6 illustrate examples for video indexing based on viewers' behavior and emotion feedback.
- video and audio monitoring 400 may be conducted by the video camera 130 to detect the viewers' casual feedback.
- the emotion detection system 100 categorizes various categories of behavior and emotion segments within a video sequence 480 , such as a “happy” segment 430 and an “astonished” segment 440 .
- a kid crawling on the floor
- a parent may be smiling.
- the emotion detection system 100 detects the emotion and automatically categorizes the segment of the video sequence as “happy” 430 .
- somebody jumps down a trampoline in a swimming pool the viewer may have an astonished expression.
- the emotion detection system 100 detects and categorizes that segment as “astonished” 440 .
- the computer 140 of the emotion detection system 100 indexes the video sequence 480 using the “happy” segment 430 and the “astonished” segment 440 as video indices 435 , 445 , respectively. Thereafter, the viewer or another user may selectively view the video sequence using the video indices 435 , 445 .
- the user can browse through “happy” and “astonished” scenes by selecting “happy” 435 and “astonished” 445 , respectively, on any video display device.
- scene change 410 and shot change 420 may also be used as video indices.
- the video indices of behavior and emotion feedback may be labeled, for example, as “happy #1”, “funny #2”, or “interesting #4”, within the video sequence for easy indexing.
- the emotion detection system 100 includes only audio sensors, such as microphones 120 , so that only audio signals are detected for video indexing.
- Audio feedback is part of emotion feedback.
- “laughter” is part of “happy”, “amused” or “funny” emotion.
- three channels are typically used, such as video, left audio, and right audio channels.
- the method and apparatus for indexing based on viewers' behavior and emotion feedback typically include at least an extra channel for audio feedback indexing to generate a hierarchical indexing structure. Two types of audio signals are generated, one representing system signal in the video sequence, the other representing the viewer's audio feedback. Thereafter, algorithms may be used to remove the viewer's audio signal from the system signal in the video sequence. A more sophisticated system may have a microphone for every viewer.
- the microphone 120 monitors audio signals. Any audio reaction of the viewers 180 may be indexed into the hierarchical index representation as a feedback input. Different techniques may be used to index the video sequence using audio signals.
- FIGS. 5A and 5B illustrate an exemplary embodiment of audio feedback indexing.
- the emotion detection system 100 detects basic oral characteristics, such as exclamations, laughter or speech, and indexes the oral characteristics into the video bitstream. Similar to the example illustrated in FIGS. 4A and 4B, in addition to scene change 510 monitoring, video and audio monitoring 500 may be conducted by the video camera 130 to detect viewers' audio feedback. The emotion detection system 100 then categorizes various categories of audio feedback within a video sequence 580 , such as a “laughter” segment 530 .
- the “laughter” segment 530 typically follows the beginning of a “funny” segment 520 after a brief delay 540 .
- the computer 140 of the emotion detection system 100 indexes the video sequence 580 using the “funny” segment 520 as a video index 525 .
- scene change 510 may also be used as a video index. Thereafter, the viewer or another user may selectively view the video sequence using either one of the video indices 510 and 525 .
- FIG. 6 illustrates another exemplary embodiment of audio feedback indexing.
- the emotion feedback system 100 tries to recognize the speech or other oral characteristics, and add relevant comments as annotations.
- video and audio monitoring 600 is conducted by the video camera 130 to detect viewers' audio feedback.
- the emotion detection system 100 then categorizes various categories of audio feedback within a video sequence 680 , such as a “speech” segment 620 .
- the computer 140 of the emotion detection system 100 indexes the video sequence 680 using the “speech” segment 620 as a video index 625 .
- scene change 610 may also be used as a video index. Thereafter, the viewer or another user may selectively view the video sequence 680 using either one of the video indices 610 and 625 .
- the computer 140 may insert a text annotation 640 , representing the speech itself.
- FIG. 7 is a flow chart illustrating the exemplary method for video indexing based on viewers' behavior and emotion feedback.
- the emotion detection system 100 may use a multimedia acquisition device 130 , such as a video camera, to monitor one or more viewers 180 during a multimedia sequence, such as a video sequence, play back (block 710 ).
- the emotion detection system 100 detect the viewers' behavior and emotion feedback (block 720 ).
- the emotion detection system 100 may use a computer 140 to categorize the viewers' behavior and emotion feedback as various categories of behavior and emotion segments (block 730 ), and index the multimedia sequence using the various categories of behavior and emotion segments (block 740 ).
- the viewers 180 or another user may use any video display device, such as a computer 140 , to selectively view the multimedia sequence using the various categories of behavior and emotion segments as video indices (block 750 ).
- the user may click on an icon from the video display device to browse through various categories of behavior and emotion segments (block 760 ).
- the indexed multimedia sequence may be sent to other users through a network 160 so that friends and family members may selectively view the multimedia sequence using the various categories of behavior and emotion segments as video indices (block 770 ).
Abstract
Description
- The technical field relates to video imaging system, and, in particular, to video indexing system.
- Users are increasingly using video cameras to record home videos, television programs, movies, concerts, or sports events on a disk or DVD for later or repeated viewing. A video camera typically records both video and audio to generate a video sequence, which can be stored in a secondary storage, such as a hard disk or a CD-ROM. Such video sequences typically have varied content or great length. Since a user normally cannot write down what is on a video sequence or where on a video sequence particular scenes, movies, events are recorded, the user may have to sit and view an entire video sequence to remember what was recorded or to retrieve a particular scene. Video indexing allows a user to have easy access to different sections of the video sequence so that the user do not need to fast forward through the whole video sequence.
- Current video indexing devices use video content analysis that automatically or semi-automatically extracts structure and meaning from visual cues in a video. After, for example, a video clip is taken from a television (TV) program or a home video, a computer will generate particular indexes so that a user can jump to a particular section within the video sequence.
- However, automatic video indexing typically generates different key frames, with the need for extensive post-processing, which involves automatic search for shot changes, scene changes, and ultimately, frames that may serve as key-frames. In addition, automatic video indexing may or may not help a user find a particular video event within a recording.
- A method for video indexing based on viewers' behavior and emotion feedback includes monitoring one or more viewers during a multimedia sequence play back and detecting the one or more viewers' behavior and emotion feedback. The method further includes categorizing the one or more viewers' behavior and emotion feedback as a plurality of categories of behavior and emotion segments, and indexing the multimedia sequence using the plurality of categories of behavior and emotion segments. As a result, the viewers or other users can selectively view the multimedia sequence using the plurality of categories of behavior and emotion segments as video indices.
- The method for video indexing based on viewers' behavior and emotion feedback is transparent to the user, and may generate custom indexing to a particular user, affording a user-friendly hierarchy for video browsing.
- The preferred embodiments of the method for video indexing based on viewers' behavior and emotion feedback will be described in detail with reference to the following figures, in which like numerals refer to like elements, and wherein:
- FIG. 1A illustrates an exemplary emotion detection system that can be connected with a network and a remote storage;
- FIG. 1B illustrates how the emotion detection system monitors viewers' feedback;
- FIG. 2 illustrates an exemplary hardware components of a computer that may be used to in connection with the exemplary method for video indexing based on viewers' behavior and emotion feedback;
- FIG. 3 illustrates an exemplary chart of behavior and emotion feedback;
- FIGS.4-6 illustrate examples for video indexing based on viewers' behavior and emotion feedback; and
- FIG. 7 is a flow chart illustrating a method for video indexing based on viewers' behavior and emotion feedback.
- A method and a corresponding apparatus for video indexing based on viewers' behavior and emotion feedback automatically index a multimedia sequence, such as a video sequence, using viewers' casual feedback while watching the multimedia sequence. The method and apparatus use a system, such as an emotion detection system that includes one or more multimedia acquisition devices, such as video cameras, to detect viewers' behavior and emotion feedback, and to produce as video indices various categories of behavior and emotion segments. In other words, human reactions are detected and indexed into video bitstream. Each video camera may include various sensors. The viewers or other users, i.e., user, may later selectively view the multimedia sequence by browsing through the video indices of various categories of emotion. The video indices are similar to chapters within digital video disc (DVD) movies that enable a user to jump directly to a particular chapter without having to fast forward. The method and apparatus for video indexing is transparent to the user, and may generate custom indexing to a particular user, affording a user-friendly hierarchy for video browsing.
- The emotion detection system is known in the art, which may be able to detect any type of emotion, some reliable and some not so reliable. Many different categories may be generated to represent different emotions, such as “amused”, “angry”, “astonished”, “curious”, “patient”, “happy”, “proud”, “scared”, and “worried”.
- The emotion detection is described, for example, in U.S. Pat. No. 5,676,138, entitled “Emotional Response Analyzer System with Multimedia Display,” issued on Oct. 14, 1997 to Zawilinski, U.S. Pat. No. 6,099,319, entitled “Neuroimaging as a Marketing Tool,” issued on Aug. 8, 2000 to Zaltman et al., and “Emotion Recognition in Human-Computer interaction” by Cowie et al, January 2001 on IEEE Signal Processing Magazine, which are incorporated herein by reference. Zawilinski discloses a multimedia computerized system for detecting emotional responses of human beings and the changes therein over time. Zaltman et al. disclose neuroimaging as a means for validating whether a stimulus such as advertisement, communication, or product evokes a certain mental response such as emotion, preference, or memory, or to predict the consequences of the stimulus on later behavior such as consumption or purchasing. Cowie et al. disclose a hybrid psychological and linguistic emotion analysis system capable of using information from faces and voices to recognize people's emotions.
- Video indexing is described, for example, in “Content-Based Browsing of Video Sequences” by Arman et al., ACM multimedia, pages 97-103, 1994; and “Content Based Video Indexing and Retrieval” by Smoliar et al., IEEE multimedia, pages 62-72, 1994, which are incorporated herein by reference. Arman et al. disclose a novel methodology to represent the contents of a video sequence. The methodology uses a content-based browsing system that forms an abstraction to represent each shot of the sequence by using a representative frame, and allows a user to easily navigate the frame, i.e., rapidly view a video sequence in order to find a particular point within the sequence. Smoliar et al. disclose a method for content-based video indexing and retrieval. The method includes parsing the video stream into generic clips, indexing the video clips when inserted into a database, and retrieval and browsing the database through queries based on text and/or visual examples.
- The method and apparatus for video indexing based on viewers' behavior and emotion feedback can be used in combination with other types of video indexing, such as traditional scene change or shot change indexing, or manual indexing using high resolution still images. Video indexing using high resolution still images is described, for example, in U.S. patent application, entitled “Video Indexing Using High Resolution Still Images,” filed on the same day herewith, which is incorporated herein by reference. Different types of video indexing may be used in the same video sequence. The user can add manual indexing, for example, within two minutes of a “happy” scene, using a still image of a person falling down, causing laughter among people.
- The
emotion detection system 100 may detect and categorize different viewers' behavior and emotion feedback every time the same video sequence is watched, and gradually generate more information regarding the video sequence. Therefore, theemotion detection system 100 can iteratively generate more accurate indices based on different viewers' feedback. For example, if every time a video sequence is play back, different viewers laugh while watching a certain segment of the video sequence, theemotion detection system 100 may accurately categorize that segment as “funny”, and index the segment accordingly. - FIG. 1A illustrates an exemplary
emotion detection system 100 that can be connected with anetwork 160 and aremote storage 150. The exemplaryemotion detection system 100 includes one or more multimedia acquisition devices, such asvideo cameras 130, and acomputer 140. Thevideo cameras 130 point to and monitor one or more viewers during a multimedia sequence, such as a video sequence, play back. Eachvideo camera 130 includes animage sensor 110 for capturing facial expression and physical movement of the viewers. Eachvideo camera 130 may also include anaudio sensor 120, such as a microphone, for capturing audio signal in the viewers' voice. In addition toimage sensors 110 andaudio sensors 120, other types ofsensor 125, such as electric or temperature sensors, may be used by theemotion detection system 100 for video indexing. For example, electrodes may be placed on human head to acquire electrical responses. Theemotion detection system 100 may include different combinational use of the sensors. - The
emotion detection system 100 then uses thecomputer 140 to categorize different behavior and emotion segments of a scene, and automatically index the video sequence according to various categories of behavior and emotion segments. The indexed video sequence may be transmitted to thenetwork 160 and/or theremote storage 150 to be stored. Theremote storage 150 may be a hard disk, a CD-ROM, a DVD RW, or a server connected to thenetwork 160. - FIG. 1B illustrates how the emotion detection system monitors viewers' feedback. Referring to FIG. 1B,
viewers 180 watch a video sequence using, for example, adigital video player 170 playing on a television set. Thevideo camera 130 is connected to thedigital video player 170 and thecomputer 140. Thevideo camera 130 monitors the viewers' behavior and emotion feedback during the video sequence play back. Thecomputer 140 detects and categorizes viewers' behavior and emotion feedback, and indexes the video sequence using various categories of behavior and emotion segments as video indices. - For example, the
emotion detection system 100 combines facial expression of the viewers, such as a smile, and audio signal in the viewers' voice, such as a laughter, to categorize the video segment as, for example, “funny”. The “funny” segment may then be used to index the video sequence. As another example, if one of theviewers 180 is snoring, that section of the video sequence can be categorized and indexed as, for example, “non-interesting”. On the other hand, if one of theviewers 180 is shouting, that section of the video sequence can be categorized and indexed as, for example, “interesting”. Physical movement may be detected and used as video indices as well. For example, if someviewers 180 are leaving the room, or not looking at the display, these behavior may show lack of interest by the viewers, and may be so noted in the video indexing. Detectingviewers 180 leaving a room may use face detection algorithms. On the other hand, if more people show up in the room, that section of the video sequence may be categorized and indexed as “interesting”, especially when people start laughing or commenting on the scene. - A certain emotion may belong to multiple categories in the
emotion detection system 100. For example, a particular segment may be “happy” and “interesting”, so that video indexing using either “happy” or “interesting” emotions may contain the same segment. Thereafter, one of theviewers 180 or another user can easily browse through specific categories of behavior and emotion segments within a stored video sequence by jumping to the particular segments directly. For example, a user may watch only “happy” moments within a video sequence. - In addition, different video indices may be generated by different users accordingly to the users' preference. For example, one user may watch “funny” segments of a video sequence, whereas another user may watch “astonished” segments of the same video sequence.
- FIG. 2 illustrates exemplary hardware components of a
computer 140 that may be used to in connection with the exemplary method for video indexing based on viewers' behavior and emotion feedback. Thecomputer 140 has a connection with anetwork 160, such as the Internet or other type of computer or telephone networks, for sending recorded multimedia sequence and video indices to friends and family by, for example, email. Thecomputer 140 typically includes amemory 202, asecondary storage device 212, aprocessor 214, aninput device 216, adisplay device 210, and anoutput device 208. - The
memory 202 may include random access memory (RAM) or similar types of memory. Thesecondary storage device 212 may include a hard disk drive, floppy disk drive, CD-ROM drive, DVD RW, or other types of non-volatile data storage. Thesecondary storage device 212 may correspond with various databases or other resources. Theprocessor 214 may execute applications or other information stored in thememory 202, thesecondary storage 212, or received from the Internet orother network 160. Theinput device 216 may include any device for entering data into thecomputer 140, such as a keyboard, key pad, cursor-control device, touch-screen (possibly with a stylus), or microphone. Thedisplay device 210 may include any type of device for presenting visual image, such as, for example, a computer monitor, flat-screen display, or display panel. Theoutput device 208 may include any type of device for presenting data in hard copy format, such as a printer, and other types of output devices including speakers or any device for providing data in audio form. Thecomputer 140 can possibly include multiple input devices, output devices, and display devices. - Although the
computer 140 is depicted with various components, one skilled in the art will appreciate that this computer can contain additional or different components. In addition, although aspects of an implementation consistent with the present invention are described as being stored in memory, one skilled in the art will appreciate that these aspects can also be stored on or read from other types of computer program products or computer-readable media, such as secondary storage devices, including hard disks, floppy disks, CD-ROM, or DVD RW; a carrier wave from the Internet or other network; or other forms of RAM or ROM. The computer-readable media may include instructions for controlling thecomputer 140 to perform a particular method. - FIG. 3 illustrates an exemplary chart of behavior and emotion feedback. A viewer's feedback may be very active301 or very passive 303, very positive 307 or very negative 305. In between, various of behavior and emotion feedback are possible, for example,
surprise 311,joy 313,anticipation 321,acceptance 323,sadness 331, disgust 333,anger 341, andfear 343. For more detailed description, see “Emotion Recognition in Human-Computer interaction” by Cowie et al, January 2001 on IEEE Signal Processing Magazine. - After the indexed multimedia sequence are processed and stored, the indexed multimedia sequence may be downloaded to the same or another
computer 140 either by transmitting over wireless channels, through a network connection, or through a wired connection, such as universal serial bus (USB) or Firewire (IEEE 1394). Alternatively, thecomputer 140 may read the local storage of thevideo camera 130 by directly connecting to a reader of thecomputer 140. Thereafter, the indexed multimedia sequence may be played back on any video display device, such as acomputer 140. The video display device may display the video indices as labeled icons for easy indexing, for example, “happy # 1”, “funny #2”, or “interesting #5”. A particular video index, for example, “happy # 1”, may be played back by clicking on an associated icon as displayed on thedisplay device 210. - As a result, the user may browse through the multimedia sequence using the various categories of behavior and emotion segments as video indices. For example, when viewing the recorded multimedia sequence using the
computer 140, a user may double click on one of the icons, for example, “funny”, and start viewing all “funny” segments within the multimedia sequence. - The indexed multimedia sequence may also be saved on a server connected to the
network 160, to be retrieved by other users. Alternatively, the indexed multimedia sequence may be transmitted to other users through thenetwork 160 by, for example, e-mail. A friend or a family member who receives the multimedia sequence and the video indices may then easily browse through particular behavior and emotion segments of the multimedia sequence. - FIGS.4-6 illustrate examples for video indexing based on viewers' behavior and emotion feedback. Referring to FIG. 4A, in addition to
scene change 410 and shotchange 420 monitoring, video and audio monitoring 400 may be conducted by thevideo camera 130 to detect the viewers' casual feedback. Theemotion detection system 100 then categorizes various categories of behavior and emotion segments within avideo sequence 480, such as a “happy”segment 430 and an “astonished”segment 440. For example, when a kid is crawling on the floor, a parent may be smiling. Theemotion detection system 100 detects the emotion and automatically categorizes the segment of the video sequence as “happy” 430. Similarly, if somebody jumps down a trampoline in a swimming pool, the viewer may have an astonished expression. Theemotion detection system 100 detects and categorizes that segment as “astonished” 440. - Referring to FIG. 4B, the
computer 140 of theemotion detection system 100 indexes thevideo sequence 480 using the “happy”segment 430 and the “astonished”segment 440 asvideo indices video indices scene change 410 and shotchange 420 may also be used as video indices. In addition, the video indices of behavior and emotion feedback may be labeled, for example, as “happy # 1”, “funny #2”, or “interesting #4”, within the video sequence for easy indexing. - In one embodiment, the
emotion detection system 100 includes only audio sensors, such asmicrophones 120, so that only audio signals are detected for video indexing. Audio feedback is part of emotion feedback. For example, “laughter” is part of “happy”, “amused” or “funny” emotion. In a video and audio bitstream, three channels are typically used, such as video, left audio, and right audio channels. The method and apparatus for indexing based on viewers' behavior and emotion feedback typically include at least an extra channel for audio feedback indexing to generate a hierarchical indexing structure. Two types of audio signals are generated, one representing system signal in the video sequence, the other representing the viewer's audio feedback. Thereafter, algorithms may be used to remove the viewer's audio signal from the system signal in the video sequence. A more sophisticated system may have a microphone for every viewer. - During a multimedia sequence, such as a video sequence, play back, the
microphone 120 monitors audio signals. Any audio reaction of theviewers 180 may be indexed into the hierarchical index representation as a feedback input. Different techniques may be used to index the video sequence using audio signals. - FIGS. 5A and 5B illustrate an exemplary embodiment of audio feedback indexing. In this embodiment, the
emotion detection system 100 detects basic oral characteristics, such as exclamations, laughter or speech, and indexes the oral characteristics into the video bitstream. Similar to the example illustrated in FIGS. 4A and 4B, in addition toscene change 510 monitoring, video andaudio monitoring 500 may be conducted by thevideo camera 130 to detect viewers' audio feedback. Theemotion detection system 100 then categorizes various categories of audio feedback within avideo sequence 580, such as a “laughter”segment 530. The “laughter”segment 530 typically follows the beginning of a “funny”segment 520 after abrief delay 540. - Referring to FIG. 5B, the
computer 140 of theemotion detection system 100 indexes thevideo sequence 580 using the “funny”segment 520 as a video index 525. Similarly,scene change 510 may also be used as a video index. Thereafter, the viewer or another user may selectively view the video sequence using either one of thevideo indices 510 and 525. - FIG. 6 illustrates another exemplary embodiment of audio feedback indexing. In this embodiment, the
emotion feedback system 100 tries to recognize the speech or other oral characteristics, and add relevant comments as annotations. In addition toscene change 610 monitoring, video andaudio monitoring 600 is conducted by thevideo camera 130 to detect viewers' audio feedback. Theemotion detection system 100 then categorizes various categories of audio feedback within avideo sequence 680, such as a “speech”segment 620. - Referring to FIG. 6B, the
computer 140 of theemotion detection system 100 indexes thevideo sequence 680 using the “speech”segment 620 as avideo index 625. Similarly,scene change 610 may also be used as a video index. Thereafter, the viewer or another user may selectively view thevideo sequence 680 using either one of thevideo indices audio file 630, thecomputer 140 may insert atext annotation 640, representing the speech itself. - FIG. 7 is a flow chart illustrating the exemplary method for video indexing based on viewers' behavior and emotion feedback. The
emotion detection system 100 may use amultimedia acquisition device 130, such as a video camera, to monitor one ormore viewers 180 during a multimedia sequence, such as a video sequence, play back (block 710). Theemotion detection system 100 then detect the viewers' behavior and emotion feedback (block 720). Next, theemotion detection system 100 may use acomputer 140 to categorize the viewers' behavior and emotion feedback as various categories of behavior and emotion segments (block 730), and index the multimedia sequence using the various categories of behavior and emotion segments (block 740). Thereafter, theviewers 180 or another user may use any video display device, such as acomputer 140, to selectively view the multimedia sequence using the various categories of behavior and emotion segments as video indices (block 750). The user may click on an icon from the video display device to browse through various categories of behavior and emotion segments (block 760). In addition, the indexed multimedia sequence may be sent to other users through anetwork 160 so that friends and family members may selectively view the multimedia sequence using the various categories of behavior and emotion segments as video indices (block 770). - While the method and apparatus for video indexing based on viewers' behavior and emotion feedback have been described in connection with an exemplary embodiment, those skilled in the art will understand that many modifications in light of these teachings are possible, and this application is intended to cover any variations thereof.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/023,810 US6585521B1 (en) | 2001-12-21 | 2001-12-21 | Video indexing based on viewers' behavior and emotion feedback |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/023,810 US6585521B1 (en) | 2001-12-21 | 2001-12-21 | Video indexing based on viewers' behavior and emotion feedback |
Publications (2)
Publication Number | Publication Date |
---|---|
US20030118974A1 true US20030118974A1 (en) | 2003-06-26 |
US6585521B1 US6585521B1 (en) | 2003-07-01 |
Family
ID=21817310
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/023,810 Expired - Fee Related US6585521B1 (en) | 2001-12-21 | 2001-12-21 | Video indexing based on viewers' behavior and emotion feedback |
Country Status (1)
Country | Link |
---|---|
US (1) | US6585521B1 (en) |
Cited By (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040101178A1 (en) * | 2002-11-25 | 2004-05-27 | Eastman Kodak Company | Imaging method and system for health monitoring and personal security |
US20050074226A1 (en) * | 2003-10-07 | 2005-04-07 | Pioneer Corporation | Index data generation apparatus, index data generation method, index data generation program and recording medium on which the program is recorded |
US20050289582A1 (en) * | 2004-06-24 | 2005-12-29 | Hitachi, Ltd. | System and method for capturing and using biometrics to review a product, service, creative work or thing |
US20070033050A1 (en) * | 2005-08-05 | 2007-02-08 | Yasuharu Asano | Information processing apparatus and method, and program |
US20070088833A1 (en) * | 2005-10-17 | 2007-04-19 | Samsung Electronics Co., Ltd. | Method and apparatus for providing multimedia data using event index |
US20070184420A1 (en) * | 2006-02-08 | 2007-08-09 | Honeywell International Inc. | Augmented tutoring |
KR100763236B1 (en) | 2006-05-09 | 2007-10-04 | 삼성전자주식회사 | Apparatus and method for editing moving picture using physiological signal |
US20080101660A1 (en) * | 2006-10-27 | 2008-05-01 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US20080107361A1 (en) * | 2006-11-07 | 2008-05-08 | Sony Corporation | Imaging apparatus, display apparatus, imaging method, and display method |
US20080126179A1 (en) * | 2006-08-31 | 2008-05-29 | Skura Corporation | System and method for tracking and predicting response to a presentation |
US20080235284A1 (en) * | 2005-09-26 | 2008-09-25 | Koninklijke Philips Electronics, N.V. | Method and Apparatus For Analysing An Emotional State of a User Being Provided With Content Information |
US20080253695A1 (en) * | 2007-04-10 | 2008-10-16 | Sony Corporation | Image storage processing apparatus, image search apparatus, image storage processing method, image search method and program |
US20090040231A1 (en) * | 2007-08-06 | 2009-02-12 | Sony Corporation | Information processing apparatus, system, and method thereof |
US20090113297A1 (en) * | 2007-10-24 | 2009-04-30 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Requesting a second content based on a user's reaction to a first content |
US20090171901A1 (en) * | 2007-12-28 | 2009-07-02 | Microsoft Corporation | Real-time annotator |
US20090253996A1 (en) * | 2007-03-02 | 2009-10-08 | Lee Michael J | Integrated Sensor Headset |
US20090276802A1 (en) * | 2008-05-01 | 2009-11-05 | At&T Knowledge Ventures, L.P. | Avatars in social interactive television |
US20100005393A1 (en) * | 2007-01-22 | 2010-01-07 | Sony Corporation | Information processing apparatus, information processing method, and program |
US20100291527A1 (en) * | 2009-05-12 | 2010-11-18 | Jennifer Baldi | Kit and process for diagnosing multiple intelligences profile |
KR100995750B1 (en) | 2003-08-11 | 2010-11-19 | 엘지전자 주식회사 | Multimedia playback controlling method and apparatus for mobile terminal |
US20120063738A1 (en) * | 2009-05-18 | 2012-03-15 | Jae Min Yoon | Digital video recorder system and operating method thereof |
US20120093481A1 (en) * | 2010-10-15 | 2012-04-19 | Microsoft Corporation | Intelligent determination of replays based on event identification |
US20120143693A1 (en) * | 2010-12-02 | 2012-06-07 | Microsoft Corporation | Targeting Advertisements Based on Emotion |
US20120324492A1 (en) * | 2011-06-20 | 2012-12-20 | Microsoft Corporation | Video selection based on environmental sensing |
WO2013003611A2 (en) * | 2011-06-30 | 2013-01-03 | Divx, Llc | Systems and methods for identifying consumer electronic products using a playback device with a product identifier |
CN103207675A (en) * | 2012-04-06 | 2013-07-17 | 微软公司 | Producing collection of media programs or expanding media programs |
US20130282829A1 (en) * | 2010-12-20 | 2013-10-24 | Alcatel Lucent | Media asset management system |
US8620113B2 (en) | 2011-04-25 | 2013-12-31 | Microsoft Corporation | Laser diode modes |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
CN103530788A (en) * | 2012-07-02 | 2014-01-22 | 纬创资通股份有限公司 | Multimedia evaluating system, multimedia evaluating device and multimedia evaluating method |
US20140086554A1 (en) * | 2012-09-25 | 2014-03-27 | Raanan YEHEZKEL | Video indexing with viewer reaction estimation and visual cue detection |
EP2721831A2 (en) * | 2011-06-17 | 2014-04-23 | Microsoft Corporation | Video highlight identification based on environmental sensing |
CN103826160A (en) * | 2014-01-09 | 2014-05-28 | 广州三星通信技术研究有限公司 | Method and device for obtaining video information, and method and device for playing video |
EP2741293A1 (en) * | 2012-12-05 | 2014-06-11 | Samsung Electronics Co., Ltd | Video processing apparatus and method |
EP2742490A1 (en) * | 2011-08-08 | 2014-06-18 | Google, Inc. | Sentimental information associated with an object within media |
WO2014093919A2 (en) * | 2012-12-13 | 2014-06-19 | Microsoft Corporation | Content reaction annotations |
US8760395B2 (en) | 2011-05-31 | 2014-06-24 | Microsoft Corporation | Gesture recognition techniques |
US20140181668A1 (en) * | 2012-12-20 | 2014-06-26 | International Business Machines Corporation | Visual summarization of video for quick understanding |
CN104040467A (en) * | 2012-01-10 | 2014-09-10 | 微软公司 | Consumption of content with reactions of an individual |
US20140270701A1 (en) * | 2013-03-15 | 2014-09-18 | First Principles, Inc. | Method on indexing a recordable event from a video recording and searching a database of recordable events on a hard drive of a computer for a recordable event |
US20140317646A1 (en) * | 2013-04-18 | 2014-10-23 | Microsoft Corporation | Linked advertisements |
US8898687B2 (en) | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
US8959541B2 (en) | 2012-05-04 | 2015-02-17 | Microsoft Technology Licensing, Llc | Determining a future portion of a currently presented media program |
FR3010557A1 (en) * | 2013-09-12 | 2015-03-13 | Orange | AUTOMATIC PROCESSING OF MULTIMEDIA DATA RELEASED BY A COMMUNICATION TERMINAL |
CN104463231A (en) * | 2014-12-31 | 2015-03-25 | 合一网络技术(北京)有限公司 | Error correction method used after facial expression recognition content is labeled |
CN104462135A (en) * | 2013-09-15 | 2015-03-25 | 联想(北京)有限公司 | Data processing method and electronic device |
WO2015074434A1 (en) * | 2013-11-21 | 2015-05-28 | 中兴通讯股份有限公司 | Contact sorting method and apparatus |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US9324244B1 (en) * | 2010-05-15 | 2016-04-26 | David Sol | Distributed multi-nodal operant conditioning system and method |
US9449218B2 (en) * | 2014-10-16 | 2016-09-20 | Software Ag Usa, Inc. | Large venue surveillance and reaction systems and methods using dynamically analyzed emotional input |
CN106792170A (en) * | 2016-12-14 | 2017-05-31 | 合网络技术(北京)有限公司 | Method for processing video frequency and device |
EP3193300A1 (en) * | 2016-01-14 | 2017-07-19 | XroadMedia GmbH | Method and system for analyzing a media signal |
US9794233B2 (en) | 2011-08-31 | 2017-10-17 | Sonic Ip, Inc. | Systems and methods for application identification |
CN107615236A (en) * | 2015-06-01 | 2018-01-19 | 华为技术有限公司 | Multi-media processing method and device |
US9922350B2 (en) | 2014-07-16 | 2018-03-20 | Software Ag | Dynamically adaptable real-time customer experience manager and/or associated method |
US20180084022A1 (en) * | 2016-09-16 | 2018-03-22 | Echostar Technologies L.L.C. | Collecting media consumer data |
CN108574875A (en) * | 2011-09-12 | 2018-09-25 | 英特尔公司 | Promote the interaction based on TV with social networking tool |
US20190166584A1 (en) * | 2016-08-03 | 2019-05-30 | Lenovo Innovations Limited (Hong Kong) | Device-to-device transmission |
US10380647B2 (en) * | 2010-12-20 | 2019-08-13 | Excalibur Ip, Llc | Selection and/or modification of a portion of online content based on an emotional state of a user |
US10380687B2 (en) | 2014-08-12 | 2019-08-13 | Software Ag | Trade surveillance and monitoring systems and/or methods |
WO2019182816A1 (en) * | 2018-03-20 | 2019-09-26 | Aic Innovations Group, Inc. | Apparatus and method for user evaluation |
RU2701508C1 (en) * | 2015-12-29 | 2019-09-27 | Хуавей Текнолоджиз Ко., Лтд. | Method and system of content recommendations based on user behavior information |
FR3080511A1 (en) * | 2018-04-20 | 2019-10-25 | Ironova | MODULE AND METHOD FOR CREATING AUDIOVISUAL SEQUENCE |
US10477009B1 (en) | 2018-05-09 | 2019-11-12 | Fuvi Cognitive Network Corp. | Apparatus, method, and system of cognitive communication assistant for enhancing ability and efficiency of users communicating comprehension |
US20200285668A1 (en) * | 2019-03-06 | 2020-09-10 | International Business Machines Corporation | Emotional Experience Metadata on Recorded Images |
US10939165B2 (en) | 2011-09-12 | 2021-03-02 | Intel Corporation | Facilitating television based interaction with social networking tools |
US11032209B2 (en) * | 2015-12-09 | 2021-06-08 | Industrial Technology Research Institute | Multimedia content cross screen synchronization apparatus and method, and display device and server |
US11206235B1 (en) * | 2018-04-26 | 2021-12-21 | Facebook, Inc. | Systems and methods for surfacing content |
US20210400342A1 (en) * | 2019-09-27 | 2021-12-23 | Apple Inc. | Content Generation Based on Audience Engagement |
US11216653B2 (en) * | 2019-11-15 | 2022-01-04 | Avio Technology, Inc. | Automated collection and correlation of reviewer response to time-based media |
US11328159B2 (en) * | 2016-11-28 | 2022-05-10 | Microsoft Technology Licensing, Llc | Automatically detecting contents expressing emotions from a video and enriching an image index |
US11533518B2 (en) * | 2020-09-25 | 2022-12-20 | International Business Machines Corporation | Audio customization in streaming environment |
Families Citing this family (148)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002042115A (en) * | 2000-07-26 | 2002-02-08 | Fujitsu Ltd | Fatigue softening system, fatigue softening method, and recording medium |
GB0129024D0 (en) * | 2001-12-04 | 2002-01-23 | Pure Resonance Ltd | Method of invoking an emotional response |
US7120873B2 (en) * | 2002-01-28 | 2006-10-10 | Sharp Laboratories Of America, Inc. | Summarization of sumo video content |
JP3772205B2 (en) * | 2002-02-06 | 2006-05-10 | 国立大学法人佐賀大学 | Teaching material learning system |
US20030212993A1 (en) * | 2002-05-07 | 2003-11-13 | Pere Obrador | System and method for efficiently managing video files |
JP3867627B2 (en) * | 2002-06-26 | 2007-01-10 | ソニー株式会社 | Audience situation estimation device, audience situation estimation method, and audience situation estimation program |
JP4285012B2 (en) * | 2003-01-31 | 2009-06-24 | 株式会社日立製作所 | Learning situation judgment program and user situation judgment system |
KR100457813B1 (en) * | 2003-02-07 | 2004-11-18 | 삼성전자주식회사 | Communication system and method |
EP1629359A4 (en) * | 2003-04-07 | 2008-01-09 | Sevenecho Llc | Method, system and software for digital media narrative personalization |
US7577636B2 (en) * | 2003-05-28 | 2009-08-18 | Fernandez Dennis S | Network-extensible reconfigurable media appliance |
US20050071865A1 (en) * | 2003-09-30 | 2005-03-31 | Martins Fernando C. M. | Annotating meta-data with user responses to digital content |
US20050079474A1 (en) * | 2003-10-14 | 2005-04-14 | Kenneth Lowe | Emotional state modification method and system |
EP1687753A1 (en) * | 2003-11-21 | 2006-08-09 | Koninklijke Philips Electronics N.V. | System and method for extracting a face from a camera picture for representation in an electronic system |
US20050131697A1 (en) * | 2003-12-10 | 2005-06-16 | International Business Machines Corporation | Speech improving apparatus, system and method |
US20050131744A1 (en) * | 2003-12-10 | 2005-06-16 | International Business Machines Corporation | Apparatus, system and method of automatically identifying participants at a videoconference who exhibit a particular expression |
WO2005069251A2 (en) * | 2004-01-06 | 2005-07-28 | Mayo Foundation For Medical Education And Research | Hypoxia awareness training system |
US7672864B2 (en) * | 2004-01-09 | 2010-03-02 | Ricoh Company Ltd. | Generating and displaying level-of-interest values |
JP4458888B2 (en) * | 2004-03-22 | 2010-04-28 | 富士通株式会社 | Conference support system, minutes generation method, and computer program |
JP2007534235A (en) * | 2004-04-15 | 2007-11-22 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Method for generating a content item having a specific emotional impact on a user |
US20050240956A1 (en) * | 2004-04-22 | 2005-10-27 | Kurt Smith | Method and apparatus for enhancing wellness |
US20050250082A1 (en) * | 2004-05-05 | 2005-11-10 | Mark Baldwin | Interpersonal cognition method and system |
US20060161587A1 (en) * | 2005-01-19 | 2006-07-20 | Tiny Engine, Inc. | Psycho-analytical system and method for audio and visual indexing, searching and retrieval |
US20060161543A1 (en) * | 2005-01-19 | 2006-07-20 | Tiny Engine, Inc. | Systems and methods for providing search results based on linguistic analysis |
US20060161553A1 (en) * | 2005-01-19 | 2006-07-20 | Tiny Engine, Inc. | Systems and methods for providing user interaction based profiles |
US7389917B2 (en) * | 2005-02-15 | 2008-06-24 | International Business Machines Corporation | Enhancing web experiences using behavioral biometric data |
MX2007011471A (en) | 2005-03-17 | 2007-12-05 | Nielsen Media Res Inc | Methods and apparatus for using audience member behavior information to determine compliance with audience measurement system usage requirements. |
US20060257834A1 (en) * | 2005-05-10 | 2006-11-16 | Lee Linda M | Quantitative EEG as an identifier of learning modality |
EP1921986A4 (en) | 2005-09-02 | 2011-11-30 | Emsense Corp | A device and method for sensing electrical activity in tissue |
CA2622365A1 (en) * | 2005-09-16 | 2007-09-13 | Imotions-Emotion Technology A/S | System and method for determining human emotion by analyzing eye properties |
KR100703705B1 (en) * | 2005-11-18 | 2007-04-06 | 삼성전자주식회사 | Multimedia comment process apparatus and method for movie |
US20070118455A1 (en) * | 2005-11-18 | 2007-05-24 | Albert William J | System and method for directed request for quote |
AU2006320692A1 (en) | 2005-11-29 | 2007-06-07 | Google Inc. | Detecting repeating content in broadcast media |
US20070150281A1 (en) * | 2005-12-22 | 2007-06-28 | Hoff Todd M | Method and system for utilizing emotion to search content |
KR100738548B1 (en) * | 2006-01-11 | 2007-07-11 | 삼성전자주식회사 | Apparatus and method for visual communication by using voip |
US7831531B1 (en) | 2006-06-22 | 2010-11-09 | Google Inc. | Approximate hashing functions for finding similar content |
US8411977B1 (en) | 2006-08-29 | 2013-04-02 | Google Inc. | Audio identification using wavelet-based signatures |
US20080201000A1 (en) * | 2007-02-20 | 2008-08-21 | Nokia Corporation | Contextual grouping of media items |
US20090088610A1 (en) * | 2007-03-02 | 2009-04-02 | Lee Hans C | Measuring Physiological Response to Media for Viewership Modeling |
US9215996B2 (en) * | 2007-03-02 | 2015-12-22 | The Nielsen Company (Us), Llc | Apparatus and method for objectively determining human response to media |
US8230457B2 (en) | 2007-03-07 | 2012-07-24 | The Nielsen Company (Us), Llc. | Method and system for using coherence of biological responses as a measure of performance of a media |
US20090070798A1 (en) * | 2007-03-02 | 2009-03-12 | Lee Hans C | System and Method for Detecting Viewer Attention to Media Delivery Devices |
US20080221969A1 (en) * | 2007-03-07 | 2008-09-11 | Emsense Corporation | Method And System For Measuring And Ranking A "Thought" Response To Audiovisual Or Interactive Media, Products Or Activities Using Physiological Signals |
US8473044B2 (en) * | 2007-03-07 | 2013-06-25 | The Nielsen Company (Us), Llc | Method and system for measuring and ranking a positive or negative response to audiovisual or interactive media, products or activities using physiological signals |
US8764652B2 (en) * | 2007-03-08 | 2014-07-01 | The Nielson Company (US), LLC. | Method and system for measuring and ranking an “engagement” response to audiovisual or interactive media, products, or activities using physiological signals |
US8782681B2 (en) * | 2007-03-08 | 2014-07-15 | The Nielsen Company (Us), Llc | Method and system for rating media and events in media based on physiological data |
JP4539712B2 (en) * | 2007-12-03 | 2010-09-08 | ソニー株式会社 | Information processing terminal, information processing method, and program |
CN101711388B (en) | 2007-03-29 | 2016-04-27 | 神经焦点公司 | The effect analysis of marketing and amusement |
WO2008137581A1 (en) | 2007-05-01 | 2008-11-13 | Neurofocus, Inc. | Neuro-feedback based stimulus compression device |
EP2142082A4 (en) | 2007-05-01 | 2015-10-28 | Neurofocus Inc | Neuro-informatics repository system |
US8126220B2 (en) * | 2007-05-03 | 2012-02-28 | Hewlett-Packard Development Company L.P. | Annotating stimulus based on determined emotional response |
US20080320037A1 (en) * | 2007-05-04 | 2008-12-25 | Macguire Sean Michael | System, method and apparatus for tagging and processing multimedia content with the physical/emotional states of authors and users |
US8316302B2 (en) * | 2007-05-11 | 2012-11-20 | General Instrument Corporation | Method and apparatus for annotating video content with metadata generated using speech recognition technology |
US8392253B2 (en) | 2007-05-16 | 2013-03-05 | The Nielsen Company (Us), Llc | Neuro-physiology and neuro-behavioral based stimulus targeting system |
US8494905B2 (en) | 2007-06-06 | 2013-07-23 | The Nielsen Company (Us), Llc | Audience response analysis using simultaneous electroencephalography (EEG) and functional magnetic resonance imaging (fMRI) |
US20090030287A1 (en) * | 2007-06-06 | 2009-01-29 | Neurofocus Inc. | Incented response assessment at a point of transaction |
US8620878B2 (en) * | 2007-07-19 | 2013-12-31 | Ustream, Inc. | System and method of distributing multimedia content |
WO2009014763A2 (en) * | 2007-07-26 | 2009-01-29 | Emsense Corporation | A method and system for creating a dynamic and automated testing of user response |
US8533042B2 (en) | 2007-07-30 | 2013-09-10 | The Nielsen Company (Us), Llc | Neuro-response stimulus and stimulus attribute resonance estimator |
KR20100047865A (en) | 2007-08-28 | 2010-05-10 | 뉴로포커스, 인크. | Consumer experience assessment system |
US8635105B2 (en) | 2007-08-28 | 2014-01-21 | The Nielsen Company (Us), Llc | Consumer experience portrayal effectiveness assessment system |
US8386313B2 (en) | 2007-08-28 | 2013-02-26 | The Nielsen Company (Us), Llc | Stimulus placement system using subject neuro-response measurements |
US8392255B2 (en) | 2007-08-29 | 2013-03-05 | The Nielsen Company (Us), Llc | Content based selection and meta tagging of advertisement breaks |
US8376952B2 (en) * | 2007-09-07 | 2013-02-19 | The Nielsen Company (Us), Llc. | Method and apparatus for sensing blood oxygen |
US8494610B2 (en) | 2007-09-20 | 2013-07-23 | The Nielsen Company (Us), Llc | Analysis of marketing and entertainment effectiveness using magnetoencephalography |
US20090083129A1 (en) * | 2007-09-20 | 2009-03-26 | Neurofocus, Inc. | Personalized content delivery using neuro-response priming data |
US9191450B2 (en) * | 2007-09-20 | 2015-11-17 | Disney Enterprises, Inc. | Measuring user engagement during presentation of media content |
US8327395B2 (en) | 2007-10-02 | 2012-12-04 | The Nielsen Company (Us), Llc | System providing actionable insights based on physiological responses from viewers of media |
HUE027777T2 (en) * | 2007-10-23 | 2016-10-28 | Mindmetic Ltd | Method, system and computer program for automated interpretation of measurements in response to stimuli |
US9582805B2 (en) * | 2007-10-24 | 2017-02-28 | Invention Science Fund I, Llc | Returning a personalized advertisement |
US9513699B2 (en) * | 2007-10-24 | 2016-12-06 | Invention Science Fund I, LL | Method of selecting a second content based on a user's reaction to a first content |
US20090112693A1 (en) * | 2007-10-24 | 2009-04-30 | Jung Edward K Y | Providing personalized advertising |
US20090112696A1 (en) * | 2007-10-24 | 2009-04-30 | Jung Edward K Y | Method of space-available advertising in a mobile device |
US20090112694A1 (en) * | 2007-10-24 | 2009-04-30 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Targeted-advertising based on a sensed physiological response by a person to a general advertisement |
JP2011505175A (en) | 2007-10-31 | 2011-02-24 | エムセンス コーポレイション | System and method for providing distributed collection and centralized processing of physiological responses from viewers |
JP5243000B2 (en) * | 2007-11-19 | 2013-07-24 | ソニー株式会社 | Information providing apparatus, information acquisition terminal, receiving terminal, information providing system, information providing method, and program |
WO2009073634A1 (en) * | 2007-11-30 | 2009-06-11 | Emsense Corporation | Correlating media instance information with physiological responses from participating subjects |
JP4538756B2 (en) * | 2007-12-03 | 2010-09-08 | ソニー株式会社 | Information processing apparatus, information processing terminal, information processing method, and program |
US8347326B2 (en) | 2007-12-18 | 2013-01-01 | The Nielsen Company (US) | Identifying key media events and modeling causal relationships between key events and reported feelings |
US8659657B2 (en) * | 2008-01-25 | 2014-02-25 | International Business Machines Corporation | System and method for pattern based thresholding applied to video surveillance monitoring |
WO2009111652A1 (en) | 2008-03-05 | 2009-09-11 | New York University | Computer-accessible medium, system and method for assessing effect of a stimulus using intersubject correlation |
US8745502B2 (en) * | 2008-05-28 | 2014-06-03 | Snibbe Interactive, Inc. | System and method for interfacing interactive systems with social networks and media playback devices |
US20100010370A1 (en) | 2008-07-09 | 2010-01-14 | De Lemos Jakob | System and method for calibrating and normalizing eye data in emotional testing |
WO2010018459A2 (en) | 2008-08-15 | 2010-02-18 | Imotions - Emotion Technology A/S | System and method for identifying the existence and position of text in visual media content and for determining a subject's interactions with the text |
US8925001B2 (en) | 2008-09-12 | 2014-12-30 | At&T Intellectual Property I, L.P. | Media stream generation based on a category of user expression |
US20100070858A1 (en) * | 2008-09-12 | 2010-03-18 | At&T Intellectual Property I, L.P. | Interactive Media System and Method Using Context-Based Avatar Configuration |
US20100095318A1 (en) * | 2008-10-14 | 2010-04-15 | William Wagner | System and Method for Monitoring Audience Response |
US20100105473A1 (en) * | 2008-10-27 | 2010-04-29 | Christopher Murphy | Video role play |
US8270814B2 (en) | 2009-01-21 | 2012-09-18 | The Nielsen Company (Us), Llc | Methods and apparatus for providing video with embedded media |
US9357240B2 (en) | 2009-01-21 | 2016-05-31 | The Nielsen Company (Us), Llc | Methods and apparatus for providing alternate media for video decoders |
US8464288B2 (en) | 2009-01-21 | 2013-06-11 | The Nielsen Company (Us), Llc | Methods and apparatus for providing personalized media in video |
US8663108B2 (en) * | 2009-01-28 | 2014-03-04 | The United States Of America As Represented By The Secretary Of The Army | Method and system for rapidly and passively identifying changes in number of open pores in the skin of an individual |
WO2010100567A2 (en) | 2009-03-06 | 2010-09-10 | Imotions- Emotion Technology A/S | System and method for determining emotional response to olfactory stimuli |
US20100250325A1 (en) | 2009-03-24 | 2010-09-30 | Neurofocus, Inc. | Neurological profiles for market matching and stimulus presentation |
US20100249636A1 (en) * | 2009-03-27 | 2010-09-30 | Neurofocus, Inc. | Personalized stimulus placement in video games |
US8655437B2 (en) | 2009-08-21 | 2014-02-18 | The Nielsen Company (Us), Llc | Analysis of the mirror neuron system for evaluation of stimulus |
US10987015B2 (en) | 2009-08-24 | 2021-04-27 | Nielsen Consumer Llc | Dry electrodes for electroencephalography |
US8756657B2 (en) * | 2009-09-29 | 2014-06-17 | Ebay Inc. | Mobile or user device authentication and tracking |
US20110106750A1 (en) | 2009-10-29 | 2011-05-05 | Neurofocus, Inc. | Generating ratings predictions using neuro-response data |
US8209224B2 (en) | 2009-10-29 | 2012-06-26 | The Nielsen Company (Us), Llc | Intracluster content management using neuro-response priming data |
US9560984B2 (en) | 2009-10-29 | 2017-02-07 | The Nielsen Company (Us), Llc | Analysis of controlled and automatic attention for introduction of stimulus material |
US8335716B2 (en) | 2009-11-19 | 2012-12-18 | The Nielsen Company (Us), Llc. | Multimedia advertisement exchange |
US8335715B2 (en) | 2009-11-19 | 2012-12-18 | The Nielsen Company (Us), Llc. | Advertisement exchange using neuro-response data |
US8625033B1 (en) | 2010-02-01 | 2014-01-07 | Google Inc. | Large-scale matching of audio and video |
US8888497B2 (en) * | 2010-03-12 | 2014-11-18 | Yahoo! Inc. | Emotional web |
US20110225049A1 (en) * | 2010-03-12 | 2011-09-15 | Yahoo! Inc. | Emoticlips |
US8442849B2 (en) * | 2010-03-12 | 2013-05-14 | Yahoo! Inc. | Emotional mapping |
US8684742B2 (en) | 2010-04-19 | 2014-04-01 | Innerscope Research, Inc. | Short imagery task (SIT) research method |
US8655428B2 (en) | 2010-05-12 | 2014-02-18 | The Nielsen Company (Us), Llc | Neuro-response data synchronization |
US10398366B2 (en) * | 2010-07-01 | 2019-09-03 | Nokia Technologies Oy | Responding to changes in emotional condition of a user |
US8683348B1 (en) * | 2010-07-14 | 2014-03-25 | Intuit Inc. | Modifying software based on a user's emotional state |
US8392250B2 (en) | 2010-08-09 | 2013-03-05 | The Nielsen Company (Us), Llc | Neuro-response evaluated stimulus in virtual reality environments |
US8392251B2 (en) | 2010-08-09 | 2013-03-05 | The Nielsen Company (Us), Llc | Location aware presentation of stimulus material |
US8396744B2 (en) | 2010-08-25 | 2013-03-12 | The Nielsen Company (Us), Llc | Effective virtual reality environments for presentation of marketing materials |
US8438590B2 (en) * | 2010-09-22 | 2013-05-07 | General Instrument Corporation | System and method for measuring audience reaction to media content |
CA2815273A1 (en) | 2010-10-21 | 2012-04-26 | Holybrain Bvba | Method and apparatus for neuropsychological modeling of human experience and purchasing behavior |
US8640021B2 (en) | 2010-11-12 | 2014-01-28 | Microsoft Corporation | Audience-based presentation and customization of content |
WO2012089906A1 (en) * | 2010-12-30 | 2012-07-05 | Nokia Corporation | Method, apparatus and computer program product for emotion detection |
US9141982B2 (en) | 2011-04-27 | 2015-09-22 | Right Brain Interface Nv | Method and apparatus for collaborative upload of content |
US8433815B2 (en) | 2011-09-28 | 2013-04-30 | Right Brain Interface Nv | Method and apparatus for collaborative upload of content |
US11064257B2 (en) * | 2011-11-07 | 2021-07-13 | Monet Networks, Inc. | System and method for segment relevance detection for digital content |
US10638197B2 (en) * | 2011-11-07 | 2020-04-28 | Monet Networks, Inc. | System and method for segment relevance detection for digital content using multimodal correlations |
US9569986B2 (en) | 2012-02-27 | 2017-02-14 | The Nielsen Company (Us), Llc | System and method for gathering and analyzing biometric user feedback for use in social media and advertising applications |
US9292858B2 (en) | 2012-02-27 | 2016-03-22 | The Nielsen Company (Us), Llc | Data collection system for aggregating biologically based measures in asynchronous geographically distributed public environments |
US9451303B2 (en) | 2012-02-27 | 2016-09-20 | The Nielsen Company (Us), Llc | Method and system for gathering and computing an audience's neurologically-based reactions in a distributed framework involving remote storage and computing |
US8922481B1 (en) | 2012-03-16 | 2014-12-30 | Google Inc. | Content annotation |
US9251406B2 (en) * | 2012-06-20 | 2016-02-02 | Yahoo! Inc. | Method and system for detecting users' emotions when experiencing a media program |
US8989835B2 (en) | 2012-08-17 | 2015-03-24 | The Nielsen Company (Us), Llc | Systems and methods to gather and analyze electroencephalographic data |
US9232247B2 (en) * | 2012-09-26 | 2016-01-05 | Sony Corporation | System and method for correlating audio and/or images presented to a user with facial characteristics and expressions of the user |
US9477993B2 (en) * | 2012-10-14 | 2016-10-25 | Ari M Frank | Training a predictor of emotional response based on explicit voting on content and eye tracking to verify attention |
US9104467B2 (en) | 2012-10-14 | 2015-08-11 | Ari M Frank | Utilizing eye tracking to reduce power consumption involved in measuring affective response |
US9320450B2 (en) | 2013-03-14 | 2016-04-26 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
KR101535432B1 (en) * | 2013-09-13 | 2015-07-13 | 엔에이치엔엔터테인먼트 주식회사 | Contents valuation system and contents valuating method using the system |
US9426538B2 (en) | 2013-11-20 | 2016-08-23 | At&T Intellectual Property I, Lp | Method and apparatus for presenting advertising in content having an emotional context |
US10798459B2 (en) | 2014-03-18 | 2020-10-06 | Vixs Systems, Inc. | Audio/video system with social media generation and methods for use therewith |
US9622702B2 (en) | 2014-04-03 | 2017-04-18 | The Nielsen Company (Us), Llc | Methods and apparatus to gather and analyze electroencephalographic data |
CN104837059B (en) * | 2014-04-15 | 2019-01-08 | 腾讯科技(北京)有限公司 | Method for processing video frequency, device and system |
US9269374B1 (en) * | 2014-10-27 | 2016-02-23 | Mattersight Corporation | Predictive video analytics system and methods |
US9668688B2 (en) | 2015-04-17 | 2017-06-06 | Mossbridge Institute, Llc | Methods and systems for content response analysis |
US10303768B2 (en) * | 2015-05-04 | 2019-05-28 | Sri International | Exploiting multi-modal affect and semantics to assess the persuasiveness of a video |
US9936250B2 (en) | 2015-05-19 | 2018-04-03 | The Nielsen Company (Us), Llc | Methods and apparatus to adjust content presented to an individual |
WO2017036953A1 (en) * | 2015-09-02 | 2017-03-09 | Thomson Licensing | Method, apparatus and system for facilitating navigation in an extended scene |
US10242252B2 (en) * | 2015-09-25 | 2019-03-26 | Intel Corporation | Expression recognition tag |
WO2017136931A1 (en) * | 2016-02-08 | 2017-08-17 | Nuralogix Corporation | System and method for conducting online market research |
CN108833989A (en) * | 2018-06-20 | 2018-11-16 | 聚好看科技股份有限公司 | A kind of method and apparatus that barrage is generated and shown |
CN108833946A (en) * | 2018-06-22 | 2018-11-16 | 青岛海信传媒网络技术有限公司 | Video broadcasting method and device |
US11620334B2 (en) | 2019-11-18 | 2023-04-04 | International Business Machines Corporation | Commercial video summaries using crowd annotation |
US11496802B2 (en) | 2019-11-29 | 2022-11-08 | International Business Machines Corporation | Media stream delivery |
US11445245B2 (en) * | 2020-06-03 | 2022-09-13 | Caavo Inc | Synchronized combinations of captured real-time media content with played-back digital content |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5243517A (en) | 1988-08-03 | 1993-09-07 | Westinghouse Electric Corp. | Method and apparatus for physiological evaluation of short films and entertainment materials |
US5940801A (en) | 1994-04-26 | 1999-08-17 | Health Hero Network, Inc. | Modular microprocessor-based diagnostic measurement apparatus and method for psychological conditions |
US5774591A (en) * | 1995-12-15 | 1998-06-30 | Xerox Corporation | Apparatus and method for recognizing facial expressions and facial gestures in a sequence of images |
US5676138A (en) | 1996-03-15 | 1997-10-14 | Zawilinski; Kenneth Michael | Emotional response analyzer system with multimedia display |
US5944530A (en) * | 1996-08-13 | 1999-08-31 | Ho; Chi Fai | Learning method and system that consider a student's concentration level |
US5828809A (en) * | 1996-10-01 | 1998-10-27 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for extracting indexing information from digital video data |
US6099319A (en) | 1998-02-24 | 2000-08-08 | Zaltman; Gerald | Neuroimaging as a marketing tool |
US6102846A (en) | 1998-02-26 | 2000-08-15 | Eastman Kodak Company | System and method of managing a psychological state of an individual using images |
JP3538112B2 (en) * | 2000-03-21 | 2004-06-14 | 日本電信電話株式会社 | Remote monitoring method, apparatus, and recording medium recording remote monitoring program |
-
2001
- 2001-12-21 US US10/023,810 patent/US6585521B1/en not_active Expired - Fee Related
Cited By (139)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040101178A1 (en) * | 2002-11-25 | 2004-05-27 | Eastman Kodak Company | Imaging method and system for health monitoring and personal security |
US7319780B2 (en) * | 2002-11-25 | 2008-01-15 | Eastman Kodak Company | Imaging method and system for health monitoring and personal security |
KR100995750B1 (en) | 2003-08-11 | 2010-11-19 | 엘지전자 주식회사 | Multimedia playback controlling method and apparatus for mobile terminal |
US20050074226A1 (en) * | 2003-10-07 | 2005-04-07 | Pioneer Corporation | Index data generation apparatus, index data generation method, index data generation program and recording medium on which the program is recorded |
EP1523010A1 (en) * | 2003-10-07 | 2005-04-13 | Pioneer Corporation | Index data generation apparatus, index data generation method, index data generation program and recording medium on which the program is recorded |
US7398003B2 (en) | 2003-10-07 | 2008-07-08 | Pioneer Corporation | Index data generation apparatus, index data generation method, index data generation program and recording medium on which the program is recorded |
US20050289582A1 (en) * | 2004-06-24 | 2005-12-29 | Hitachi, Ltd. | System and method for capturing and using biometrics to review a product, service, creative work or thing |
US8407055B2 (en) * | 2005-08-05 | 2013-03-26 | Sony Corporation | Information processing apparatus and method for recognizing a user's emotion |
US20070033050A1 (en) * | 2005-08-05 | 2007-02-08 | Yasuharu Asano | Information processing apparatus and method, and program |
US20080235284A1 (en) * | 2005-09-26 | 2008-09-25 | Koninklijke Philips Electronics, N.V. | Method and Apparatus For Analysing An Emotional State of a User Being Provided With Content Information |
US20070088833A1 (en) * | 2005-10-17 | 2007-04-19 | Samsung Electronics Co., Ltd. | Method and apparatus for providing multimedia data using event index |
US8718447B2 (en) | 2005-10-17 | 2014-05-06 | Samsung Electronics Co., Ltd. | Method and apparatus for providing multimedia data using event index |
US20070184420A1 (en) * | 2006-02-08 | 2007-08-09 | Honeywell International Inc. | Augmented tutoring |
US9721480B2 (en) * | 2006-02-08 | 2017-08-01 | Honeywell International Inc. | Augmented tutoring |
KR100763236B1 (en) | 2006-05-09 | 2007-10-04 | 삼성전자주식회사 | Apparatus and method for editing moving picture using physiological signal |
US20070263978A1 (en) * | 2006-05-09 | 2007-11-15 | Samsung Electronics Co., Ltd. | System, method and medium editing moving pictures using biometric signals |
US8157729B2 (en) | 2006-05-09 | 2012-04-17 | Samsung Electronics Co., Ltd. | System, method and medium editing moving pictures using biometric signals |
US20080126179A1 (en) * | 2006-08-31 | 2008-05-29 | Skura Corporation | System and method for tracking and predicting response to a presentation |
US9560411B2 (en) | 2006-10-27 | 2017-01-31 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US20110219042A1 (en) * | 2006-10-27 | 2011-09-08 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US20080101660A1 (en) * | 2006-10-27 | 2008-05-01 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US8605958B2 (en) * | 2006-10-27 | 2013-12-10 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US7953254B2 (en) * | 2006-10-27 | 2011-05-31 | Samsung Electronics Co., Ltd. | Method and apparatus for generating meta data of content |
US20080107361A1 (en) * | 2006-11-07 | 2008-05-08 | Sony Corporation | Imaging apparatus, display apparatus, imaging method, and display method |
US20100005393A1 (en) * | 2007-01-22 | 2010-01-07 | Sony Corporation | Information processing apparatus, information processing method, and program |
US20090253996A1 (en) * | 2007-03-02 | 2009-10-08 | Lee Michael J | Integrated Sensor Headset |
US8687925B2 (en) * | 2007-04-10 | 2014-04-01 | Sony Corporation | Image storage processing apparatus, image search apparatus, image storage processing method, image search method and program |
US20080253695A1 (en) * | 2007-04-10 | 2008-10-16 | Sony Corporation | Image storage processing apparatus, image search apparatus, image storage processing method, image search method and program |
US10262449B2 (en) | 2007-08-06 | 2019-04-16 | Sony Corporation | Information processing apparatus, system, and method for displaying bio-information or kinetic information |
US10937221B2 (en) | 2007-08-06 | 2021-03-02 | Sony Corporation | Information processing apparatus, system, and method for displaying bio-information or kinetic information |
US8797331B2 (en) | 2007-08-06 | 2014-08-05 | Sony Corporation | Information processing apparatus, system, and method thereof |
US20090040231A1 (en) * | 2007-08-06 | 2009-02-12 | Sony Corporation | Information processing apparatus, system, and method thereof |
US9568998B2 (en) | 2007-08-06 | 2017-02-14 | Sony Corporation | Information processing apparatus, system, and method for displaying bio-information or kinetic information |
US9972116B2 (en) | 2007-08-06 | 2018-05-15 | Sony Corporation | Information processing apparatus, system, and method for displaying bio-information or kinetic information |
US10529114B2 (en) | 2007-08-06 | 2020-01-07 | Sony Corporation | Information processing apparatus, system, and method for displaying bio-information or kinetic information |
US20090113297A1 (en) * | 2007-10-24 | 2009-04-30 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Requesting a second content based on a user's reaction to a first content |
US20090171901A1 (en) * | 2007-12-28 | 2009-07-02 | Microsoft Corporation | Real-time annotator |
US8131750B2 (en) * | 2007-12-28 | 2012-03-06 | Microsoft Corporation | Real-time annotator |
US20110225603A1 (en) * | 2008-05-01 | 2011-09-15 | At&T Intellectual Property I, L.P. | Avatars in Social Interactive Television |
US20090276802A1 (en) * | 2008-05-01 | 2009-11-05 | At&T Knowledge Ventures, L.P. | Avatars in social interactive television |
US7953255B2 (en) * | 2008-05-01 | 2011-05-31 | At&T Intellectual Property I, L.P. | Avatars in social interactive television |
US8818054B2 (en) | 2008-05-01 | 2014-08-26 | At&T Intellectual Property I, L.P. | Avatars in social interactive television |
US8098905B2 (en) * | 2008-05-01 | 2012-01-17 | At&T Intellectual Property I, L.P. | Avatars in social interactive television |
US8311295B2 (en) | 2008-05-01 | 2012-11-13 | At&T Intellectual Property I, L.P. | Avatars in social interactive television |
US20100291527A1 (en) * | 2009-05-12 | 2010-11-18 | Jennifer Baldi | Kit and process for diagnosing multiple intelligences profile |
US20120063738A1 (en) * | 2009-05-18 | 2012-03-15 | Jae Min Yoon | Digital video recorder system and operating method thereof |
US9324244B1 (en) * | 2010-05-15 | 2016-04-26 | David Sol | Distributed multi-nodal operant conditioning system and method |
US20120093481A1 (en) * | 2010-10-15 | 2012-04-19 | Microsoft Corporation | Intelligent determination of replays based on event identification |
US9484065B2 (en) * | 2010-10-15 | 2016-11-01 | Microsoft Technology Licensing, Llc | Intelligent determination of replays based on event identification |
US20120143693A1 (en) * | 2010-12-02 | 2012-06-07 | Microsoft Corporation | Targeting Advertisements Based on Emotion |
US20130282829A1 (en) * | 2010-12-20 | 2013-10-24 | Alcatel Lucent | Media asset management system |
US10380647B2 (en) * | 2010-12-20 | 2019-08-13 | Excalibur Ip, Llc | Selection and/or modification of a portion of online content based on an emotional state of a user |
US9674250B2 (en) * | 2010-12-20 | 2017-06-06 | Alcatel Lucent | Media asset management system |
US8620113B2 (en) | 2011-04-25 | 2013-12-31 | Microsoft Corporation | Laser diode modes |
US8760395B2 (en) | 2011-05-31 | 2014-06-24 | Microsoft Corporation | Gesture recognition techniques |
US9372544B2 (en) | 2011-05-31 | 2016-06-21 | Microsoft Technology Licensing, Llc | Gesture recognition techniques |
US10331222B2 (en) | 2011-05-31 | 2019-06-25 | Microsoft Technology Licensing, Llc | Gesture recognition techniques |
EP2721831A2 (en) * | 2011-06-17 | 2014-04-23 | Microsoft Corporation | Video highlight identification based on environmental sensing |
EP2721831A4 (en) * | 2011-06-17 | 2015-04-15 | Microsoft Technology Licensing Llc | Video highlight identification based on environmental sensing |
US20120324492A1 (en) * | 2011-06-20 | 2012-12-20 | Microsoft Corporation | Video selection based on environmental sensing |
WO2013003611A2 (en) * | 2011-06-30 | 2013-01-03 | Divx, Llc | Systems and methods for identifying consumer electronic products using a playback device with a product identifier |
WO2013003611A3 (en) * | 2011-06-30 | 2014-05-08 | Sonic Ip, Inc. | Systems and methods for identifying consumer electronic products using a playback device with a product identifier |
US9092646B2 (en) | 2011-06-30 | 2015-07-28 | Sonic Ip, Inc. | Systems and methods for identifying consumer electronic products based on a product identifier |
EP2742490A4 (en) * | 2011-08-08 | 2015-04-08 | Google Inc | Sentimental information associated with an object within media |
US11947587B2 (en) | 2011-08-08 | 2024-04-02 | Google Llc | Methods, systems, and media for generating sentimental information associated with media content |
US11080320B2 (en) | 2011-08-08 | 2021-08-03 | Google Llc | Methods, systems, and media for generating sentimental information associated with media content |
EP2742490A1 (en) * | 2011-08-08 | 2014-06-18 | Google, Inc. | Sentimental information associated with an object within media |
US10341306B2 (en) | 2011-08-31 | 2019-07-02 | Divx, Llc | Systems and methods for application identification |
US9794233B2 (en) | 2011-08-31 | 2017-10-17 | Sonic Ip, Inc. | Systems and methods for application identification |
US11870758B2 (en) | 2011-08-31 | 2024-01-09 | Divx, Llc | Systems and methods for application identification |
CN108574875A (en) * | 2011-09-12 | 2018-09-25 | 英特尔公司 | Promote the interaction based on TV with social networking tool |
US10939165B2 (en) | 2011-09-12 | 2021-03-02 | Intel Corporation | Facilitating television based interaction with social networking tools |
US9154837B2 (en) | 2011-12-02 | 2015-10-06 | Microsoft Technology Licensing, Llc | User interface presenting an animated avatar performing a media reaction |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US10798438B2 (en) | 2011-12-09 | 2020-10-06 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US9628844B2 (en) | 2011-12-09 | 2017-04-18 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US10045077B2 (en) | 2012-01-10 | 2018-08-07 | Microsoft Technology Licensing, Llc | Consumption of content with reactions of an individual |
CN104040467A (en) * | 2012-01-10 | 2014-09-10 | 微软公司 | Consumption of content with reactions of an individual |
US8898687B2 (en) | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
CN103207675A (en) * | 2012-04-06 | 2013-07-17 | 微软公司 | Producing collection of media programs or expanding media programs |
US9788032B2 (en) | 2012-05-04 | 2017-10-10 | Microsoft Technology Licensing, Llc | Determining a future portion of a currently presented media program |
US8959541B2 (en) | 2012-05-04 | 2015-02-17 | Microsoft Technology Licensing, Llc | Determining a future portion of a currently presented media program |
CN103530788A (en) * | 2012-07-02 | 2014-01-22 | 纬创资通股份有限公司 | Multimedia evaluating system, multimedia evaluating device and multimedia evaluating method |
CN104541514A (en) * | 2012-09-25 | 2015-04-22 | 英特尔公司 | Video indexing with viewer reaction estimation and visual cue detection |
EP2901707A4 (en) * | 2012-09-25 | 2016-06-22 | Intel Corp | Video indexing with viewer reaction estimation and visual cue detection |
US9247225B2 (en) * | 2012-09-25 | 2016-01-26 | Intel Corporation | Video indexing with viewer reaction estimation and visual cue detection |
WO2014051992A1 (en) | 2012-09-25 | 2014-04-03 | Intel Corporation | Video indexing with viewer reaction estimation and visual cue detection |
US20140086554A1 (en) * | 2012-09-25 | 2014-03-27 | Raanan YEHEZKEL | Video indexing with viewer reaction estimation and visual cue detection |
EP2741293A1 (en) * | 2012-12-05 | 2014-06-11 | Samsung Electronics Co., Ltd | Video processing apparatus and method |
WO2014093919A2 (en) * | 2012-12-13 | 2014-06-19 | Microsoft Corporation | Content reaction annotations |
US10678852B2 (en) | 2012-12-13 | 2020-06-09 | Microsoft Technology Licensing, Llc | Content reaction annotations |
EP2932457A4 (en) * | 2012-12-13 | 2016-08-10 | Microsoft Technology Licensing Llc | Content reaction annotations |
WO2014093919A3 (en) * | 2012-12-13 | 2014-10-09 | Microsoft Corporation | Content reaction annotations |
US9721010B2 (en) | 2012-12-13 | 2017-08-01 | Microsoft Technology Licensing, Llc | Content reaction annotations |
US20140181668A1 (en) * | 2012-12-20 | 2014-06-26 | International Business Machines Corporation | Visual summarization of video for quick understanding |
US9961403B2 (en) | 2012-12-20 | 2018-05-01 | Lenovo Enterprise Solutions (Singapore) PTE., LTD. | Visual summarization of video for quick understanding by determining emotion objects for semantic segments of video |
US20140270701A1 (en) * | 2013-03-15 | 2014-09-18 | First Principles, Inc. | Method on indexing a recordable event from a video recording and searching a database of recordable events on a hard drive of a computer for a recordable event |
CN105264603A (en) * | 2013-03-15 | 2016-01-20 | 第一原理公司 | Method on indexing a recordable event from a video recording and searching a database of recordable events on a hard drive of a computer for a recordable event |
US20140317646A1 (en) * | 2013-04-18 | 2014-10-23 | Microsoft Corporation | Linked advertisements |
US9015737B2 (en) * | 2013-04-18 | 2015-04-21 | Microsoft Technology Licensing, Llc | Linked advertisements |
FR3010557A1 (en) * | 2013-09-12 | 2015-03-13 | Orange | AUTOMATIC PROCESSING OF MULTIMEDIA DATA RELEASED BY A COMMUNICATION TERMINAL |
CN104462135A (en) * | 2013-09-15 | 2015-03-25 | 联想(北京)有限公司 | Data processing method and electronic device |
WO2015074434A1 (en) * | 2013-11-21 | 2015-05-28 | 中兴通讯股份有限公司 | Contact sorting method and apparatus |
CN103826160A (en) * | 2014-01-09 | 2014-05-28 | 广州三星通信技术研究有限公司 | Method and device for obtaining video information, and method and device for playing video |
US9922350B2 (en) | 2014-07-16 | 2018-03-20 | Software Ag | Dynamically adaptable real-time customer experience manager and/or associated method |
US10380687B2 (en) | 2014-08-12 | 2019-08-13 | Software Ag | Trade surveillance and monitoring systems and/or methods |
US9996736B2 (en) | 2014-10-16 | 2018-06-12 | Software Ag Usa, Inc. | Large venue surveillance and reaction systems and methods using dynamically analyzed emotional input |
US9449218B2 (en) * | 2014-10-16 | 2016-09-20 | Software Ag Usa, Inc. | Large venue surveillance and reaction systems and methods using dynamically analyzed emotional input |
CN104463231A (en) * | 2014-12-31 | 2015-03-25 | 合一网络技术(北京)有限公司 | Error correction method used after facial expression recognition content is labeled |
CN107615236A (en) * | 2015-06-01 | 2018-01-19 | 华为技术有限公司 | Multi-media processing method and device |
EP3306463A4 (en) * | 2015-06-01 | 2018-06-06 | Huawei Technologies Co. Ltd. | Method and device for processing multimedia |
US20180160174A1 (en) * | 2015-06-01 | 2018-06-07 | Huawei Technologies Co., Ltd. | Method and device for processing multimedia |
US11032209B2 (en) * | 2015-12-09 | 2021-06-08 | Industrial Technology Research Institute | Multimedia content cross screen synchronization apparatus and method, and display device and server |
US11500907B2 (en) * | 2015-12-29 | 2022-11-15 | Futurewei Technologies, Inc. | System and method for user-behavior based content recommendations |
RU2701508C1 (en) * | 2015-12-29 | 2019-09-27 | Хуавей Текнолоджиз Ко., Лтд. | Method and system of content recommendations based on user behavior information |
US10664500B2 (en) | 2015-12-29 | 2020-05-26 | Futurewei Technologies, Inc. | System and method for user-behavior based content recommendations |
EP3193300A1 (en) * | 2016-01-14 | 2017-07-19 | XroadMedia GmbH | Method and system for analyzing a media signal |
US20190166584A1 (en) * | 2016-08-03 | 2019-05-30 | Lenovo Innovations Limited (Hong Kong) | Device-to-device transmission |
US10986609B2 (en) * | 2016-08-03 | 2021-04-20 | Lenovo Innovations Limited (Hong Kong) | Device-to-device transmission |
US10390096B2 (en) * | 2016-09-16 | 2019-08-20 | DISH Technologies L.L.C. | Collecting media consumer data |
US20180084022A1 (en) * | 2016-09-16 | 2018-03-22 | Echostar Technologies L.L.C. | Collecting media consumer data |
US11328159B2 (en) * | 2016-11-28 | 2022-05-10 | Microsoft Technology Licensing, Llc | Automatically detecting contents expressing emotions from a video and enriching an image index |
CN106792170A (en) * | 2016-12-14 | 2017-05-31 | 合网络技术(北京)有限公司 | Method for processing video frequency and device |
WO2019182816A1 (en) * | 2018-03-20 | 2019-09-26 | Aic Innovations Group, Inc. | Apparatus and method for user evaluation |
US11627877B2 (en) | 2018-03-20 | 2023-04-18 | Aic Innovations Group, Inc. | Apparatus and method for user evaluation |
FR3080511A1 (en) * | 2018-04-20 | 2019-10-25 | Ironova | MODULE AND METHOD FOR CREATING AUDIOVISUAL SEQUENCE |
US11206235B1 (en) * | 2018-04-26 | 2021-12-21 | Facebook, Inc. | Systems and methods for surfacing content |
US10686928B2 (en) | 2018-05-09 | 2020-06-16 | Fuvi Cognitive Network Corp. | Apparatus, method, and system of cognitive communication assistant for enhancing ability and efficiency of users communicating comprehension |
WO2019217677A1 (en) * | 2018-05-09 | 2019-11-14 | Fuvi Cognitive Network Corp. | Apparatus, method, and system of cognitive communication assistant for enhancing ability and efficiency of users communicating comprehension |
US10477009B1 (en) | 2018-05-09 | 2019-11-12 | Fuvi Cognitive Network Corp. | Apparatus, method, and system of cognitive communication assistant for enhancing ability and efficiency of users communicating comprehension |
US11163822B2 (en) * | 2019-03-06 | 2021-11-02 | International Business Machines Corporation | Emotional experience metadata on recorded images |
US11157549B2 (en) * | 2019-03-06 | 2021-10-26 | International Business Machines Corporation | Emotional experience metadata on recorded images |
US20200285669A1 (en) * | 2019-03-06 | 2020-09-10 | International Business Machines Corporation | Emotional Experience Metadata on Recorded Images |
US20200285668A1 (en) * | 2019-03-06 | 2020-09-10 | International Business Machines Corporation | Emotional Experience Metadata on Recorded Images |
US20210400342A1 (en) * | 2019-09-27 | 2021-12-23 | Apple Inc. | Content Generation Based on Audience Engagement |
US11949949B2 (en) * | 2019-09-27 | 2024-04-02 | Apple Inc. | Content generation based on audience engagement |
US11216653B2 (en) * | 2019-11-15 | 2022-01-04 | Avio Technology, Inc. | Automated collection and correlation of reviewer response to time-based media |
US11533518B2 (en) * | 2020-09-25 | 2022-12-20 | International Business Machines Corporation | Audio customization in streaming environment |
Also Published As
Publication number | Publication date |
---|---|
US6585521B1 (en) | 2003-07-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6585521B1 (en) | Video indexing based on viewers' behavior and emotion feedback | |
Drucker et al. | Smartskip: consumer level browsing and skipping of digital video content | |
JP5091086B2 (en) | Method and graphical user interface for displaying short segments of video | |
Chang | The holy grail of content-based media analysis | |
RU2440606C2 (en) | Method and apparatus for automatic generation of summary of plurality of images | |
Dimitrova et al. | Applications of video-content analysis and retrieval | |
Boreczky et al. | An interactive comic book presentation for exploring video | |
KR101058054B1 (en) | Extract video | |
US20070223871A1 (en) | Method of Generating a Content Item Having a Specific Emotional Influence on a User | |
US7181757B1 (en) | Video summary description scheme and method and system of video summary description data generation for efficient overview and browsing | |
JP4733328B2 (en) | Video summary description structure for efficient overview and browsing, and video summary description data generation method and system | |
JP2011217209A (en) | Electronic apparatus, content recommendation method, and program | |
JP2006525537A (en) | Method and apparatus for summarizing music videos using content analysis | |
JP5868978B2 (en) | Method and apparatus for providing community-based metadata | |
JP5079817B2 (en) | Method for creating a new summary for an audiovisual document that already contains a summary and report and receiver using the method | |
US6931201B2 (en) | Video indexing using high quality sound | |
Gkonela et al. | VideoSkip: event detection in social web videos with an implicit user heuristic | |
WO2008087742A1 (en) | Moving picture reproducing system, information terminal device and information display method | |
JP5343658B2 (en) | Recording / playback apparatus and content search program | |
EP1456845B1 (en) | Video indexing using high resolution still images | |
Leftheriotis et al. | Efficient video indexing on the web: A system that crowdsources user interactions with a video player | |
US20140189769A1 (en) | Information management device, server, and control method | |
Shubin et al. | Assessment of end-user response to sports highlights extraction for personal video recorders | |
Campanella | Balancing automation and user control in a home video editing system | |
WO2007099050A1 (en) | Method for displaying data extracted from a document consisting of reports and receiver implementing said method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HEWLETT-PACKARD COMPANY, COLORADO Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OBRADOR, PERE;REEL/FRAME:013445/0566 Effective date: 20011218 |
|
AS | Assignment |
Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEWLETT-PACKARD COMPANY;REEL/FRAME:013862/0623 Effective date: 20030728 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20150701 |