US20080155410A1 - Method for aligning demonstrated user actions with existing documentation - Google Patents

Method for aligning demonstrated user actions with existing documentation Download PDF

Info

Publication number
US20080155410A1
US20080155410A1 US12/041,123 US4112308A US2008155410A1 US 20080155410 A1 US20080155410 A1 US 20080155410A1 US 4112308 A US4112308 A US 4112308A US 2008155410 A1 US2008155410 A1 US 2008155410A1
Authority
US
United States
Prior art keywords
alignment
user interface
alignment device
user
documented procedure
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/041,123
Inventor
Lawrence D. Bergman
Vittorio Castelli
Eugene Rogan Creswick
Tessa A. Lau
Daniel A. Oblinger
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Activision Publishing Inc
Original Assignee
Bergman Lawrence D
Vittorio Castelli
Eugene Rogan Creswick
Lau Tessa A
Oblinger Daniel A
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bergman Lawrence D, Vittorio Castelli, Eugene Rogan Creswick, Lau Tessa A, Oblinger Daniel A filed Critical Bergman Lawrence D
Priority to US12/041,123 priority Critical patent/US20080155410A1/en
Publication of US20080155410A1 publication Critical patent/US20080155410A1/en
Assigned to ACTIVISION PUBLISHING, INC. reassignment ACTIVISION PUBLISHING, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing

Definitions

  • the present invention generally relates to documented procedures and, more particularly, to a method and apparatus for aligning steps in a documented procedure with demonstrated user actions.
  • an apparatus for aligning document actions specified in a documented procedure with demonstrated user actions includes an alignment device for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.
  • an apparatus for guiding user actions with respect to a documented procedure displayed on a user interface The documented procedure has a plurality of steps specified therein.
  • the apparatus includes an alignment device, a highlighting device, and a scrolling device.
  • the alignment device is for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.
  • the highlighting device connected in signal communication with the alignment device, is for visually highlighting on the user interface at least a portion of a step in the documented procedure based on the alignment computed by the alignment device.
  • the scrolling device connected in signal communication with the alignment device, is for automatically scrolling through each of the plurality of steps of the documented procedure on the user interface based on the alignment computed by the alignment device.
  • a method for aligning document actions specified in a documented procedure with demonstrated user actions includes the step of computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.
  • a method for guiding user actions with respect to a documented procedure displayed on a user interface includes the steps of computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric, visually highlighting on the user interface at least a portion of a step in the documented procedure based on the computed alignment, and automatically scrolling through each of the plurality of steps of the documented procedure on the user interface based on the computed alignment.
  • FIG. 1 is a block diagram illustrating a system 100 for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention
  • FIG. 2 is a flow diagram illustrating a method for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention.
  • FIG. 3 is a flow diagram further illustrating step 230 of FIG. 2 , according to an illustrative embodiment of the present invention.
  • the present invention is directed to a method and apparatus for aligning document actions specified in a documented procedure with corresponding demonstrated user actions.
  • a user opens an online documentation file, such as an operating manual or an IBM Redbook, turns to a particular documented procedure and, as the user performs the procedure, the corresponding text and/or graphics are highlighted and advance (e.g., scrolling to the next step, next page, etc.) along with user.
  • the next step to be taken in the documentation may be highlighted.
  • the highlighting and advancing may be performed directly in the user interface.
  • the present invention may advantageously employ the following features to help guide an individual to perform steps in a documented procedure: automatically highlighting the target widget on the screen; automatically detecting and branching on conditionals; and providing a visual indicator of which step the user is currently “up to” in the procedure. It is to be appreciated that the present invention is not limited to the preceding features to aid a user in performing a documented procedure and, thus, given the teachings of the present invention provided herein, other features may also be employed while maintaining the spirit of the present invention.
  • highlighting and “highlight” refer to any visual, audible, and tactile indication or cue that may be used to identify (or make stand out) an item of interest in a documented procedure.
  • Such items may include, but are not limited to, a step or portion thereof, a word, a phrase, a symbol, a hyperlink, a graphic, and any other item that may be found in a documented procedure.
  • the present invention is not limited to the preceding ways in which to identify an item, or to the items specified and, thus, other ways and other items may be used to provide “highlighting” as used herein, while maintaining the spirit and scope of the present invention.
  • processor or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage.
  • DSP digital signal processor
  • ROM read-only memory
  • RAM random access memory
  • any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that software to perform the function.
  • the invention as defined by such claims resides in the fact that the functionalities provided by the various recited means are combined and brought together in the manner which the claims call for. Applicant thus regards any means that can provide those functionalities as equivalent to those shown herein.
  • FIG. 1 is a block diagram illustrating a system 100 for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention.
  • the system 100 is intended to cooperate with (or be included in) a computer 199 that is to be used to perform the documented procedure.
  • the computer 199 may simply be a desktop computer, a laptop computer, a personal digital assistant, an electronic tablet, and so forth that is capable of displaying a documented procedure for a user to execute.
  • the computer 199 may optionally have the capability to access a network (e.g., the Internet, a company intranet, etc.) to download a documented procedure.
  • a network e.g., the Internet, a company intranet, etc.
  • the computer 199 includes one or more processors (hereinafter “processor”) 188 , one or more memories (hereinafter “memory”) 187 , a network interface 186 , and one or more buses (hereinafter “bus”) 185 .
  • the computer 199 further includes a user interface 198 for allowing a user to interact with the documented procedure.
  • the user interface 198 may include a keyboard or keypad (hereinafter “keyboard”) 197 and a display 196 .
  • the display 196 may be a touch-screen display capable of receiving user inputs via the display.
  • the present invention is not limited to interacting with only computers having the configuration as shown in FIG. 1 and, thus, other computers, computing devices and corresponding configurations may also be employed in accordance with the present invention while maintaining the spirit of the present invention.
  • the system 100 includes an alignment device 110 .
  • the alignment device 110 may receive inputs from the user interface 198 or may include a camera (also interchangeably referred to herein as “imaging device”) 111 or other device for use in determining demonstrated user actions.
  • the alignment device 110 preferably further includes one or more processors (hereinafter “processor”) 112 , one or more memories (hereinafter “memory”) 113 , and one or more buses (hereinafter “bus”) 114 for interconnecting the elements of the alignment device 110 .
  • the alignment device 110 may include a step extractor 119 for extracting steps from a documented procedure.
  • the step extractor 119 may include a scanner and/or other device capable of optical character recognition, a parser, a memory for storing words/phrases/etc. that may be compared against parsed words/phrases/etc., and/or so forth.
  • the alignment device 110 computes an alignment between the document actions specified in a documented procedure and corresponding demonstrated user actions. Based on the computed alignment, a step or portion of a step (e.g., word(s), graphic(s), and so forth) in the documented procedure is highlighted so that the user can identify where he or she is (e.g., which step is the current step for which a user action needs to be taken) within the documented procedure. Optionally, a next step may also be highlighted, preferably in a different manner than the current step.
  • the alignment may be used to automatically scroll through the procedure as the user progresses there through.
  • connection configuration depicted in FIG. 1 is merely illustrative and, thus, given the teachings of the present invention provided herein, one of ordinary skill in the related art will contemplate these and other connection configurations and other elements that may be employed with respect to a system for aligning document actions specified in a documented procedure with corresponding demonstrated user actions while maintaining the spirit of the present invention.
  • FIG. 2 is a flow diagram illustrating a method for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention.
  • the document is marked up to identify a plurality of document actions (step 205 ).
  • a document action is an action expected or mandated to be taken with respect to a given document.
  • a list that includes the plurality of document actions is generated (step 210 ). Steps 205 and 210 are preferably performed prior to a user attempting to perform the steps specified in the documented procedure.
  • the next step in the procedure is highlighted (step 215 ).
  • the system state and demonstrated user action are recorded (step 220 ).
  • An alignment is computed between the demonstrated user action and the document actions in the list based on at least one or more similarity metrics (step 230 ). Based on the alignment, the procedure may be advanced, if necessary, and the most likely document action in the procedure (i.e., the current document action to be taken by the user) is highlighted (step 240 ).
  • step 250 It is then determined whether or not the user has completed the procedure (step 250 ). If the user has not completed the procedure, then the method returns to step 215 to highlight the next step in the procedure. Preferably, the next step is highlighted in a manner different from that employed with respect to the current step, so as to avoid user confusion. Otherwise, if the user has completed the procedure, then the method is terminated. It is to be appreciated that highlighting with respect to the current step and/or the next step may involve highlighting a whole textual and/or graphical description of a step, a portion thereof, an element of the user interface (including, but not limited to, action buttons (e.g., cut, paste, save, etc.), and so forth.
  • action buttons e.g., cut, paste, save, etc.
  • FIG. 3 is a flow diagram further illustrating step 230 of FIG. 2 , according to an illustrative embodiment of the present invention.
  • Step 230 can also be considered as computing a similarity between a user-demonstrated state/action and steps in a documented procedure (document actions, as specified in the list).
  • similarity may be based on a probabilistic combination of similarity parameters.
  • the number of words in common between a textual description of a demonstrated user action and document action is determined (step 310 ).
  • the similarity of the action types involved e.g., clicking button and clicking checkbox are more similar than typing
  • the similarity of the action types involved e.g., clicking button and clicking checkbox are more similar than typing
  • a probability distribution (e.g., including, but not limited to, a Gaussian distribution) over two or more steps is determined (step 330 ).
  • the next step in temporal order is usually the most likely.
  • Any common strings in the user interface e.g., button labels
  • the documented procedure are determined (step 340 ).
  • An alignment based on one or more of steps 310 through 340 is then calculated (step 350 ). It is to be appreciated that the alignment may also be based on experiences of previous users, using any of the results of steps 310 through even 350 for the previous users.
  • the teachings of the present invention are implemented as a combination of hardware and software.
  • the software is preferably implemented as an application program tangibly embodied on a program storage unit.
  • the application program may be uploaded to, and executed by, a machine comprising any suitable architecture.
  • the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPU”), a random access memory (“RAM”), and input/output (“I/O”) interfaces.
  • CPU central processing units
  • RAM random access memory
  • I/O input/output
  • the computer platform may also include an operating system and microinstruction code.
  • the various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU.
  • various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.

Abstract

There is provided an apparatus for aligning document actions specified in a documented procedure with demonstrated user actions. The apparatus includes an alignment device for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.

Description

    FIELD OF THE INVENTION
  • The present invention generally relates to documented procedures and, more particularly, to a method and apparatus for aligning steps in a documented procedure with demonstrated user actions.
  • BACKGROUND OF THE INVENTION
  • Technical support procedures are typically very complex. Users often have trouble following printed instruction describing how to perform these procedures.
  • In a study to investigate how end users perform technical support procedures while following printed directions, the mouse and keyboard actions of eleven subjects were captured as they followed five pages of written instructions (including screenshots). These instructions were copied from IBM's internal technical support website. The instructions described a procedure to modify and verify the Domain Name Server (DNS) configuration of a laptop computer. Each subject was presented with a different initial configuration, with the eleven configurations chosen to be distinct and to span the space of possible procedure pathways. The subjects were instructed to follow the directions to restore the system to the correct configuration.
  • Nearly all participants, regardless of MICROSOFT WINDOWS knowledge, had difficulties following printed instructions. In particular, it was noted that people often had difficulty translating from text on the page to widgets on the screen. Moreover, it was noted that people tended to miss portions of the instructions, particularly separate portions of text that described branches of a conditional instruction. Also, it was noted that people had difficulty following out-of-order instructions and screenshots, especially those that were not identical to what they saw on the screen.
  • Accordingly, it would be desirable and highly advantageous to have a method and apparatus for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, so as provide a guided approach to technical support procedures.
  • SUMMARY OF THE INVENTION
  • These and other drawbacks and disadvantages of the prior art are addressed by the present invention, which is directed to a method and apparatus for aligning document actions specified in a documented procedure with corresponding demonstrated user actions.
  • According to an aspect of the present invention, there is provided an apparatus for aligning document actions specified in a documented procedure with demonstrated user actions. The apparatus includes an alignment device for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.
  • According to another aspect of the present invention, there is provided an apparatus for guiding user actions with respect to a documented procedure displayed on a user interface. The documented procedure has a plurality of steps specified therein. The apparatus includes an alignment device, a highlighting device, and a scrolling device. The alignment device is for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric. The highlighting device, connected in signal communication with the alignment device, is for visually highlighting on the user interface at least a portion of a step in the documented procedure based on the alignment computed by the alignment device. The scrolling device, connected in signal communication with the alignment device, is for automatically scrolling through each of the plurality of steps of the documented procedure on the user interface based on the alignment computed by the alignment device.
  • According to yet another aspect of the present invention, there is provided a method for aligning document actions specified in a documented procedure with demonstrated user actions. The method includes the step of computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric.
  • According to a further aspect of the present invention, there is provided a method for guiding user actions with respect to a documented procedure displayed on a user interface. The documented procedure has a plurality of steps specified therein. The method includes the steps of computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric, visually highlighting on the user interface at least a portion of a step in the documented procedure based on the computed alignment, and automatically scrolling through each of the plurality of steps of the documented procedure on the user interface based on the computed alignment.
  • These and other aspects, features and advantages of the present invention will become apparent from the following detailed description of exemplary embodiments, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention may be better understood in accordance with the following exemplary figures, in which:
  • FIG. 1 is a block diagram illustrating a system 100 for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention;
  • FIG. 2 is a flow diagram illustrating a method for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention; and
  • FIG. 3 is a flow diagram further illustrating step 230 of FIG. 2, according to an illustrative embodiment of the present invention.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • The present invention is directed to a method and apparatus for aligning document actions specified in a documented procedure with corresponding demonstrated user actions.
  • A brief description will now be given of an implementation of the present invention to illustrate some of the many attendant advantages and features of the present invention. According to one illustrative embodiment of the present invention, a user opens an online documentation file, such as an operating manual or an IBM Redbook, turns to a particular documented procedure and, as the user performs the procedure, the corresponding text and/or graphics are highlighted and advance (e.g., scrolling to the next step, next page, etc.) along with user. Optionally, the next step to be taken in the documentation may be highlighted. In the likely event the user is performing the procedure using a user interface, the highlighting and advancing may be performed directly in the user interface.
  • Thus, in accordance with the present invention, novice users could benefit from having annotations explaining the key concepts in the procedure, and experts could benefit from partial automation that helps them avoid mistakes. Several ways have been identified in which an interactive help system could be an improvement over printed instructions. For example, in accordance with the principles of the present invention, the present invention may advantageously employ the following features to help guide an individual to perform steps in a documented procedure: automatically highlighting the target widget on the screen; automatically detecting and branching on conditionals; and providing a visual indicator of which step the user is currently “up to” in the procedure. It is to be appreciated that the present invention is not limited to the preceding features to aid a user in performing a documented procedure and, thus, given the teachings of the present invention provided herein, other features may also be employed while maintaining the spirit of the present invention.
  • It is to be appreciated that as used herein the words “highlighting” and “highlight” refer to any visual, audible, and tactile indication or cue that may be used to identify (or make stand out) an item of interest in a documented procedure. Such items may include, but are not limited to, a step or portion thereof, a word, a phrase, a symbol, a hyperlink, a graphic, and any other item that may be found in a documented procedure. It is to be further appreciated that the present invention is not limited to the preceding ways in which to identify an item, or to the items specified and, thus, other ways and other items may be used to provide “highlighting” as used herein, while maintaining the spirit and scope of the present invention.
  • The present description illustrates the principles of the present invention. It will thus be appreciated that those skilled in the art will be able to devise various arrangements that, although not explicitly described or shown herein, embody the principles of the invention and are included within its spirit and scope.
  • All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the principles of the invention and the concepts contributed by the inventor to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions.
  • Moreover, all statements herein reciting principles, aspects, and embodiments of the invention, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
  • Thus, for example, it will be appreciated by those skilled in the art that the block diagrams presented herein represent conceptual views of illustrative circuitry embodying the principles of the invention. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudocode, and the like represent various processes which may be substantially represented in computer readable media and so executed by a computer or processor, whether or not such computer or processor is explicitly shown.
  • The functions of the various elements shown in the figures may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software. When provided by a processor, the functions may be provided by a single dedicated processor, by a single shared processor, or by a plurality of individual processors, some of which may be shared. Moreover, explicit use of the term “processor” or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage.
  • Other hardware, conventional and/or custom, may also be included. Similarly, any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • In the claims hereof, any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that software to perform the function. The invention as defined by such claims resides in the fact that the functionalities provided by the various recited means are combined and brought together in the manner which the claims call for. Applicant thus regards any means that can provide those functionalities as equivalent to those shown herein.
  • FIG. 1 is a block diagram illustrating a system 100 for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention.
  • The system 100 is intended to cooperate with (or be included in) a computer 199 that is to be used to perform the documented procedure. For example, the computer 199 may simply be a desktop computer, a laptop computer, a personal digital assistant, an electronic tablet, and so forth that is capable of displaying a documented procedure for a user to execute. The computer 199 may optionally have the capability to access a network (e.g., the Internet, a company intranet, etc.) to download a documented procedure. Thus, it is to be appreciated that the documented procedure may be located locally or remotely with respect to the computer 199. The computer 199 includes one or more processors (hereinafter “processor”) 188, one or more memories (hereinafter “memory”) 187, a network interface 186, and one or more buses (hereinafter “bus”) 185. The computer 199 further includes a user interface 198 for allowing a user to interact with the documented procedure. The user interface 198 may include a keyboard or keypad (hereinafter “keyboard”) 197 and a display 196. The display 196 may be a touch-screen display capable of receiving user inputs via the display. The present invention is not limited to interacting with only computers having the configuration as shown in FIG. 1 and, thus, other computers, computing devices and corresponding configurations may also be employed in accordance with the present invention while maintaining the spirit of the present invention.
  • The system 100 includes an alignment device 110. The alignment device 110 may receive inputs from the user interface 198 or may include a camera (also interchangeably referred to herein as “imaging device”) 111 or other device for use in determining demonstrated user actions. The alignment device 110 preferably further includes one or more processors (hereinafter “processor”) 112, one or more memories (hereinafter “memory”) 113, and one or more buses (hereinafter “bus”) 114 for interconnecting the elements of the alignment device 110. Additionally, the alignment device 110 may include a step extractor 119 for extracting steps from a documented procedure. The step extractor 119 may include a scanner and/or other device capable of optical character recognition, a parser, a memory for storing words/phrases/etc. that may be compared against parsed words/phrases/etc., and/or so forth.
  • The alignment device 110 computes an alignment between the document actions specified in a documented procedure and corresponding demonstrated user actions. Based on the computed alignment, a step or portion of a step (e.g., word(s), graphic(s), and so forth) in the documented procedure is highlighted so that the user can identify where he or she is (e.g., which step is the current step for which a user action needs to be taken) within the documented procedure. Optionally, a next step may also be highlighted, preferably in a different manner than the current step. The alignment may be used to automatically scroll through the procedure as the user progresses there through.
  • It is to be appreciated that the connection configuration depicted in FIG. 1 is merely illustrative and, thus, given the teachings of the present invention provided herein, one of ordinary skill in the related art will contemplate these and other connection configurations and other elements that may be employed with respect to a system for aligning document actions specified in a documented procedure with corresponding demonstrated user actions while maintaining the spirit of the present invention.
  • FIG. 2 is a flow diagram illustrating a method for aligning document actions specified in a documented procedure with corresponding demonstrated user actions, according to an illustrative embodiment of the present invention.
  • The document is marked up to identify a plurality of document actions (step 205). A document action is an action expected or mandated to be taken with respect to a given document. A list that includes the plurality of document actions is generated (step 210). Steps 205 and 210 are preferably performed prior to a user attempting to perform the steps specified in the documented procedure.
  • The next step in the procedure is highlighted (step 215). The system state and demonstrated user action are recorded (step 220). An alignment is computed between the demonstrated user action and the document actions in the list based on at least one or more similarity metrics (step 230). Based on the alignment, the procedure may be advanced, if necessary, and the most likely document action in the procedure (i.e., the current document action to be taken by the user) is highlighted (step 240).
  • It is then determined whether or not the user has completed the procedure (step 250). If the user has not completed the procedure, then the method returns to step 215 to highlight the next step in the procedure. Preferably, the next step is highlighted in a manner different from that employed with respect to the current step, so as to avoid user confusion. Otherwise, if the user has completed the procedure, then the method is terminated. It is to be appreciated that highlighting with respect to the current step and/or the next step may involve highlighting a whole textual and/or graphical description of a step, a portion thereof, an element of the user interface (including, but not limited to, action buttons (e.g., cut, paste, save, etc.), and so forth.
  • FIG. 3 is a flow diagram further illustrating step 230 of FIG. 2, according to an illustrative embodiment of the present invention. Step 230 can also be considered as computing a similarity between a user-demonstrated state/action and steps in a documented procedure (document actions, as specified in the list). According to an illustrative embodiment of the present invention, similarity may be based on a probabilistic combination of similarity parameters.
  • The number of words in common between a textual description of a demonstrated user action and document action is determined (step 310).
  • The similarity of the action types involved (e.g., clicking button and clicking checkbox are more similar than typing) with respect to the demonstrated user action and the document actions is determined (step 320).
  • A probability distribution (e.g., including, but not limited to, a Gaussian distribution) over two or more steps is determined (step 330). In such an approach, the next step in temporal order is usually the most likely.
  • Any common strings in the user interface (e.g., button labels) and the documented procedure are determined (step 340).
  • An alignment based on one or more of steps 310 through 340 is then calculated (step 350). It is to be appreciated that the alignment may also be based on experiences of previous users, using any of the results of steps 310 through even 350 for the previous users.
  • It is to be appreciated that the present invention is not limited to the preceding alignment or similarity parameters and, thus, given the teachings of the present invention provided herein, other parameters may also be employed to determine alignment or similarity while maintaining the spirit of the present invention.
  • These and other features and advantages of the present invention may be readily ascertained by one of ordinary skill in the pertinent art based on the teachings herein. It is to be understood that the teachings of the present invention may be implemented in various forms of hardware, software, firmware, special purpose processors, or combinations thereof.
  • Most preferably, the teachings of the present invention are implemented as a combination of hardware and software. Moreover, the software is preferably implemented as an application program tangibly embodied on a program storage unit. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPU”), a random access memory (“RAM”), and input/output (“I/O”) interfaces. The computer platform may also include an operating system and microinstruction code. The various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.
  • It is to be further understood that, because some of the constituent system components and methods depicted in the accompanying drawings are preferably implemented in software, the actual connections between the system components or the process function blocks may differ depending upon the manner in which the present invention is programmed. Given the teachings herein, one of ordinary skill in the pertinent art will be able to contemplate these and similar implementations or configurations of the present invention.
  • Although the illustrative embodiments have been described herein with reference to the accompanying drawings, it is to be understood that the present invention is not limited to those precise embodiments, and that various changes and modifications may be effected therein by one of ordinary skill in the pertinent art without departing from the scope or spirit of the present invention. All such changes and modifications are intended to be included within the scope of the present invention as set forth in the appended claims.

Claims (17)

1. (canceled)
2. The apparatus of claim 10, wherein said alignment device identifies a current step in the documented procedure based on the alignment.
3. The apparatus of claim 2, wherein the documented procedure is displayed on a user interface, and said alignment device provides an identification of the current step to the user interface for use in highlighting the current step on the user interface.
4. The apparatus of claim 2, further comprising a user interface, in signal communication with said alignment device, for displaying the documented procedure, and wherein said alignment device outputs an identification of the current step to the user interface, and said user interface highlights the current step based on the identification.
5. The apparatus of claim 10, wherein said alignment device identifies a next step in the documented procedure based on the alignment.
6. The apparatus of claim 5, wherein the documented procedure is displayed on a user interface, and said alignment device provides an identification of the next step to the user interface for use in highlighting the next step on the user interface.
7. The apparatus of claim 5, further comprising a user interface, in signal communication with said alignment device, for displaying the documented procedure, and wherein said alignment device outputs an identification of the next step to the user interface, and said user interface highlights the next step based on the identification.
8. The apparatus of claim 10, wherein the at least one similarity metric is based on at least one of a number of words in common between a textual description of a demonstrated user action and a document action, a similarity of an action type involved in both a demonstrated user action and a document action, a probability distribution over at least two steps of the documented procedure.
9. The apparatus of claim 10 wherein the demonstrated user actions are performed with respect to a user interface, and the at least one similarity metric is based on common strings in the user interface and the documented procedure.
10. An apparatus for aligning document actions specified in a documented procedure with demonstrated user actions, comprising:
an alignment device for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric; and
a camera, connected in signal communication with said alignment device, for capturing images of at least some of the demonstrated user actions, and
wherein said alignment device uses at least some of the captured images to calculate the alignment.
11. The apparatus of claim 10 wherein said alignment device computes the alignment further based on previous user experiences with respect to performing the documented procedure.
12. The apparatus of claim 10 wherein the documented procedure has a plurality of steps specified therein, and the apparatus further comprises a document action extractor for extracting at least some of the plurality of steps.
13. (canceled)
14. The apparatus of claim 15 wherein the at least one similarity metric is based on at least one of a number of words in common between a textual description of a demonstrated user action and a document action, a similarity of an action type involved in both a demonstrated user action and a document action, a probability distribution over at least two steps of the documented procedure, and common strings in the user interface and the documented procedure.
15. An apparatus for guiding user actions with respect to a documented procedure displayed on a user interface, the documented procedure having a plurality of steps specified therein, the apparatus comprising:
an alignment device for computing an alignment between the document actions and the demonstrated user actions based on at least one similarity metric;
a highlighting device, connected in signal communication with said alignment device, for visually highlighting on the user interface at least a portion of a step in the documented procedure based on the alignment computed by said alignment device;
a scrolling device, connected in signal communication with said alignment device, for automatically scrolling through each of the plurality of steps of the documented procedure on the user interface based on the alignment computed by said alignment device; and
a camera, connected in signal communication with said alignment device, for capturing images of at least some of the demonstrated user actions, and
wherein said alignment device uses at least some of the captured images to calculate the alignment.
16. The apparatus of claim 15 wherein said alignment device computes the alignment further based on previous user experiences with respect to performing the documented procedure.
17-32. (canceled)
US12/041,123 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation Abandoned US20080155410A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/041,123 US20080155410A1 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/993,377 US8176422B2 (en) 2004-11-19 2004-11-19 Method for aligning demonstrated user actions with existing documentation
US12/041,123 US20080155410A1 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/993,377 Continuation US8176422B2 (en) 2004-11-19 2004-11-19 Method for aligning demonstrated user actions with existing documentation

Publications (1)

Publication Number Publication Date
US20080155410A1 true US20080155410A1 (en) 2008-06-26

Family

ID=36462280

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/993,377 Active 2031-01-22 US8176422B2 (en) 2004-11-19 2004-11-19 Method for aligning demonstrated user actions with existing documentation
US12/041,123 Abandoned US20080155410A1 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation
US12/041,142 Active 2028-11-01 US8607142B2 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation
US12/041,171 Abandoned US20080155395A1 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/993,377 Active 2031-01-22 US8176422B2 (en) 2004-11-19 2004-11-19 Method for aligning demonstrated user actions with existing documentation

Family Applications After (2)

Application Number Title Priority Date Filing Date
US12/041,142 Active 2028-11-01 US8607142B2 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation
US12/041,171 Abandoned US20080155395A1 (en) 2004-11-19 2008-03-03 Method for aligning demonstrated user actions with existing documentation

Country Status (1)

Country Link
US (4) US8176422B2 (en)

Families Citing this family (61)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8127235B2 (en) 2007-11-30 2012-02-28 International Business Machines Corporation Automatic increasing of capacity of a virtual space in a virtual world
US20090164919A1 (en) 2007-12-24 2009-06-25 Cary Lee Bates Generating data for managing encounters in a virtual world environment
JP5159375B2 (en) 2008-03-07 2013-03-06 インターナショナル・ビジネス・マシーンズ・コーポレーション Object authenticity determination system and method in metaverse, and computer program thereof
US9205328B2 (en) 2010-02-18 2015-12-08 Activision Publishing, Inc. Videogame system and method that enables characters to earn virtual fans by completing secondary objectives
US9682324B2 (en) 2010-05-12 2017-06-20 Activision Publishing, Inc. System and method for enabling players to participate in asynchronous, competitive challenges
TW201227631A (en) * 2010-12-22 2012-07-01 Altek Corp Interactive learning system and method thereof
US10137376B2 (en) 2012-12-31 2018-11-27 Activision Publishing, Inc. System and method for creating and streaming augmented game sessions
US10286326B2 (en) 2014-07-03 2019-05-14 Activision Publishing, Inc. Soft reservation system and method for multiplayer video games
US11351466B2 (en) 2014-12-05 2022-06-07 Activision Publishing, Ing. System and method for customizing a replay of one or more game events in a video game
US10118099B2 (en) 2014-12-16 2018-11-06 Activision Publishing, Inc. System and method for transparently styling non-player characters in a multiplayer video game
US10486068B2 (en) 2015-05-14 2019-11-26 Activision Publishing, Inc. System and method for providing dynamically variable maps in a video game
US10286314B2 (en) 2015-05-14 2019-05-14 Activision Publishing, Inc. System and method for providing continuous gameplay in a multiplayer video game through an unbounded gameplay session
US10315113B2 (en) 2015-05-14 2019-06-11 Activision Publishing, Inc. System and method for simulating gameplay of nonplayer characters distributed across networked end user devices
US10668367B2 (en) 2015-06-15 2020-06-02 Activision Publishing, Inc. System and method for uniquely identifying physical trading cards and incorporating trading card game items in a video game
US10471348B2 (en) 2015-07-24 2019-11-12 Activision Publishing, Inc. System and method for creating and sharing customized video game weapon configurations in multiplayer video games via one or more social networks
US10099140B2 (en) 2015-10-08 2018-10-16 Activision Publishing, Inc. System and method for generating personalized messaging campaigns for video game players
US11185784B2 (en) 2015-10-08 2021-11-30 Activision Publishing, Inc. System and method for generating personalized messaging campaigns for video game players
US10245509B2 (en) 2015-10-21 2019-04-02 Activision Publishing, Inc. System and method of inferring user interest in different aspects of video game streams
US10376781B2 (en) 2015-10-21 2019-08-13 Activision Publishing, Inc. System and method of generating and distributing video game streams
US10232272B2 (en) 2015-10-21 2019-03-19 Activision Publishing, Inc. System and method for replaying video game streams
US10694352B2 (en) 2015-10-28 2020-06-23 Activision Publishing, Inc. System and method of using physical objects to control software access
US10300390B2 (en) 2016-04-01 2019-05-28 Activision Publishing, Inc. System and method of automatically annotating gameplay of a video game based on triggering events
US10226701B2 (en) 2016-04-29 2019-03-12 Activision Publishing, Inc. System and method for identifying spawn locations in a video game
US10179289B2 (en) 2016-06-21 2019-01-15 Activision Publishing, Inc. System and method for reading graphically-encoded identifiers from physical trading cards through image-based template matching
US10573065B2 (en) 2016-07-29 2020-02-25 Activision Publishing, Inc. Systems and methods for automating the personalization of blendshape rigs based on performance capture data
US10042613B2 (en) * 2016-08-19 2018-08-07 International Business Machines Corporation System, method, and recording medium for validating computer documentation
US10463964B2 (en) 2016-11-17 2019-11-05 Activision Publishing, Inc. Systems and methods for the real-time generation of in-game, locally accessible heatmaps
US10709981B2 (en) 2016-11-17 2020-07-14 Activision Publishing, Inc. Systems and methods for the real-time generation of in-game, locally accessible barrier-aware heatmaps
US10500498B2 (en) 2016-11-29 2019-12-10 Activision Publishing, Inc. System and method for optimizing virtual games
US10055880B2 (en) 2016-12-06 2018-08-21 Activision Publishing, Inc. Methods and systems to modify a two dimensional facial image to increase dimensional depth and generate a facial image that appears three dimensional
US10861079B2 (en) 2017-02-23 2020-12-08 Activision Publishing, Inc. Flexible online pre-ordering system for media
US10818060B2 (en) 2017-09-05 2020-10-27 Activision Publishing, Inc. Systems and methods for guiding motion capture actors using a motion reference system
US10561945B2 (en) 2017-09-27 2020-02-18 Activision Publishing, Inc. Methods and systems for incentivizing team cooperation in multiplayer gaming environments
US10974150B2 (en) 2017-09-27 2021-04-13 Activision Publishing, Inc. Methods and systems for improved content customization in multiplayer gaming environments
US11040286B2 (en) 2017-09-27 2021-06-22 Activision Publishing, Inc. Methods and systems for improved content generation in multiplayer gaming environments
US10463971B2 (en) 2017-12-06 2019-11-05 Activision Publishing, Inc. System and method for validating video gaming data
US10537809B2 (en) 2017-12-06 2020-01-21 Activision Publishing, Inc. System and method for validating video gaming data
US10981051B2 (en) 2017-12-19 2021-04-20 Activision Publishing, Inc. Synchronized, fully programmable game controllers
US10864443B2 (en) 2017-12-22 2020-12-15 Activision Publishing, Inc. Video game content aggregation, normalization, and publication systems and methods
US10596471B2 (en) 2017-12-22 2020-03-24 Activision Publishing, Inc. Systems and methods for enabling audience participation in multi-player video game play sessions
US11278813B2 (en) 2017-12-22 2022-03-22 Activision Publishing, Inc. Systems and methods for enabling audience participation in bonus game play sessions
US11263670B2 (en) 2018-11-19 2022-03-01 Activision Publishing, Inc. Systems and methods for dynamically modifying video game content based on non-video gaming content being concurrently experienced by a user
US11192028B2 (en) 2018-11-19 2021-12-07 Activision Publishing, Inc. Systems and methods for the real-time customization of video game content based on player data
US11115712B2 (en) 2018-12-15 2021-09-07 Activision Publishing, Inc. Systems and methods for indexing, searching for, and retrieving digital media
US11679330B2 (en) 2018-12-18 2023-06-20 Activision Publishing, Inc. Systems and methods for generating improved non-player characters
US11305191B2 (en) 2018-12-20 2022-04-19 Activision Publishing, Inc. Systems and methods for controlling camera perspectives, movements, and displays of video game gameplay
US11344808B2 (en) 2019-06-28 2022-05-31 Activision Publishing, Inc. Systems and methods for dynamically generating and modulating music based on gaming events, player profiles and/or player reactions
US11097193B2 (en) 2019-09-11 2021-08-24 Activision Publishing, Inc. Methods and systems for increasing player engagement in multiplayer gaming environments
US11423605B2 (en) 2019-11-01 2022-08-23 Activision Publishing, Inc. Systems and methods for remastering a game space while maintaining the underlying game simulation
US11712627B2 (en) 2019-11-08 2023-08-01 Activision Publishing, Inc. System and method for providing conditional access to virtual gaming items
US11537209B2 (en) 2019-12-17 2022-12-27 Activision Publishing, Inc. Systems and methods for guiding actors using a motion capture reference system
US11420122B2 (en) 2019-12-23 2022-08-23 Activision Publishing, Inc. Systems and methods for controlling camera perspectives, movements, and displays of video game gameplay
US11563774B2 (en) 2019-12-27 2023-01-24 Activision Publishing, Inc. Systems and methods for tracking and identifying phishing website authors
US11351459B2 (en) 2020-08-18 2022-06-07 Activision Publishing, Inc. Multiplayer video games with virtual characters having dynamically generated attribute profiles unconstrained by predefined discrete values
US11524234B2 (en) 2020-08-18 2022-12-13 Activision Publishing, Inc. Multiplayer video games with virtual characters having dynamically modified fields of view
US11724188B2 (en) 2020-09-29 2023-08-15 Activision Publishing, Inc. Methods and systems for selecting a level of detail visual asset during the execution of a video game
US11833423B2 (en) 2020-09-29 2023-12-05 Activision Publishing, Inc. Methods and systems for generating level of detail visual assets in a video game
US11717753B2 (en) 2020-09-29 2023-08-08 Activision Publishing, Inc. Methods and systems for generating modified level of detail visual assets in a video game
US11439904B2 (en) 2020-11-11 2022-09-13 Activision Publishing, Inc. Systems and methods for imparting dynamic and realistic movement to player-controlled avatars in video games
US11853439B2 (en) 2020-12-30 2023-12-26 Activision Publishing, Inc. Distributed data storage system providing enhanced security
US11794107B2 (en) 2020-12-30 2023-10-24 Activision Publishing, Inc. Systems and methods for improved collision detection in video games

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5481667A (en) * 1992-02-13 1996-01-02 Microsoft Corporation Method and system for instructing a user of a computer system how to perform application program tasks
US5493658A (en) * 1992-03-26 1996-02-20 International Business Machines Corporation Interactive online tutorial system with monitoring function for software products
US5715415A (en) * 1996-06-05 1998-02-03 Microsoft Corporation Computer application with help pane integrated into workspace
US5825356A (en) * 1996-03-18 1998-10-20 Wall Data Incorporated Help system with semitransparent window for disabling controls
US6259445B1 (en) * 1997-07-07 2001-07-10 Informix, Inc. Computer-based documentation and instruction
US6262730B1 (en) * 1996-07-19 2001-07-17 Microsoft Corp Intelligent user assistance facility
US6601012B1 (en) * 2000-03-16 2003-07-29 Microsoft Corporation Contextual models and methods for inferring attention and location
US7000187B2 (en) * 1999-07-01 2006-02-14 Cisco Technology, Inc. Method and apparatus for software technical support and training
US7233933B2 (en) * 2001-06-28 2007-06-19 Microsoft Corporation Methods and architecture for cross-device activity monitoring, reasoning, and visualization for providing status and forecasts of a users' presence and availability
US7640548B1 (en) * 2002-06-21 2009-12-29 Siebel Systems, Inc. Task based user interface

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5434963A (en) * 1988-09-03 1995-07-18 Hitachi, Ltd. Method and system of help-information control method and system
US5175812A (en) * 1988-11-30 1992-12-29 Hewlett-Packard Company System for providing help information during a help mode based on selected operation controls and a current-state of the system
JPH08137647A (en) * 1994-11-07 1996-05-31 Fujitsu Ltd Method and device for reproducing operation guide information reproduction, method and device for reproducing multimedia information

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5481667A (en) * 1992-02-13 1996-01-02 Microsoft Corporation Method and system for instructing a user of a computer system how to perform application program tasks
US5493658A (en) * 1992-03-26 1996-02-20 International Business Machines Corporation Interactive online tutorial system with monitoring function for software products
US5825356A (en) * 1996-03-18 1998-10-20 Wall Data Incorporated Help system with semitransparent window for disabling controls
US5715415A (en) * 1996-06-05 1998-02-03 Microsoft Corporation Computer application with help pane integrated into workspace
US6262730B1 (en) * 1996-07-19 2001-07-17 Microsoft Corp Intelligent user assistance facility
US6259445B1 (en) * 1997-07-07 2001-07-10 Informix, Inc. Computer-based documentation and instruction
US7000187B2 (en) * 1999-07-01 2006-02-14 Cisco Technology, Inc. Method and apparatus for software technical support and training
US6601012B1 (en) * 2000-03-16 2003-07-29 Microsoft Corporation Contextual models and methods for inferring attention and location
US7233933B2 (en) * 2001-06-28 2007-06-19 Microsoft Corporation Methods and architecture for cross-device activity monitoring, reasoning, and visualization for providing status and forecasts of a users' presence and availability
US7640548B1 (en) * 2002-06-21 2009-12-29 Siebel Systems, Inc. Task based user interface

Also Published As

Publication number Publication date
US8176422B2 (en) 2012-05-08
US20060112322A1 (en) 2006-05-25
US20080155393A1 (en) 2008-06-26
US20080155395A1 (en) 2008-06-26
US8607142B2 (en) 2013-12-10

Similar Documents

Publication Publication Date Title
US8176422B2 (en) Method for aligning demonstrated user actions with existing documentation
US7895030B2 (en) Visualization method for machine translation
US7983897B2 (en) Machine translation feedback
US20080077393A1 (en) Virtual keyboard adaptation for multilingual input
US20190018560A1 (en) Terminal and method for determining type of input method editor
US9218066B2 (en) Method for character correction
CN105580384A (en) Actionable content displayed on a touch screen
WO2012055714A1 (en) Automated adjustment of input configuration
CN104412212A (en) Input method editor
CN109918685B (en) Computer-aided translation method, device, computer equipment and storage medium
US7290220B2 (en) Method and apparatus for non-sequential access of form fields
US20200111491A1 (en) Speech enabled user interaction
RU2665274C2 (en) Pop-up verification panel
CN112631437A (en) Information recommendation method and device and electronic equipment
CN111880668A (en) Input display method and device and electronic equipment
CN114417784A (en) Comment display method, device, equipment and medium for document
JP2018010532A (en) Information processing device, program and information processing method
US20050183033A1 (en) Apparatus and methods for displaying dialog box text messages including languages having different reading orders
CN113626441A (en) Text management method, device and equipment based on scanning equipment and storage medium
CN107908792B (en) Information pushing method and device
JP2011154590A (en) Program and information processing apparatus
JP2019215936A (en) Automatic translation device and automatic translation program
Cuff et al. Statistical analysis of surgical pathology data using the R program
KR102118322B1 (en) Document translation server and translation method for generating original and translation files individually
CN111966267A (en) Application comment method and device and electronic equipment

Legal Events

Date Code Title Description
AS Assignment

Owner name: ACTIVISION PUBLISHING, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:029900/0285

Effective date: 20121231

STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION