US20070058037A1 - User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same - Google Patents
User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same Download PDFInfo
- Publication number
- US20070058037A1 US20070058037A1 US11/407,217 US40721706A US2007058037A1 US 20070058037 A1 US20070058037 A1 US 20070058037A1 US 40721706 A US40721706 A US 40721706A US 2007058037 A1 US2007058037 A1 US 2007058037A1
- Authority
- US
- United States
- Prior art keywords
- information
- receptacle
- prohibited object
- image
- user interface
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000012216 screening Methods 0.000 title claims abstract description 58
- 238000001514 detection method Methods 0.000 claims abstract description 119
- 238000000034 method Methods 0.000 claims abstract description 80
- 238000012545 processing Methods 0.000 claims description 56
- 230000003287 optical effect Effects 0.000 claims description 43
- 230000035945 sensitivity Effects 0.000 claims description 22
- 238000004891 communication Methods 0.000 claims description 15
- 238000002591 computed tomography Methods 0.000 claims description 10
- 230000005855 radiation Effects 0.000 claims description 7
- 230000005251 gamma ray Effects 0.000 claims description 5
- 230000008569 process Effects 0.000 description 18
- 230000004044 response Effects 0.000 description 13
- 238000010586 diagram Methods 0.000 description 12
- 238000007781 pre-processing Methods 0.000 description 11
- 230000000007 visual effect Effects 0.000 description 11
- 239000002131 composite material Substances 0.000 description 8
- 238000001914 filtration Methods 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 4
- 230000007812 deficiency Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 238000012015 optical character recognition Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000001427 coherent effect Effects 0.000 description 2
- 230000006378 damage Effects 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 2
- 239000002184 metal Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000149 penetrating effect Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 230000001143 conditioned effect Effects 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000034994 death Effects 0.000 description 1
- 231100000517 death Toxicity 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 229910003460 diamond Inorganic materials 0.000 description 1
- 239000010432 diamond Substances 0.000 description 1
- 235000019800 disodium phosphate Nutrition 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000009021 linear effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000000513 principal component analysis Methods 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 238000001931 thermography Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
- 230000003245 working effect Effects 0.000 description 1
Images
Classifications
-
- G01V5/22—
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N23/00—Investigating or analysing materials by the use of wave or particle radiation, e.g. X-rays or neutrons, not covered by groups G01N3/00 – G01N17/00, G01N21/00 or G01N22/00
- G01N23/02—Investigating or analysing materials by the use of wave or particle radiation, e.g. X-rays or neutrons, not covered by groups G01N3/00 – G01N17/00, G01N21/00 or G01N22/00 by transmitting the radiation through the material
- G01N23/04—Investigating or analysing materials by the use of wave or particle radiation, e.g. X-rays or neutrons, not covered by groups G01N3/00 – G01N17/00, G01N21/00 or G01N22/00 by transmitting the radiation through the material and forming images of the material
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/88—Image or video recognition using optical means, e.g. reference filters, holographic masks, frequency domain filters or spatial domain filters
- G06V10/89—Image or video recognition using optical means, e.g. reference filters, holographic masks, frequency domain filters or spatial domain filters using frequency domain filters, e.g. Fourier masks implemented on spatial light modulators
- G06V10/893—Image or video recognition using optical means, e.g. reference filters, holographic masks, frequency domain filters or spatial domain filters using frequency domain filters, e.g. Fourier masks implemented on spatial light modulators characterised by the kind of filter
- G06V10/898—Image or video recognition using optical means, e.g. reference filters, holographic masks, frequency domain filters or spatial domain filters using frequency domain filters, e.g. Fourier masks implemented on spatial light modulators characterised by the kind of filter combination of filters, e.g. phase-only filters
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/05—Recognition of patterns representing particular kinds of hidden objects, e.g. weapons, explosives, drugs
Definitions
- the present invention relates generally to security systems and, more particularly, to a user interface for use in screening luggage, mail parcels or cargo containers to identify certain objects located therein or for screening persons to identify certain objects located thereon and to a method and system for implementing such a user interface.
- security-screening systems at airports make use of devices generating penetrating radiation, such as x-ray devices, to scan individual pieces of luggage to generate an image conveying the contents of the luggage.
- the image is displayed on a screen and is examined by a human operator whose task it is to identify, on the basis of the image, potentially threatening objects located in the luggage.
- a deficiency with current systems is that they are entirely reliant on the human operator to identify potentially threatening objects.
- the performance of the human operator greatly varies according to such factors as poor training and fatigue.
- the process of detection and identification of threatening objects is highly susceptible to human error.
- the invention provides a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening.
- the method comprises receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle.
- the method also includes displaying first information conveying an image associated with the receptacle on the basis of the image signal.
- the method also includes displaying second information conveying a presence of at least one prohibited object in the receptacle, the second information being displayed simultaneously with the first information.
- the method also includes providing a control allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the at least one prohibited object.
- the first information and second information are displayed in a first viewing window and the third information is displayed in a second viewing window.
- the control is adapted for causing the second viewing window to be displayed to a user.
- control allows the user to cause the third information to be displayed by using any suitable input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
- control may be embodied in any suitable form including, for example, a selection box.
- the graphical user interface module displays a prohibited object list, the prohibited object list including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected.
- at least one entry in the prohibited object list conveys a level of confidence associated to detection of the corresponding prohibited object in the receptacle.
- the graphical user interface module is adapted for enabling a user to select an entry from the plurality of entries in the prohibited object list and to display information associated with the entry selected by the user.
- the type of information displayed associated with the entry selected can vary from one implementation to the other depending on the specific application. Examples of the type of information that can be displayed include an image of the selected prohibited object detected and a risk levels associated to the prohibited object detected.
- the second information conveys the presence of at least one prohibited object in the receptacle in either textual format or graphical format.
- the second information conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
- the method comprises receiving a detection signal conveying the presence of at least one prohibited object in the receptacle and processing the detection signal to derive the second information.
- the detection signal may convey a plurality of information items.
- the detection signal may provide information allowing to identify the type of prohibited object detected, the level of confidence associated to the detection process and the position of the prohibited object in the receptacle amongst others.
- the graphical user interface module may be operative for processing the data element indicative of the level of confidence in combination with a detection sensitivity level.
- the second information may be omitted from the user interface module.
- the user interface provides a control allowing a user to modify the detection sensitivity level.
- receptacle is used to broadly describe an entity adapted for receiving objects therein such as, for example, a luggage item, a cargo container or a mail parcel.
- the expression “luggage item” is used to broadly describe luggage, suitcases, handbags, backpacks, briefcases, boxes, parcels or any other similar type of item suitable for containing objects therein.
- the invention provides and apparatus suitable for implementing a user interface for use in screening a receptacle to detect the presence of one or more prohibited objects in accordance with the above described method.
- the invention provides a computer readable storage medium including a program element suitable for execution by a CPU for implementing a graphical user interface module for displaying information associated to the content of a receptacle in accordance with the above described method.
- the invention provides an apparatus for implementing a user interface module for use in screening receptacles to detect the presence of one or more prohibited objects.
- the apparatus comprises means for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle.
- the apparatus also comprises means for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle.
- the apparatus also comprises means for implementing the user interface module.
- the user interface module is adapted for displaying first information conveying an image associated with the receptacle on the basis of the image signal.
- the user interface module is also adapted for displaying second information conveying a presence of at least one prohibited object in the receptacle, the second information being displayed simultaneously with the first information and being derived at least in part on the basis of the detection signal.
- the user interface module is also adapted for providing a control allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the at least one prohibited object.
- the apparatus also comprises means for releasing a signal adapted to cause a display module to display the user interface module.
- the invention provides a system for detecting the presence of one or more prohibited objects in a receptacle.
- the system includes an input, an optical correlator and an apparatus for implementing a user interface module.
- the input is for receiving data conveying graphic information regarding the contents of the receptacle.
- the optical correlator is in communication with the input and is operative for processing the graphic information to detect a depiction of the one or more prohibited objects in the receptacle.
- the apparatus implementing the user interface module is in communication with the input and with the optical correlator.
- the user interface module implemented by the apparatus is operative for displaying first information conveying an image associated with the receptacle on the basis of the data conveying graphic information regarding the contents of the receptacle.
- the user interface module is also adapted for displaying second information conveying a presence of one or more prohibited objects in the receptacle, the second information being displayed simultaneously with the first information.
- the user interface module is also operative for releasing a signal adapted to cause a display module to display the user interface module.
- the invention provides a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening.
- the method comprises receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle and displaying first information conveying an image associated with the receptacle on the basis of the image signal.
- the method also comprises receiving a detection signal conveying a presence of at least one prohibited object in the receptacle and displaying second information conveying a presence of at least one prohibited object in the receptacle.
- the second information is displayed simultaneously with the first information and is derived at least in part on the basis of the detection signal.
- the invention provides an apparatus suitable for implementing a user interface for use in screening a receptacle to detect the presence of one or more prohibited objects in accordance with the above described method.
- the invention provides a computer readable storage medium including a program element suitable for execution by a CPU for implementing a graphical user interface module for displaying information associated to the content of a receptacle in accordance with the above described method.
- the invention provides an apparatus for implementing a user interface module for use in screening a person to detect the presence of one or more prohibited objects.
- the apparatus comprising a first input for receiving an image signal associated with the person and a second input for receiving a detection signal conveying a presence of at least one prohibited object on the person.
- the apparatus also includes a processing unit operative for implementing the user interface module.
- the user interface module is adapted for displaying first information conveying an image associated with the person on the basis of the image signal and displaying second information conveying a presence of at least one prohibited object on the person.
- the second information is displayed simultaneously with the first information and is derived at least in part on the basis of the detection signal.
- the apparatus also includes an output for releasing a signal adapted to cause a display module to display said user interface module.
- FIG. 1 is a high-level block diagram of a system for screening a receptacle to detect therein the presence of one or more prohibited objects in accordance with a specific example of implementation of the present invention
- FIG. 2 is a block diagram of an output module suitable for use in connection with the system depicted in FIG. 1 in accordance with a specific example of implementation of the present invention
- FIG. 3 is a block diagram of an apparatus suitable for implementing a user interface for use in screening a receptacle in accordance with a specific example of implementation of the present invention
- FIG. 4 shows a flow diagram depicting a process for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening;
- FIG. 5 a and 5 b depict viewing windows of a user interface module displayed by the output module of FIG. 2 in accordance with a specific example of implementation of the present invention
- FIG. 5 c depicts a viewing window of a user interface module displayed by the output module of FIG. 2 in accordance with an alternative specific example of implementation of the present invention
- FIG. 5 d depicts a control window of a user interface module displayed by the apparatus of FIG. 3 allowing a user to select screening options in accordance with a specific example of implementation of the present invention
- FIG. 6 is a block diagram of an apparatus for processing images suitable for use in connection with the system depicted in FIG. 1 in accordance with a specific example of implementation of the present invention
- FIG. 7 is a flow diagram depicting a process for detecting a presence of at least one target object in the receptacle in accordance with specific examples of implementation of the present invention.
- FIG. 8 shows three images associated to a target object suitable for use in connection with the system depicted in FIG. 1 , each image depicting the target object in a different orientation, in accordance with a specific example of implementation of the present invention
- FIG. 9 shows a mosaic image including a plurality of sub-images associated with a target object suitable for use in connection with the system depicted in FIG. 1 , each sub-image depicting the target object in a different orientation and scale, in accordance with a specific example of implementation of the present invention
- FIG. 10 is a functional block diagram a luggage screening system including an optical correlator in accordance with a specific example of implementation of the present invention.
- FIG. 11 is a block diagram depicting the functioning of an optical correlator in accordance with a specific example of implementation of the present invention.
- FIG. 12 depicts a Fourier transform, amplitude and phase, of the spatial domain image for number 2 ;
- FIG. 13 shows two images associated to a person suitable for use in a system for screening a person to detect the presence of one or more prohibited objects in accordance with a specific example of implementation of the present invention
- FIG. 14 is a block diagram of an apparatus suitable for implementing a user interface for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening in accordance with a specific example of implementation of the present invention
- FIG. 15 is a block diagram of an alternative implementation of an apparatus suitable for for use in detecting the presence of one or more prohibited objects in a receptacle during security screening in accordance with a specific example of implementation of the present invention
- FIG. 16 shows a functional block diagram of a client-server system suitable for use in screening a receptacle to detect therein the presence of one or more prohibited objects in accordance with an alternative specific example of implementation of the present invention.
- FIG. 1 Shown in FIG. 1 is a system 100 for screening a receptacle in accordance with a specific example of implementation of the present invention.
- receptacle as used for the purposes of the present description, is used to broadly describe an entity adapted for receiving objects therein such as, for example, a luggage item, a cargo container or a mail parcel.
- the expression “luggage item” is used to broadly describe luggage, suitcases, handbags, backpacks, briefcases, boxes, parcels or any other similar type of item suitable for containing objects therein.
- the system 100 includes an image generation device 102 ; an apparatus 106 in communication with the image generation device 102 and an output module 108 .
- the image generation device 102 generates an image signal associated with a receptacle 104 .
- the image signal conveys information related to the contents of the receptacle 104 .
- the apparatus 106 receives the image signal associated with the receptacle 104 and processes that image signal in combination with a plurality of target images associated with target objects to detect a presence of at least one target object in the receptacle 104 .
- the plurality of target images is stored in a database of target images 110 .
- the apparatus 106 In response to detection of the presence of at least one target object in the receptacle 104 , the apparatus 106 generates a detection signal conveying the presence of the target object in the receptacle 104 . Examples of the manner in which the detection signal can be derived are described later on in the specification.
- the output module 108 conveys to a user information derived at least in part on the basis of the detection signal to a user of the system.
- the system 100 provides assistance to the human security personnel using the system in detecting certain target objects, including prohibited objects, and decreases the susceptibility of the screening process to human error.
- the image generation device 102 uses penetrating radiation or emitted radiation to generate the image associated with the receptacle 104 .
- penetrating radiation or emitted radiation include, without being limited to, x-ray, gamma ray, computed tomography (CT scans), thermal imaging, TeraHertz and millimeter wave devices.
- CT scans computed tomography
- thermal imaging TeraHertz
- millimeter wave devices such devices are known in the art and as such will not be described further here.
- the image generation device 102 is a conventional x-ray machine adapted for generating an x-ray image of the receptacle 104 .
- the image signal generated by the image generation device 102 and associated with the receptacle 104 may be conveyed as a two-dimensional (2-D) image or as a three-dimensional (3-D) image and may be in any suitable format. Possible formats include, without being limited to, JPEG, GIF, TIFF and bitmap amongst others. Preferably, the image signal is in a format that can be displayed on a display screen.
- the database of target images 110 includes a plurality of entries associated to respective target objects that the system 100 is designed to detect.
- target image for each entry associated to a target object at least one image (hereinafter referred to as a “target image”) is provided in the database of target images 110 .
- the format of the target images will depend upon the image processing algorithm implemented by the apparatus 106 . More specifically, the format of the target images is such that a comparison operation can be performed by the apparatus 106 between the target images and data derived from the image signal associated with the receptacle 104 .
- a set of images is provided in the database of target images 1 10 .
- images depicting the target object in various orientations may be provided.
- FIG. 6 of the drawings depicts an example of arbitrary 3D orientations of a target object.
- characteristics of the target object are provided. Such characteristics may include, without being limited to, the name of the target object, its associated threat level, the recommended handling procedure when such a target object is detected and any other suitable information.
- the threat level information associated to the target object convey the relative threat level of a prohibited object compared to other prohibited objects in the database of target images 110 .
- a gun would be given a relatively high threat level while a metallic nail file would be given a relatively low level threat level and a pocket knife would be given a threat level between that of the nail file and the gun.
- each entry in the database of target images 110 is also associated to a respective target object identifier data element.
- the database of target images 110 includes at least one entry associated to a prohibited object, such as a weapon.
- a prohibited object such as a weapon.
- luggage screening in an airport facility for example
- the prohibited object typically constitutes a potential threat to the safety of the passenger or aircraft.
- the prohibited object is typically an object that is normally not permitted to be sent through the mail, such as guns (in Canada) for example, due to registration requirements/permits and so on.
- the database of target images 110 includes one or more entries associated to objects which are not prohibited but which may represent potential threats. For example, the presence of a metal plate or a metal canister in a piece of luggage going through luggage security screening is not prohibited in itself. However such objects may conceal one or more dangerous objects. As such, it is desirable to be able to detect the presence of such objects in receptacle such as to bring them to the attention of the security screeners.
- the specific design and content of the database of target images 110 may vary from one implementation to the next without detracting from the spirit of the invention.
- the design of the database is not critical to the present invention and as such will not be described further here.
- the database of target images 110 has been shown in FIG. 1 to be a component separate from the apparatus 106 , it will be appreciated that in certain embodiments the database of target images 110 may be part of apparatus 106 and that such implementations do not detract from the spirit of the invention. In addition, it will also be appreciated that in certain implementations, the database of target images 110 is shared between multiple apparatuses 106 .
- the output module 108 conveys to a user of the system information derived at least in part on the basis of the detection signal.
- the output module 108 includes an output device 202 and an output controller unit 200 .
- the output device 202 may be any device suitable for conveying information to a user of the system 100 regarding the presence of a prohibited object in the receptacle 104 .
- the information may be conveyed in visual format, audio format or as a combination of visual and audio formats.
- the output device 202 is in communication with the output module 200 and includes a display unit adapted for displaying in visual format information related to the presence of a prohibited object in the receptacle 104 on the basis of a signal received from the output module 200 .
- the output device 202 includes a printer adapted for displaying in printed format information related to the presence of a prohibited object in the receptacle 104 .
- the output device 202 includes an audio output unit adapted for releasing an audio signal conveying information related to the presence of a prohibited object in the receptacle 104 .
- the output device 202 includes a set of visual elements, such as lights or other suitable visual elements, adapted for conveying in visual format information related to the presence of a prohibited object in the receptacle 104 .
- the person skilled in the art will readily appreciate, in light of the present specification, that other suitable types of output devices may be used here without detracting from the spirit of the invention.
- the output controller unit 200 receives the detection signal conveying the presence of the at least one prohibited object in the receptacle 104 from apparatus 106 (shown in FIG. 1 ).
- the detection signal conveys position information related to a certain prohibited object detected in the receptacle 104 and information allowing for the identification of the prohibited object.
- the detection signal also conveys a prohibited object identifier data element.
- the prohibited object identifier data element is associated to an entry in the database of target images 110 .
- the output controller unit 200 includes an apparatus 1510 of the type depicted in FIG. 3 implementing a user interface module.
- the output controller unit 200 is adapted for communicating with an output device 202 including a display screen for causing the latter to display the user interface module generated by the apparatus 1510 .
- the apparatus 1510 implements a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening.
- a specific example of a method implemented by the apparatus 1510 will now be described with reference to FIG. 4 .
- an image signal associated with a receptacle is received, the image signal conveying information related to the contents of the receptacle.
- first information conveying an image associated with the receptacle on the basis of the image signal is caused to be displayed.
- second information is displayed, the second information conveying a presence of at least one prohibited object in the receptacle. The second information is displayed simultaneously with the first information.
- the second information is derived from a detection received from the image processing apparatus 106 and conveying the presence of at least one prohibited object in the receptacle.
- a control is provided for allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the prohibited object whose presence in the receptacle was detected.
- the apparatus 1510 includes a first input 1512 , a second input 1502 , a third input 1504 , a user input 1550 , a processing unit 1506 and an output 1508 .
- the first input 1512 is for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle.
- the image signal is derived from a signal generated by the image generation device 102 (shown in FIG. 1 ).
- the second input 1502 is for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle.
- the detection signal is provided by the image processing apparatus 106 .
- the type of information received at the second input 1502 depends on the specific implementation of the image processing apparatus 106 and may vary from one implementation to the next without detracting from the spirit of the invention. Examples of the type of information that may be received include information on the position of the prohibited object detected, information about the level of confidence of the detection and data allowing identifying the prohibited object detected.
- the third input 1504 is for receiving information associated to the one or more prohibited objects detected in the receptacle from the database of target images 110 .
- the type of information received at the third input 1504 depends on the content of the database of target images 110 and may vary from one implementation to the next. Examples of the type of information that may be received include images depicting the target object and characteristics of the target object. Such characteristics may include, without being limited to, the name of the target object, dimensions of the target object, its associated threat level, the recommended handling procedure when such a target object is detected and any other suitable information.
- the user input 1550 is for receiving signals from a user input device, the signals conveying commands for controlling the information displayed by the user interface module or for annotating the information displayed.
- Any suitable user input device for providing user commands may be used such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen.
- the processing unit 1506 is in communication with the first input 1512 , the second input 1502 , the third input 1504 and the user input 1550 and implements a user interface module for displaying information for use in screening receptacles to detect the presence of one or more prohibited objects.
- the output 1508 is for releasing a signal for causing the output device 202 (shown in FIG. 2 ) to display the graphical user interface module implemented by processing unit 1506 .
- a graphical user interface module implemented by apparatus 1510 in accordance with a specific example of implementation is described in greater detail herein below with reference to figures 5 a , 5 b , 5 c and 5 d.
- FIG. 5 a there is shown a display generated by a graphical user interface module in accordance with a non-limiting implementation on the invention.
- the user interface module displays first information 1604 conveying an image associated with a receptacle on the basis of the image signal received at input 1512 (shown in FIG. 3 ).
- the image associated with the receptacle may be in any suitable format and will depend on the format of the image signal received at input 1512 .
- the image may be of type x-ray, gamma-ray, computed tomography (CT scans), TeraHertz, millimeter wave or emitted radiation amongst others.
- the user interface module also displays second information 1606 conveying a presence of one or more prohibited objects in the receptacle on the basis of the detection signal received at input 1502 (shown in FIG. 3 ).
- the second information 1606 is derived at least in part on the basis of the detection signal received at second input 1502 .
- the second information 1606 is displayed simultaneously with the first information 1604 .
- the second information 1606 conveys position information related to one or more prohibited objects whose presence in the receptacle was detected.
- the second information may convey the presence of one or more prohibited object in the receptacle in textual format in graphical format or as a combination of graphical information and textual information.
- the second information may appear in a dialog box with a message of the form “A ### prohibited object name ### has been detected.”
- the second information 1606 includes visual graphic indicators in the form of circles positioned such as to identify the location of the one or more prohibited objects in the image associated with the receptacle. The location of the circles is derived on the basis of the content of the detection signal received at input 1502 (shown in FIG. 3 ). It will be readily apparent that visual indicators of any suitable shape (e.g. square, arrows, etc . . .
- ⁇ may be used to identify the location of the one or more prohibited objects in the image associated with the receptacle and that the examples depicted in the figures have been presented for the purpose of illustration only. Moreover, functionality may be provided to the user to allow the latter to modify the appearance, such as size, shape and/or color, of the visual indicators used to identify the location of the one or more prohibited objects in the image associated with the receptacle.
- the manner in which such a functionality would be provided is not critical to the present invention and as such will not be described further here.
- the user interface module also provides a control 1608 allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the one or more prohibited objects.
- the control 1608 allows the user to cause the third information to be displayed by using an input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
- the control 1608 is in the form of a selection box including an actuation button that can be selectively actuated by a user.
- the control 1608 is provided as a physical button (or key) on a keyboard or other input device that can be selectively actuated by a user.
- the physical button (or key) is in communication with the apparatus 1510 (shown in FIG. 3 ) through user input 1550 .
- the first information 1604 and the second information 1606 are displayed in a first viewing window 1602 and the third information is displayed in a second viewing window 1630 of the type depicted in figure 5 b .
- the first and second viewing windows 1602 and 1630 may be displayed concurrently on same display screen, concurrently on separate display screens or separately such that when the second viewing window 1630 is displayed the first viewing window 1602 is partially or fully concealed.
- the control 1608 allows a user to cause the second viewing window displaying third information to be displayed.
- FIG. 5 c of the drawings depicts an alternative embodiment of a user interface module where the first and second viewing windows 1602 and 1630 are displayed concurrently.
- a second viewing window 1630 for displaying third information conveying at least one characteristic associated to the one or more prohibited objects detected in the receptacle.
- the type of data conveyed by the third information will vary from one implementation to another.
- the third information conveys an image 1632 and object characteristics 1638 including a description, a risk level and a level of confidence for the detection, each of the above being associated with one of the prohibited objects that was detected.
- Other types of information that may be conveyed include, without being limited to: handling procedure when such a prohibited object is detected, dimensions of the prohibited object or any other characteristics of the prohibited object that could assist the user in validating the information provided, confirm its presence, or facilitate its handling.
- the third information may be conveyed in textual formal or graphical format.
- the third information may include information related to the level of confidence for the detection using a color schema.
- a non-limiting example of a color scheme that may be used in the following:
- the third information may include information related to the level of confidence for the detection using a shape schema.
- the use of a shape based scheme to show information related to the level of threat is particularly useful for individuals who are color blind or for use with monochromatic display screens.
- a non-limiting example of a shape scheme that may be used in the following:
- the processing unit 1506 is adapted to transmit a query signal to the database of prohibited objects 110 (shown in FIG. 1 ), on the basis of information received through input 1502 in the detection signal, in order to obtain certain information elements associated to a detected prohibited object, for example an image, a description, a risk level and a handling procedure amongst others.
- the database of prohibited objects 110 (shown in FIG. 1 ) transmits the requested information with the processing unit 1506 through input 1504 .
- a signal conveying information associated with one of the prohibited objects that was detected can automatically provided to the apparatus 1510 without requiring a query.
- the graphical user interface module displays a prohibited object list 1634 including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected.
- the prohibited object list 1634 is displayed in the second viewing window 1630 however it will be readily apparent that such a list may alternatively be displayed in the first viewing window 1602 or in yet another viewing window distinct from the first viewing window 1602 and the second viewing window 1630 without detracting from the spirit of the invention.
- the prohibited object list 1634 may be displayed in the first viewing window 1602 and may perform the functionality of the control 1608 . More specifically, in this alternative implementation, the control 1608 (shown in FIG.
- 5 a is embodied in the form of a list of prohibited objects including entries associated to prohibited objects detected in the receptacle.
- the user is enabled to select one or more entries from the list of detected prohibited objects.
- third information conveying at least one characteristic associated to the one or more selected prohibited objects is caused to be displayed by the user interface.
- each entry in the list of entries 1634 includes information conveying a level of confidence associated to the presence of the corresponding prohibited object in the receptacle.
- the information conveying a level of confidence is extracted from the detection signal received at input 1502 .
- the graphical user interface module is operative for processing a data element indicative of the level of confidence received in the detection signal in combination with a detection sensitivity level.
- the level of confidence associated to the presence of a prohibited object in the receptacle conveyed by the data element in the detection signal is below the detection sensitivity level
- the second information associated with the prohibited object is omitted from the user interface module.
- the prohibited object is not listed in the list of entries 1634 . As such, only information associated to prohibited objects for which detection levels of confidence exceeds the detection sensitivity level is provided by the user interface.
- each entry in the list of entries 1634 includes information conveying a threat level (not shown in the figures) associated to the corresponding prohibited object in the receptacle.
- the information conveying a threat level is extracted from the signal received from the database of target images 110 received at third input 1504 .
- the threat level information associated to the prohibited object may convey the relative threat level of a prohibited object compared to other prohibited objects in the database of target images 110 . For example, a gun would be given a relatively high threat level while a metallic nail file would be given a relatively low threat level and perhaps a pocket knife would be given a threat level between that of the nail file and the gun.
- functionality is provided to the user for allowing the latter to sort the entries in the list of entries 1634 on the basis of one or more selection criteria.
- criteria may include, without being limited to, the detection levels of confidence and/or the threat level.
- such functionality may be enabled by displaying a control (not shown on the figures) on the user interface in the form of a pull-down menu providing a user with a set of sorting criteria and allowing the user to select the criteria via a user input device.
- the entries in the list of entries 1634 are sorting on the basis of the criteria selected by the user.
- functionality is provided to the user for allowing the latter to add and/or remove one or more entries in the list of entries 1634 .
- Removing an entry may be desirable for example when the screening personnel observes the detection results and decides that the detection was erroneous or, alternatively, that the object detected is not particularly problematic.
- Adding an entry may be desirable for example when the screening personnel observes the presence of a prohibited object on the image displayed which was not detected.
- the user is prompted to enter information conveying a reason why the entry was removed/added from/to the list of entries.
- the information may be entered using any suitable user input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen.
- the graphical user interface module enables a user to select one or more entries from the plurality of entries in the prohibited object list 1634 for which third information is to be displayed in the second viewing window 1630 .
- the user can select one or more entries entry from the prohibited object list 1634 by using an input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
- the user selection is received at user input 1550 .
- information associated with the one or more entries selected in the prohibited object list 1634 is displayed in the second viewing window 1630 .
- the user interface module is adapted for displaying a second control (not shown in the figures) for allowing a user to cause the second information to be removed from the user interface module.
- the user interface module is adapted for displaying additional controls 1636 for allowing a user to modify the configuration of the user interface.
- the user interface module displays a control window of the type depicted in FIG. 5 d, in response to actuation of control button 1680 , allowing a user to select screening options.
- the user is enabled to select between the following screening options:
- the setting threshold sensitivity/confidence level 1660 may only be made available to user having certain privileges (examples screening supervisors or security directors).
- the user interface module may include some type of user identification functionality, such as a login process, to identify the user of the screening system.
- the user interface module upon selection by the user of the setting threshold sensitivity/confidence level 1660 option, may prompt the user to enter a password for allowing the user to modify the detection sensitivity level of the screening system.
- the user interface module is adapted to allow the user to add complementary information to the information being displayed on the user interface.
- the user is enabled to insert markings in the form of text and/or visual indicators in the image displayed on the user interface.
- the markings may be used, for example, to emphasize certain portions of the receptacle.
- the marked-up image may then be transmitted to a third party location, such as a checking station, so that the checking station is alerted to verify the marked portion of the receptacle to locate a prohibited object.
- the user input 1550 receives signals from a user input device, the signals conveying commands for marking the image displayed in the user interface.
- Any suitable user input device for providing user commands may be used such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen.
- a mouse keyboard
- pointing device pointing device
- speech recognition unit touch sensitive screen
- the user interface module is adapted to store a history of the image signals received at first input 1512 conveying information related to the contents of previously screened receptacles.
- the image signals may be stored in association with the corresponding detection signals received at input 1502 and any corresponding user input received at input 1550 .
- the history of prior images may be accessed through a suitable control (not shown in the figures) provided on the user interface.
- the control may be actuated to cause a list for prior images to be displayed to the user.
- the user may then be enabled to select one or more entries in the list of prior images.
- the selection may be effected on the basis of the images themselves or by allowing the user to specify either a time or time period associated to the images in the history of prior images.
- the one or more images from the history of prior images may then be displayed to the user along with information regarding the prohibited objects detected in those images. When multiple images are selected, the selected images may be displayed concurrently with another or may be displayed separately.
- the user interface module is adapted to assign a classification to the receptacle depending upon the detection signal received at second input 1502 and optionally on the basis of information associated with the prohibited objects conveyed by the detection signal.
- the classification criteria may vary from one implementation to the other and may be further conditioned on the basis of external factors such as national security levels.
- the classification may be a two level classification, such as an “ACCEPTED/REJECTED” type of classification or alternatively may be a multi-level classification.
- An example of a multi-level classification is a three level classification where the receptacles are classified as “LOW/MEDIUM/ HIGH RISK”. The classifications may then be associated to respective handling procedures.
- each class is associated to a set of criteria.
- criteria may include, without being limited to: a threshold confidence level associated to the detection process, the level of risk associated with the prohibited object detection and whether a prohibited object was detected. It will be readily apparent to the person skilled in the art in light of the specification that other criteria may be used without detracting from the spirit of the invention.
- the apparatus 200 for implementing a user interface then releases a signal for causing the output device 202 , in the form of a display, to convey the user interface to a user of the system.
- the output controller unit 200 is adapted to cause an audio unit to convey information related to the certain target object in the receptacle 104 .
- the output controller unit 200 generates audio data conveying the presence of the certain target object in the receptacle 104 , the location of the certain target object in the receptacle 104 and the characteristics of the target object.
- the apparatus 106 includes a first input 310 , a second input 314 , an output 312 and a processing unit, generally comprising a pre-processing module 300 , an image comparison module 302 and a detection signal generator module 306 .
- the first input 310 is for receiving an image signal associated with a receptacle from the image generation device 102 (shown in FIG. 1 ).
- the second input 314 is for receiving target images from the database of target images 110 . It will be appreciated that in embodiments where the database of target images 110 is part of apparatus 106 , the second input 314 may be omitted.
- the output 312 is for releasing a detection signal conveying the presence of a target object in the receptacle 104 for transmittal to output module 108 .
- the processing unit of the apparatus 106 receives the image signal associated with the receptacle 104 from the first input 310 and processes that image signal in combination with a plurality of target images associated with target objects received at input 314 to detect a presence of a target object in the receptacle 104 . In response to detection of the presence of at least one target object in the receptacle 104 , the processing unit of the apparatus 106 generates and releases at output 312 a detection signal conveying the presence of the target object in the receptacle 104 .
- the pre-processing module 300 receives an image signal associated with the receptacle 104 is received via the first input 310 .
- the pre-processing module 300 processes the image signal in order to enhance the image, remove extraneous information therefrom and remove noise artefacts in order to obtain more accurate comparison results.
- the complexity of the requisite level of pre-processing and the related tradeoffs between speed and accuracy depend on the application. Examples of pre-processing may include, without being limited to, brightness and contrast manipulation, histogram modification, noise removal and filtering amongst others.
- the pre-processing module 300 may actually be external to the apparatus 106 , e.g., it may be integrated as part of the image generation device 102 or as an external component. It will also be appreciated that the pre-processing module 300 (and hence step 501 ) may be omitted in certain embodiments of the present invention without detracting from the spirit of the invention. As part of step 501 , the pre-processing module 300 releases a modified image signal for processing by the image comparison module 302 .
- the image comparison module 302 verifies whether there remain any unprocessed target images in the database of target images 110 . In the affirmative, the image comparison module 302 proceeds to step 503 where the next target image is accessed and the image comparison module 302 then proceeds to step 504 . If at step 502 all target images in the database of target images 110 have been processed, the image comparison module 302 proceeds to step 508 and the process is completed.
- the image comparison module 302 compares the image signal associated with the receptacle 104 against the target image accessed at step 503 to determine whether a match exists.
- the comparison may be effected using any image processing algorithm suitable for comparing two images. Examples of algorithms that can be used to perform image processing and comparison include without being limited to:
- the image comparison module 302 includes an edge detector to perform part of the comparison at step 504 .
- the comparison performed at step 504 includes effecting a correlation operation between data derived from the image signal and the target images in the database 110 .
- the correlation operation is performed by an optical correlator.
- an optical correlator suitable for use in comparing two images will be described later on in the specification.
- the correlation operation is performed by a digital correlator.
- the image comparison module 302 then proceeds to step 506 where the result of the comparison effected at step 504 is processed to determine whether a match exists between the image signal associated with the receptacle 104 and the target image. In the absence of a match, the image comparison module 302 returns to step 502 . In response to detection of a match, the image comparison module 302 triggers the detection signal generation module 306 to execute step 510 . Then, the image comparison module 302 returns to step 502 to continue processing with respect to the next target image.
- the detection signal generation module 306 generates a detection signal conveying the presence of the target object in the receptacle 104 , and the detection signal is released at output 312 .
- the detection signal may simply convey the fact that a target object has been detected as present in the receptacle 104 , without necessarily specifying the identity of the target object.
- the detection signal may convey the actual identity of the detected target object detected as being present in the receptacle 104 .
- the detection signal may include information related to the positioning of the target object within the receptacle 104 and optionally a target object identifier data element associated to the target object determined to be a potential match.
- step 504 which involves a comparison between the image signal associated with the receptacle 104 and the target images from the database of target images 110 , is performed using a correlation operation.
- the correlation operation multiplies together the Fourier transform of the image signal associated with the receptacle 104 with the Fourier transform complex conjugate of a target image.
- the result of the correlation operation provides a measure of the degree of similarity between the two images.
- the image comparison module 302 includes an optical correlator unit for computing the correlation between the image signal associated with the receptacle 104 and a target image from the database of target images 110 .
- the optical correlator include a joint transform correlator (JTC) and a focal plane correlator (FPC).
- the optical correlator multiplies together the Fourier transform of the image signal associated with the receptacle 104 with the Fourier transform complex conjugate of a target image and records the result with a camera. An energy peak measured with that camera indicates a match between the image signal associated with the receptacle 104 and the target image.
- an optical correlator performs the correlation operation physically through light-based computation, rather than by using software running on a silicon-based computer, which allows computations to be performed at a higher speed than is possible with a software implementation and thus provides for improved real-time performance.
- the correlation computation may also be implemented using a digital correlator.
- the correlation operation is computationally intensive and, in certain implementations requiring real-time performance, the use of a digital correlator may not provide a suitable performance. In such implementations, an optical correlator will be preferred.
- the correlation computation is performed between an images associated with the receptacle 104 and the target images from the database of target images 110 , which includes a plurality of target images associated to objects, which the system 100 is designed to detect.
- the content and format of the database of target images 110 may vary from one implementation to the next.
- the next paragraphs describe manners in which the database 110 can be generated when a correlation computation is used to effect a comparison between an image associated with the receptacle 104 and the target images from the database of target images 110 .
- the skilled person in the art will readily appreciate in fight of the present description that other manners for generating the database 110 may be used without detracting from the spirit of the invention.
- the database of target images 110 includes data indicative of the Fourier transform of the target image.
- This data will herein be referred to as a template or filter.
- the Fourier transform of the target image is digitally pre-computed such as to improve the speed of the correlation operation when the system is in use.
- Image processing and enhancement can be performed on an original image of a target object to obtain better matching performance depending on the environment and application.
- the generation of the reference template or filter is performed in a few steps. First, the background is removed from the target image.
- the target image is extracted from the background and the background is replaced by a black background.
- the resulting image is then processed through a Fourier transform function.
- the result of this transform is a complex image.
- a phase only filter (POF) for example will only contain the complex conjugate of the phase information (between zero and 2 pi) which is mapped to a 0 to 255 range values. These 256 values correspond in fact to the 256 levels of gray of an image.
- phase only filters Phase only filters
- a MACE Minimum Average Correlation Energy
- the MACE filter combines several different 2D projections of a given object and encodes them in a single MACE filter instead of having one 2D projection per filter.
- One of the benefits of using MACE filters is that the resulting database of target images 110 would take less space since it would include fewer items. Also, since the number of correlation operations needed to identify a single target object would be reduced, the total processing time to determine whether a given object is present would also be reduced.
- MACE filters Mahalanobis, A., B. V. K. Vijaya Kumar, and D. Casasent (1987); Minimum average correlation energy filters, Appl. Opt. 26 no. 17, 3633-3640. The contents of this document are incorporated herein by reference.
- Another way of reducing the processing time of the correlation computation is to take advantage of the linear properties of the Fourier transform.
- a composite image can be formed, herein referred to as a mosaic.
- the correlation is computed simultaneously on all the sub-images without incurring any substantial time penalty.
- a mosaic may contain several different target objects or several different orientations of the same target object or a combination of both.
- FIG. 9 of the drawings depicts a mosaic including a target object in various orientations and scales. The parallel processing 5 capabilities of a mosaic effectively increase the throughput of an optical correlator.
- FIG. 10 depicts a high level functional block diagram of a receptacle screening system using an optical correlator as part of the image comparison module 302 .
- an image 800 associated with a receptacle is generated by the image generation device 102 and provided as input to the pre-processing module 300 .
- the pre-processing module 300 performs image acquisition and pre-processing operations and forwards the pre-processed signal to the optical correlator, which is part of the image comparison module 302 .
- the pre-processed image undergoes an optical Fourier transformation 840 .
- the result of the transformation is multiplied 820 by the (previously computed) Fourier transform complex conjugate of a target image 804 obtained from the database of target images 110 .
- the optical correlator then processes the result of the multiplication of the two Fourier transforms by applying another optical Fourier transform 822 .
- the resulting signal is captured by a camera at what is referred to as the correlation plane, which yields the correlation output.
- the correlation output is released for transmission to the detection signal generator 306 where it is analyzed. A peak in the correlation output indicates a match between the image 800 associated with the receptacle 104 and the target image 804 .
- the result of the processing is then conveyed to the user by output module 108 .
- the Fourier transform of the image 800 associated with the receptacle 104 is performed as follows: The image is displayed internally on a small Liquid Crystal Display (LCD). A collimated coherent light beam projects the image through a lens that performs the equivalent of a Fourier transform on the image.
- the multiplication 820 of the Fourier transform of the image 800 by the (previously computed) Fourier transform complex conjugate of a target image 804 is performed by projecting the Fourier transform of the image 800 on a second LCD screen on which is displayed the template or filter associated to the target image 804 .
- the detection signal generator module 306 includes a frame grabber implemented by a digital computer. The digital computer is programmed to detect correlation peaks captured by the CCD.
- FIG. 11 The inner workings of the aforementioned non-limiting example optical correlator are illustrated in FIG. 11 .
- a laser source 900 that generates a coherent light beam used to project images across the correlator.
- the light beam is directed first through a small set of lenses 902 used to expand its diameter in order to illuminate, in parallel, the whole surface of a first LCD screen 904 .
- the image 800 associated with the receptacle 104 is displayed on the first LCD screen 904 either through a direct camera interface or provided as a VGA image by a computing device.
- the first LCD screen 904 is illuminated by the light beam and the image is propagated through the correlator.
- the image 800 captured by the camera is that of a gun on a conveyor belt.
- the light beam modulated by the first image on the first LCD screen 904 is then propagated through a second set of lenses 906 , referred to as a Fourier lens since it performs the equivalent of the Fourier transform mathematical operation.
- the inherent properties of light are used to physically perform the appropriate calculations. Specifically, the propagation of light is a function which corresponds to the kernel of the Fourier transform operation, thus the propagation of light along the axis of a Fourier lens represents a sufficiently strong approximation of this natural phenomenon to assert that the light beam undergoes a Fourier transform. Otherwise stated, a lens has the inherent property of performing a Fourier transform on images observed at its front focal plane, provided that this image is displayed at its back focal plane.
- the Fourier transform which can normally be rather computation-intensive when calculated by a digital computer, is performed in the optical correlator simply by the propagation of the light.
- the mathematics behind this optical realization is equivalent to the exact Fourier transform function and can be modeled with standard fast Fourier algorithms.
- Fourier transforms the reader is invited to consider B. V. K. Vijaya Kumar, Marios Savvides, Krithika Venkataramani,and Chunyan Xie , “Spatial frequency domain image processing for biometric recognition”, Biometrics ICIP Conference 2002 or alternatively J. W. Goodman, Introduction to Fourier Optics, 2nd Edition, McGraw-Hill, 1996. The contents of these documents are incorporated herein by reference.
- the signal is projected on a second LCD screen 908 on which is displayed the target template, i.e., Fourier transform of the target image.
- the target template i.e., Fourier transform of the target image.
- the Fourier transform of the image associated with the receptacle goes through the second LCD screen 908 on which the target template is displayed, the light beam crosses a second Fourier lens 910 which, again, optically computes the equivalent of a Fourier transform multiplication.
- This operation corresponds to a correlation in the spatial domain.
- the target image displayed on the second LCD screen 908 in fact induces a phase variation on the incoming light beam. Each pixel can potentially induce a phase change whose magnitude is equivalent to its grey level.
- the Fourier transform displayed on the first LCD screen 904 is multiplied with the Fourier transform of the target image, which is equivalent to performing a correlation.
- the second Fourier lens 910 finally concentrates the light beam on a small area camera or CCD 912 where the result of the correlation is measured, so to speak.
- the CCD (camera) 912 in fact measures energy peaks on the correlation plane. The position of a correlation peak corresponds in fact to the location of the target object center in the image 800 associated with the receptacle.
- the CCD communicates the signal from the optical correlator to the detection signal generator module 306 .
- the detection signal generator module 306 is a computing unit including a frame grabber and software.
- the software is adapted to processing the signal received from the correlator to detect energy peaks as gray level video signals varying between 0 and 255 .
- a strong intensity peak on the correlation plane indicates a match between the image 800 associated with the receptacle and the target image 804 .
- the location of the energy peak also indicates the location of the center of the target image in the image 800 associated with the receptacle.
- the detection signal generator module 306 generates a detection signal.
- the detection signal may provide, for example, information about the level of the peak(s) and, optionally, the position of the peak(s).
- the detection signal may also include data allowing identifying the target object for which the level of the peak(s) and, optionally, the position of the peak(s) is being provided.
- the Fourier transform as applied to images will now be described in general terms.
- the Fourier transform is a mathematical tool used to convert the information present within an object's image into its frequency representation.
- an image can be seen as a superposition of various spatial frequencies and the Fourier transform is a mathematical operation used to compute the intensity of each of these frequencies within the original image.
- the spatial frequencies represent the rate of variation of image intensity in space. Consequently, a smooth or uniform pattern mainly contains low frequencies. Sharply contoured patterns, by contrast, exhibit a higher frequency content.
- the Fourier transform is a global operator: changing a single frequency of the Fourier transform affects the whole object in the spatial domain.
- ⁇ and ⁇ represent the pixel coordinates in the correlation plane
- C( ⁇ , ⁇ ) stands for the correlation
- x and y identify the pixel coordinates of the input image
- f(x, y) is the original input image
- h *( ⁇ , ⁇ ) is the complex conjugate of the correlation filter.
- the use of optics for computing a correlation operation allows the computation to be performed in a shorter time than by using a digital implementation of the correlation. It turns out that an optical lens properly positioned (i.e. input and output images are located on the lens's focal planes) automatically computes the Fourier transform of the input image.
- the Fourier transform of a target image can be computed beforehand and submitted to the correlator as a mask or template.
- the target template (or filter in short) is generated by computing the Fourier transform of the reference template. This type of filter is called a matched filter.
- FIG. 10 depicts the Fourier transform of the spatial domain image of a ‘ 2 ’. It can be seen that most of the energy (bright areas) is contained in the central portion of the Fourier transform image which correspond to low spatial frequencies (the images are centred on the origin of the Fourier plane). The energy is somewhat more dispersed in the medium frequencies and is concentrated in orientations representative of the shape of the input image. Finally, little energy is contained in the upper frequencies.
- the right-hand-side image shows the phase content of the Fourier transform. The phase is coded from black (0°) to white ( 360°).
- Matched filters are specifically adapted to respond to one image in particular: they are optimized to respond to an object with respect to its energy content.
- the contour of an object corresponds to its high frequency content. This can be easily understood as the contour represent areas where the intensity varies rapidly (hence a high frequency).
- the matched filter can be divided by its module (the image is normalized), over the whole Fourier transform image.
- phase only filters Phase - Only Matched Filtering ”, Joseph L. Homer and Peter D. Gianino, Appl. Opt. Vol. 23 no. 6, 15 Mar. 1994, pp. 812-816. The contents of this document are incorporated herein by reference.
- the discrimination provided by the POF filter has some disadvantages. It turns out that, although the optical correlator is somewhat insensitive to the size of the objects to be recognized, the images are expected to be properly sized, otherwise the features might not be registered properly. To understand this requirement, imagine a filter defined out of a given instance of a ‘2’. If that filter is applied to a second instance of a ‘2’ whose contour is slightly different, the correlation peak will be significantly reduced as a result of the great sensitivity of the filter to the original shape.
- a different type of filter termed a composite filter, was introduced to overcome these limitations. The reader is invited to refer to the following document for additional information regarding this different type of composite filter: H. J. Caufield and W. T. Maloney, Improved discrimination in optical character recognition, Appl. Opt., 8, 2354, 1969. The contents of this document are incorporated herein by reference.
- filters can be designed by:
- composite filters are composed of the response of individual POF filters to the same symbol.
- a filter generated in this fashion is likely to be more robust to minor signature variations as the irrelevant high frequency features will be averaged out.
- the net effect is an equalization of the response of the filter to the different instances of a given symbol.
- Composite filters can also be used to reduce the response of the filter to the other classes of symbols.
- the coefficient b for example, is set to a negative value, then the filter response to a symbol of class b will be significantly reduced.
- the correlation peak will be high if h a ( x,y ) is at the input image, and low if h b ( x,y ) is present at the input.
- OCR Optical character recognition
- a system for screening people includes components similar to those described in connection with the system depicted in FIG. 1 .
- the image generation device 102 is configured to scan a person and possibly to scan the person along various axes and/or views to generate multiple images associated to the person.
- the image or images associated with the person convey information related to the objects carried by the person.
- FIG. 13 depicts two images associated with a person suitable for use in connection with a specific implementation of the system. Each image is then processed in accordance with the method described in the present specification to detect the presence of prohibited objects on the person.
- the database of target objects 110 may further include entries associated to non-prohibited objects and/or objects that do not represent a potential threat.
- Such entries may be used to detect objects commonly carried by people such as cell-phones, watches and rings, for example, which are not prohibited and not threatening.
- identifying such objects and indicating to the screeners that such objects are not prohibited and/or do not represent a potential and as such can be ignored, unnecessary manual verifications can be avoided.
- Certain portions of the apparatus 1510 for implementing a user interface can be implemented on a general purpose digital computer 1300 , of the type depicted in FIG. 14 , including a processing unit 1302 and a memory 1304 connected by a communication bus.
- the memory includes data 1308 and program instructions 1306 .
- the processing unit 1302 is adapted to process the data 1308 and the program instructions 1306 in order to implement the functional blocks described in the specification and depicted in the drawings.
- the digital computer 1300 may also comprise an I/O interface 1310 for receiving or sending data elements to external devices.
- certain portions of the image processing apparatus 106 can also be implemented on a general purpose digital computer having a similar structure as that described in connection with FIG. 14 .
- Certain portions of the image processing apparatus 106 and of the apparatus 1510 for implementing a user interface may be implemented on a same general purpose digital computer without detracting from the spirit of the invention.
- the above-described image processing apparatus 106 can be implemented on a dedicated hardware platform where electrical/optical components implement the functional blocks described in the specification and depicted in the drawings. Specific implementations may be realized using ICs, ASICs, DSPs, FPGA, an optical correlator, digital correlator or other suitable hardware platform.
- image processing apparatus 106 and the apparatus 1510 for implementing a user interface can be implemented as a combination of dedicated hardware and software such as apparatus 1200 of the type depicted in FIG. 15 .
- an implementation comprises an optical correlator 1208 or other dedicated image processing hardware and a general purpose computing unit 1206 including a CPU 1212 and a memory 1214 connected by a communication bus.
- the memory includes data 1218 and program instructions 1216 .
- the CPU 1212 is adapted to process the data 1218 and the program instructions 1216 in order to implement the functional blocks described in the specification and depicted in the drawings.
- the CPU 1212 is also adapted to exchange data with the optical correlator 1208 over communication link 1210 to make use of the optical correlator's image processing capabilities.
- the apparatus 1202 may also comprise I/O interfaces 1202 1204 for receiving or sending data elements to external devices.
- a single optical correlator 1208 can be shared by multiple general purpose computing units 1206 .
- conventional parallel processing techniques can be used for sharing a common hardware resource.
- the optical correlator suitable for use in the system described includes two video inputs.
- the video inputs are suitable for receiving a signal derived from an image generation device and a signal derived from a database of target images.
- the video inputs are suitable for receiving a signal in an NTSC compatible format or a VGA compatible format. It will be appreciated that either one of the video inputs may be adapted for receiving signals of lower or higher resolution than the VGA compatible format signal.
- the video input suitable for receiving a signal in an NTSC compatible format may be adapted for receiving signals in suitable formats such as, but not limited to, PAL and SECAM.
- the optical correlator is adapted to process an image received at the video input having an area of 640 ⁇ 480 pixels.
- the optical correlator's processing capability is independent of the size of the image, as opposed to digital systems that require more processing time and power as images get larger.
- the system for screening a receptacle 100 may also be of a distributed nature where the image signals associated with the receptacles are obtained at one location or more locations and transmitted over a network to a server unit implementing the method described above.
- the server unit may then transmit a signal for causing an output unit to display information to the user.
- the output unit may be located in the same location where the image signal associated with the receptacle was obtained or in the same location as the server unit or in yet another location.
- the output unit is part of a centralized receptacle screening facility.
- FIG. 16 illustrates a network-based client-server system 1300 for system for screening receptacles.
- the client-server system 1300 includes a plurality of client systems 1302 , 1304 , 1306 and 1308 connected to a server system 1310 through network 1312 .
- the communication links 1314 between the client systems 1302 , 1304 , 1306 and 1308 and the server system 1310 can be metallic conductors, optical fibres or wireless, without departing from the spirit of the invention.
- the network 1312 may be any suitable network including but not limited to a global public network such as the Internet, a private network and a wireless network.
- the server 1310 may be adapted to process and issue signals concurrently using suitable methods known in the computer related arts.
- the server system 1310 includes a program element 1316 for execution by a CPU.
- Program element 1316 includes functionality to implement the methods described above, including a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects during security screening of a receptacle, and includes the necessary networking functionality to allow the server system 1310 to communicate with the client systems 1302 , 1304 , 1306 and 1308 over network 1312 .
- the client systems 1302 , 1304 , 1306 and 1308 include display units responsive to signals received from the server system 1310 for displaying user interface module implementation by the server system 1310 .
- server system 1310 also includes an optical correlator unit.
Abstract
An apparatus and method for implementing a user interface module for use in screening a receptacle to detect therein the presence of one or more prohibited objects is provided. An image signal associated with the receptacle conveying information related to the receptacle's contents and a detection signal conveying a presence of at least one prohibited object in the receptacle are received. A user interface module is adapted for displaying first information conveying an image associated with the receptacle on the basis of the image signal. The user interface module is also adapted for displaying, simultaneously with the first information, second information conveying the presence of the prohibited object in the receptacle. The second information is derived at least in part on the basis of the detection signal. As a variant, the user interface module is adapted for providing a control allowing a user to cause third information to be displayed. The third information conveys at least one characteristic associated to the prohibited object. Alternative implementations of the user interface module may be made for use in screening a person to detect thereon the presence of one or more prohibited objects.
Description
- This application is a continuation-in-part application of international PCT patent application serial number PCT/CA2005/000716 filed May 11, 2005 designating the United States.
- The contents of the above referenced applications are incorporated herein by reference.
- The present invention relates generally to security systems and, more particularly, to a user interface for use in screening luggage, mail parcels or cargo containers to identify certain objects located therein or for screening persons to identify certain objects located thereon and to a method and system for implementing such a user interface.
- Security in airports, train stations, ports, mail sorting facilities, office buildings and other public or private venues is becoming increasingly important in particular in light of recent violent events.
- Typically, security-screening systems at airports make use of devices generating penetrating radiation, such as x-ray devices, to scan individual pieces of luggage to generate an image conveying the contents of the luggage. The image is displayed on a screen and is examined by a human operator whose task it is to identify, on the basis of the image, potentially threatening objects located in the luggage.
- A deficiency with current systems is that they are entirely reliant on the human operator to identify potentially threatening objects. However, the performance of the human operator greatly varies according to such factors as poor training and fatigue. As such, the process of detection and identification of threatening objects is highly susceptible to human error.
- Another deficiency with current systems is that the labour costs associated with such systems are significant since human operators must view the images.
- Yet another deficiency is that the images displayed on the x-ray machines provide little, if any, guidance as to what is being observed. It will be appreciated that failure to identify a threatening object, such as a weapon for example, may have serious consequences, such as property damage, injuries and human deaths.
- Consequently, there is a need in the industry for providing a method and system for use in screening luggage items, mail parcels, cargo containers or persons to identify certain objects that alleviate at least in part the deficiencies of the prior art.
- In accordance with a broad aspect, the invention provides a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening. The method comprises receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle. The method also includes displaying first information conveying an image associated with the receptacle on the basis of the image signal. The method also includes displaying second information conveying a presence of at least one prohibited object in the receptacle, the second information being displayed simultaneously with the first information. The method also includes providing a control allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the at least one prohibited object.
- In accordance with a specific implementation, the first information and second information are displayed in a first viewing window and the third information is displayed in a second viewing window. Optionally, the control is adapted for causing the second viewing window to be displayed to a user.
- In accordance with a specific implementation, the control allows the user to cause the third information to be displayed by using any suitable input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen. In addition, the control may be embodied in any suitable form including, for example, a selection box.
- In accordance with a specific implementation, the graphical user interface module displays a prohibited object list, the prohibited object list including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected. Optionally, at least one entry in the prohibited object list conveys a level of confidence associated to detection of the corresponding prohibited object in the receptacle. The graphical user interface module is adapted for enabling a user to select an entry from the plurality of entries in the prohibited object list and to display information associated with the entry selected by the user. The type of information displayed associated with the entry selected can vary from one implementation to the other depending on the specific application. Examples of the type of information that can be displayed include an image of the selected prohibited object detected and a risk levels associated to the prohibited object detected.
- In accordance with a specific implementation, the second information conveys the presence of at least one prohibited object in the receptacle in either textual format or graphical format. Optionally, the second information conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
- In accordance with a specific implementation, the method comprises receiving a detection signal conveying the presence of at least one prohibited object in the receptacle and processing the detection signal to derive the second information. The detection signal may convey a plurality of information items. For example, the detection signal may provide information allowing to identify the type of prohibited object detected, the level of confidence associated to the detection process and the position of the prohibited object in the receptacle amongst others. In the specific example where the detection signal provides information associated level of confidence associated to the detection process, the graphical user interface module may be operative for processing the data element indicative of the level of confidence in combination with a detection sensitivity level. When the level of confidence associated to the presence a prohibited object in the receptacle is below the detection sensitivity level, the second information may be omitted from the user interface module. Optionally, the user interface provides a control allowing a user to modify the detection sensitivity level.
- For the purpose of this specification, the expression “receptacle” is used to broadly describe an entity adapted for receiving objects therein such as, for example, a luggage item, a cargo container or a mail parcel.
- For the purpose of this specification, the expression “luggage item” is used to broadly describe luggage, suitcases, handbags, backpacks, briefcases, boxes, parcels or any other similar type of item suitable for containing objects therein.
- In accordance with another broad aspect, the invention provides and apparatus suitable for implementing a user interface for use in screening a receptacle to detect the presence of one or more prohibited objects in accordance with the above described method.
- In accordance with another broad aspect, the invention provides a computer readable storage medium including a program element suitable for execution by a CPU for implementing a graphical user interface module for displaying information associated to the content of a receptacle in accordance with the above described method.
- In accordance with another broad aspect, the invention provides an apparatus for implementing a user interface module for use in screening receptacles to detect the presence of one or more prohibited objects. The apparatus comprises means for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle. The apparatus also comprises means for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle. The apparatus also comprises means for implementing the user interface module. The user interface module is adapted for displaying first information conveying an image associated with the receptacle on the basis of the image signal. The user interface module is also adapted for displaying second information conveying a presence of at least one prohibited object in the receptacle, the second information being displayed simultaneously with the first information and being derived at least in part on the basis of the detection signal. The user interface module is also adapted for providing a control allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the at least one prohibited object. The apparatus also comprises means for releasing a signal adapted to cause a display module to display the user interface module.
- In accordance with yet another broad aspect, the invention provides a system for detecting the presence of one or more prohibited objects in a receptacle. The system includes an input, an optical correlator and an apparatus for implementing a user interface module. The input is for receiving data conveying graphic information regarding the contents of the receptacle. The optical correlator is in communication with the input and is operative for processing the graphic information to detect a depiction of the one or more prohibited objects in the receptacle. The apparatus implementing the user interface module is in communication with the input and with the optical correlator. The user interface module implemented by the apparatus is operative for displaying first information conveying an image associated with the receptacle on the basis of the data conveying graphic information regarding the contents of the receptacle. The user interface module is also adapted for displaying second information conveying a presence of one or more prohibited objects in the receptacle, the second information being displayed simultaneously with the first information. The user interface module is also operative for releasing a signal adapted to cause a display module to display the user interface module.
- In accordance with another broad aspect, the invention provides a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening. The method comprises receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle and displaying first information conveying an image associated with the receptacle on the basis of the image signal. The method also comprises receiving a detection signal conveying a presence of at least one prohibited object in the receptacle and displaying second information conveying a presence of at least one prohibited object in the receptacle. The second information is displayed simultaneously with the first information and is derived at least in part on the basis of the detection signal.
- In accordance with another broad aspect, the invention provides an apparatus suitable for implementing a user interface for use in screening a receptacle to detect the presence of one or more prohibited objects in accordance with the above described method.
- In accordance with another broad aspect, the invention provides a computer readable storage medium including a program element suitable for execution by a CPU for implementing a graphical user interface module for displaying information associated to the content of a receptacle in accordance with the above described method.
- In accordance with yet another broad aspect, the invention provides an apparatus for implementing a user interface module for use in screening a person to detect the presence of one or more prohibited objects. The apparatus comprising a first input for receiving an image signal associated with the person and a second input for receiving a detection signal conveying a presence of at least one prohibited object on the person. The apparatus also includes a processing unit operative for implementing the user interface module. The user interface module is adapted for displaying first information conveying an image associated with the person on the basis of the image signal and displaying second information conveying a presence of at least one prohibited object on the person. The second information is displayed simultaneously with the first information and is derived at least in part on the basis of the detection signal. The apparatus also includes an output for releasing a signal adapted to cause a display module to display said user interface module.
- Other aspects and features of the present invention will become apparent to those ordinarily skilled in the art upon review of the following description of specific embodiments of the invention in conjunction with the accompanying Figures.
- A detailed description of the embodiments of the present invention is provided herein below, by way of example only, with reference to the accompanying drawings, in which:
-
FIG. 1 is a high-level block diagram of a system for screening a receptacle to detect therein the presence of one or more prohibited objects in accordance with a specific example of implementation of the present invention; -
FIG. 2 is a block diagram of an output module suitable for use in connection with the system depicted inFIG. 1 in accordance with a specific example of implementation of the present invention; -
FIG. 3 is a block diagram of an apparatus suitable for implementing a user interface for use in screening a receptacle in accordance with a specific example of implementation of the present invention; -
FIG. 4 shows a flow diagram depicting a process for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening; -
FIG. 5 a and 5 b depict viewing windows of a user interface module displayed by the output module ofFIG. 2 in accordance with a specific example of implementation of the present invention; -
FIG. 5 c depicts a viewing window of a user interface module displayed by the output module ofFIG. 2 in accordance with an alternative specific example of implementation of the present invention; -
FIG. 5 d depicts a control window of a user interface module displayed by the apparatus ofFIG. 3 allowing a user to select screening options in accordance with a specific example of implementation of the present invention; -
FIG. 6 is a block diagram of an apparatus for processing images suitable for use in connection with the system depicted inFIG. 1 in accordance with a specific example of implementation of the present invention; -
FIG. 7 is a flow diagram depicting a process for detecting a presence of at least one target object in the receptacle in accordance with specific examples of implementation of the present invention; -
FIG. 8 shows three images associated to a target object suitable for use in connection with the system depicted inFIG. 1 , each image depicting the target object in a different orientation, in accordance with a specific example of implementation of the present invention; -
FIG. 9 shows a mosaic image including a plurality of sub-images associated with a target object suitable for use in connection with the system depicted inFIG. 1 , each sub-image depicting the target object in a different orientation and scale, in accordance with a specific example of implementation of the present invention; -
FIG. 10 is a functional block diagram a luggage screening system including an optical correlator in accordance with a specific example of implementation of the present invention; -
FIG. 11 is a block diagram depicting the functioning of an optical correlator in accordance with a specific example of implementation of the present invention; -
FIG. 12 depicts a Fourier transform, amplitude and phase, of the spatial domain image fornumber 2; -
FIG. 13 shows two images associated to a person suitable for use in a system for screening a person to detect the presence of one or more prohibited objects in accordance with a specific example of implementation of the present invention; -
FIG. 14 is a block diagram of an apparatus suitable for implementing a user interface for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in a receptacle during security screening in accordance with a specific example of implementation of the present invention; -
FIG. 15 is a block diagram of an alternative implementation of an apparatus suitable for for use in detecting the presence of one or more prohibited objects in a receptacle during security screening in accordance with a specific example of implementation of the present invention; -
FIG. 16 shows a functional block diagram of a client-server system suitable for use in screening a receptacle to detect therein the presence of one or more prohibited objects in accordance with an alternative specific example of implementation of the present invention. - In the drawings, the embodiments of the invention are illustrated by way of examples. It is to be expressly understood that the description and drawings are only for the purpose of illustration and are an aid for understanding. They are not intended to be a definition of the limits of the invention.
- Shown in
FIG. 1 is asystem 100 for screening a receptacle in accordance with a specific example of implementation of the present invention. It is to be understood that the expression “receptacle”, as used for the purposes of the present description, is used to broadly describe an entity adapted for receiving objects therein such as, for example, a luggage item, a cargo container or a mail parcel. In addition, the expression “luggage item” is used to broadly describe luggage, suitcases, handbags, backpacks, briefcases, boxes, parcels or any other similar type of item suitable for containing objects therein. - As depicted, the
system 100 includes animage generation device 102; anapparatus 106 in communication with theimage generation device 102 and anoutput module 108. - The
image generation device 102 generates an image signal associated with areceptacle 104. The image signal conveys information related to the contents of thereceptacle 104. Theapparatus 106 receives the image signal associated with thereceptacle 104 and processes that image signal in combination with a plurality of target images associated with target objects to detect a presence of at least one target object in thereceptacle 104. In a specific implementation, the plurality of target images is stored in a database oftarget images 110. In response to detection of the presence of at least one target object in thereceptacle 104, theapparatus 106 generates a detection signal conveying the presence of the target object in thereceptacle 104. Examples of the manner in which the detection signal can be derived are described later on in the specification. Theoutput module 108 conveys to a user information derived at least in part on the basis of the detection signal to a user of the system. - Advantageously, the
system 100 provides assistance to the human security personnel using the system in detecting certain target objects, including prohibited objects, and decreases the susceptibility of the screening process to human error. -
Image Generation Device 102 - In a specific example of implementation, the
image generation device 102 uses penetrating radiation or emitted radiation to generate the image associated with thereceptacle 104. Specific examples of such devices include, without being limited to, x-ray, gamma ray, computed tomography (CT scans), thermal imaging, TeraHertz and millimeter wave devices. Such devices are known in the art and as such will not be described further here. In a non-limiting example of implementation, theimage generation device 102 is a conventional x-ray machine adapted for generating an x-ray image of thereceptacle 104. - The image signal generated by the
image generation device 102 and associated with thereceptacle 104 may be conveyed as a two-dimensional (2-D) image or as a three-dimensional (3-D) image and may be in any suitable format. Possible formats include, without being limited to, JPEG, GIF, TIFF and bitmap amongst others. Preferably, the image signal is in a format that can be displayed on a display screen. - Database of
Target Images 110 - In a specific example of implementation, the database of
target images 110 includes a plurality of entries associated to respective target objects that thesystem 100 is designed to detect. - In a non-limiting implementation, for each entry associated to a target object at least one image (hereinafter referred to as a “target image”) is provided in the database of
target images 110. The format of the target images will depend upon the image processing algorithm implemented by theapparatus 106. More specifically, the format of the target images is such that a comparison operation can be performed by theapparatus 106 between the target images and data derived from the image signal associated with thereceptacle 104. - Optionally, for each entry associated to a target object, a set of images is provided in the database of
target images 1 10. For example, images depicting the target object in various orientations may be provided.FIG. 6 of the drawings depicts an example of arbitrary 3D orientations of a target object. - Optionally still, for each entry associated to a target object, characteristics of the target object are provided. Such characteristics may include, without being limited to, the name of the target object, its associated threat level, the recommended handling procedure when such a target object is detected and any other suitable information. In a specific implementation, the threat level information associated to the target object convey the relative threat level of a prohibited object compared to other prohibited objects in the database of
target images 110. For example, a gun would be given a relatively high threat level while a metallic nail file would be given a relatively low level threat level and a pocket knife would be given a threat level between that of the nail file and the gun. Optionally still, each entry in the database oftarget images 110 is also associated to a respective target object identifier data element. In a non-limiting example of implementation, the database oftarget images 110 includes at least one entry associated to a prohibited object, such as a weapon. In the case of luggage screening (in an airport facility for example) the prohibited object typically constitutes a potential threat to the safety of the passenger or aircraft. - In the case of mail parcel screening, the prohibited object is typically an object that is normally not permitted to be sent through the mail, such as guns (in Canada) for example, due to registration requirements/permits and so on.
- In a non-limiting example of implementation, the database of
target images 110 includes one or more entries associated to objects which are not prohibited but which may represent potential threats. For example, the presence of a metal plate or a metal canister in a piece of luggage going through luggage security screening is not prohibited in itself. However such objects may conceal one or more dangerous objects. As such, it is desirable to be able to detect the presence of such objects in receptacle such as to bring them to the attention of the security screeners. - The specific design and content of the database of
target images 110 may vary from one implementation to the next without detracting from the spirit of the invention. The design of the database is not critical to the present invention and as such will not be described further here. - Although the database of
target images 110 has been shown inFIG. 1 to be a component separate from theapparatus 106, it will be appreciated that in certain embodiments the database oftarget images 110 may be part ofapparatus 106 and that such implementations do not detract from the spirit of the invention. In addition, it will also be appreciated that in certain implementations, the database oftarget images 110 is shared betweenmultiple apparatuses 106. -
Output Module 108 - In a specific example of implementation, the
output module 108 conveys to a user of the system information derived at least in part on the basis of the detection signal. - A specific example of implementation of the
output module 108 is shown inFIG. 2 of the drawings. As depicted, the output module includes anoutput device 202 and anoutput controller unit 200. - The
output device 202 may be any device suitable for conveying information to a user of thesystem 100 regarding the presence of a prohibited object in thereceptacle 104. The information may be conveyed in visual format, audio format or as a combination of visual and audio formats. In a first specific example of implementation, theoutput device 202 is in communication with theoutput module 200 and includes a display unit adapted for displaying in visual format information related to the presence of a prohibited object in thereceptacle 104 on the basis of a signal received from theoutput module 200. In a second specific example of implementation, theoutput device 202 includes a printer adapted for displaying in printed format information related to the presence of a prohibited object in thereceptacle 104. In a third specific example of implementation, theoutput device 202 includes an audio output unit adapted for releasing an audio signal conveying information related to the presence of a prohibited object in thereceptacle 104. In a fourth specific example of implementation, theoutput device 202 includes a set of visual elements, such as lights or other suitable visual elements, adapted for conveying in visual format information related to the presence of a prohibited object in thereceptacle 104. The person skilled in the art will readily appreciate, in light of the present specification, that other suitable types of output devices may be used here without detracting from the spirit of the invention. - The
output controller unit 200 receives the detection signal conveying the presence of the at least one prohibited object in thereceptacle 104 from apparatus 106 (shown inFIG. 1 ). In a specific implementation, the detection signal conveys position information related to a certain prohibited object detected in thereceptacle 104 and information allowing for the identification of the prohibited object. Optionally, the detection signal also conveys a prohibited object identifier data element. The prohibited object identifier data element is associated to an entry in the database oftarget images 110. - In a first specific example of implementation, the
output controller unit 200 includes anapparatus 1510 of the type depicted inFIG. 3 implementing a user interface module. In such an implementation, theoutput controller unit 200 is adapted for communicating with anoutput device 202 including a display screen for causing the latter to display the user interface module generated by theapparatus 1510. - The
apparatus 1510 implements a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening. A specific example of a method implemented by theapparatus 1510 will now be described with reference toFIG. 4 . Atstep 1700, an image signal associated with a receptacle is received, the image signal conveying information related to the contents of the receptacle. Atstep 1702, first information conveying an image associated with the receptacle on the basis of the image signal is caused to be displayed. Atstep 1704, second information is displayed, the second information conveying a presence of at least one prohibited object in the receptacle. The second information is displayed simultaneously with the first information. The second information is derived from a detection received from theimage processing apparatus 106 and conveying the presence of at least one prohibited object in the receptacle. Optionally, atstep 1706, a control is provided for allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the prohibited object whose presence in the receptacle was detected. - The
apparatus 1510 includes afirst input 1512, asecond input 1502, athird input 1504, auser input 1550, aprocessing unit 1506 and anoutput 1508. - The
first input 1512 is for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle. In a specific implementation, the image signal is derived from a signal generated by the image generation device 102 (shown inFIG. 1 ). - The
second input 1502 is for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle. In a specific implementation, the detection signal is provided by theimage processing apparatus 106. The type of information received at thesecond input 1502 depends on the specific implementation of theimage processing apparatus 106 and may vary from one implementation to the next without detracting from the spirit of the invention. Examples of the type of information that may be received include information on the position of the prohibited object detected, information about the level of confidence of the detection and data allowing identifying the prohibited object detected. - The
third input 1504 is for receiving information associated to the one or more prohibited objects detected in the receptacle from the database oftarget images 110. The type of information received at thethird input 1504 depends on the content of the database oftarget images 110 and may vary from one implementation to the next. Examples of the type of information that may be received include images depicting the target object and characteristics of the target object. Such characteristics may include, without being limited to, the name of the target object, dimensions of the target object, its associated threat level, the recommended handling procedure when such a target object is detected and any other suitable information. - The
user input 1550 is for receiving signals from a user input device, the signals conveying commands for controlling the information displayed by the user interface module or for annotating the information displayed. Any suitable user input device for providing user commands may be used such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen. - The
processing unit 1506 is in communication with thefirst input 1512, thesecond input 1502, thethird input 1504 and theuser input 1550 and implements a user interface module for displaying information for use in screening receptacles to detect the presence of one or more prohibited objects. - The
output 1508 is for releasing a signal for causing the output device 202 (shown inFIG. 2 ) to display the graphical user interface module implemented byprocessing unit 1506. A graphical user interface module implemented byapparatus 1510 in accordance with a specific example of implementation is described in greater detail herein below with reference to figures 5 a, 5 b, 5 c and 5 d. - With reference to
FIG. 5 a, there is shown a display generated by a graphical user interface module in accordance with a non-limiting implementation on the invention. - As depicted, the user interface module displays
first information 1604 conveying an image associated with a receptacle on the basis of the image signal received at input 1512 (shown inFIG. 3 ). The image associated with the receptacle may be in any suitable format and will depend on the format of the image signal received atinput 1512. For example, the image may be of type x-ray, gamma-ray, computed tomography (CT scans), TeraHertz, millimeter wave or emitted radiation amongst others. - The user interface module also displays
second information 1606 conveying a presence of one or more prohibited objects in the receptacle on the basis of the detection signal received at input 1502 (shown inFIG. 3 ). Thesecond information 1606 is derived at least in part on the basis of the detection signal received atsecond input 1502. Preferably, thesecond information 1606 is displayed simultaneously with thefirst information 1604. In a specific example, thesecond information 1606 conveys position information related to one or more prohibited objects whose presence in the receptacle was detected. The second information may convey the presence of one or more prohibited object in the receptacle in textual format in graphical format or as a combination of graphical information and textual information. In textual format, the second information may appear in a dialog box with a message of the form “A ### prohibited object name ### has been detected.” In the specific implementation depicted inFIG. 5 a, thesecond information 1606 includes visual graphic indicators in the form of circles positioned such as to identify the location of the one or more prohibited objects in the image associated with the receptacle. The location of the circles is derived on the basis of the content of the detection signal received at input 1502 (shown inFIG. 3 ). It will be readily apparent that visual indicators of any suitable shape (e.g. square, arrows, etc . . . ) may be used to identify the location of the one or more prohibited objects in the image associated with the receptacle and that the examples depicted in the figures have been presented for the purpose of illustration only. Moreover, functionality may be provided to the user to allow the latter to modify the appearance, such as size, shape and/or color, of the visual indicators used to identify the location of the one or more prohibited objects in the image associated with the receptacle. The manner in which such a functionality would be provided is not critical to the present invention and as such will not be described further here. - The user interface module also provides a
control 1608 allowing a user to cause third information to be displayed, the third information conveying at least one characteristic associated to the one or more prohibited objects. In a specific implementation, thecontrol 1608 allows the user to cause the third information to be displayed by using an input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen. In the example depicted, thecontrol 1608 is in the form of a selection box including an actuation button that can be selectively actuated by a user. In an alternative embodiment, thecontrol 1608 is provided as a physical button (or key) on a keyboard or other input device that can be selectively actuated by a user. In such an implementation, the physical button (or key) is in communication with the apparatus 1510 (shown inFIG. 3 ) throughuser input 1550. - In a specific example of implementation, the
first information 1604 and thesecond information 1606 are displayed in afirst viewing window 1602 and the third information is displayed in asecond viewing window 1630 of the type depicted infigure 5 b. The first andsecond viewing windows second viewing window 1630 is displayed thefirst viewing window 1602 is partially or fully concealed. In a specific implementation, thecontrol 1608 allows a user to cause the second viewing window displaying third information to be displayed.FIG. 5 c of the drawings depicts an alternative embodiment of a user interface module where the first andsecond viewing windows - With reference to
FIG. 5 b, there is shown asecond viewing window 1630 for displaying third information conveying at least one characteristic associated to the one or more prohibited objects detected in the receptacle. The type of data conveyed by the third information will vary from one implementation to another. - In the specific example depicted in
FIG. 5 b, the third information conveys animage 1632 andobject characteristics 1638 including a description, a risk level and a level of confidence for the detection, each of the above being associated with one of the prohibited objects that was detected. Other types of information that may be conveyed include, without being limited to: handling procedure when such a prohibited object is detected, dimensions of the prohibited object or any other characteristics of the prohibited object that could assist the user in validating the information provided, confirm its presence, or facilitate its handling. The third information may be conveyed in textual formal or graphical format. For example, the third information may include information related to the level of confidence for the detection using a color schema. A non-limiting example of a color scheme that may be used in the following: -
- Red: threat positively detected.
- Yellow: possible threat detected.
- Green: No threat detected.
- In yet another example, the third information may include information related to the level of confidence for the detection using a shape schema. The use of a shape based scheme to show information related to the level of threat is particularly useful for individuals who are color blind or for use with monochromatic display screens. A non-limiting example of a shape scheme that may be used in the following:
-
- Diamond: threat positively detected.
- Triangle: possible threat detected.
- Square: No threat detected.
- In a specific example of implementation, the
processing unit 1506 is adapted to transmit a query signal to the database of prohibited objects 110 (shown inFIG. 1 ), on the basis of information received throughinput 1502 in the detection signal, in order to obtain certain information elements associated to a detected prohibited object, for example an image, a description, a risk level and a handling procedure amongst others. In response to the query signal, the database of prohibited objects 110 (shown inFIG. 1 ) transmits the requested information with theprocessing unit 1506 throughinput 1504. Alternatively, a signal conveying information associated with one of the prohibited objects that was detected can automatically provided to theapparatus 1510 without requiring a query. - In the specific example of implementation depicted in
figure 5 b, the graphical user interface module displays a prohibitedobject list 1634 including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected. In the example depicted, the prohibitedobject list 1634 is displayed in thesecond viewing window 1630 however it will be readily apparent that such a list may alternatively be displayed in thefirst viewing window 1602 or in yet another viewing window distinct from thefirst viewing window 1602 and thesecond viewing window 1630 without detracting from the spirit of the invention. Alternatively still, the prohibitedobject list 1634 may be displayed in thefirst viewing window 1602 and may perform the functionality of thecontrol 1608. More specifically, in this alternative implementation, the control 1608 (shown inFIG. 5 a) is embodied in the form of a list of prohibited objects including entries associated to prohibited objects detected in the receptacle. The user is enabled to select one or more entries from the list of detected prohibited objects. In response to the user's selection, third information conveying at least one characteristic associated to the one or more selected prohibited objects is caused to be displayed by the user interface. - Optionally, each entry in the list of
entries 1634 includes information conveying a level of confidence associated to the presence of the corresponding prohibited object in the receptacle. The information conveying a level of confidence is extracted from the detection signal received atinput 1502. In a specific example of implementation, the graphical user interface module is operative for processing a data element indicative of the level of confidence received in the detection signal in combination with a detection sensitivity level. When the level of confidence associated to the presence of a prohibited object in the receptacle conveyed by the data element in the detection signal is below the detection sensitivity level, the second information associated with the prohibited object is omitted from the user interface module. In addition, the prohibited object is not listed in the list ofentries 1634. As such, only information associated to prohibited objects for which detection levels of confidence exceeds the detection sensitivity level is provided by the user interface. - Optionally, each entry in the list of
entries 1634 includes information conveying a threat level (not shown in the figures) associated to the corresponding prohibited object in the receptacle. The information conveying a threat level is extracted from the signal received from the database oftarget images 110 received atthird input 1504. The threat level information associated to the prohibited object may convey the relative threat level of a prohibited object compared to other prohibited objects in the database oftarget images 110. For example, a gun would be given a relatively high threat level while a metallic nail file would be given a relatively low threat level and perhaps a pocket knife would be given a threat level between that of the nail file and the gun. - Optionally still, functionality is provided to the user for allowing the latter to sort the entries in the list of
entries 1634 on the basis of one or more selection criteria. Such criteria may include, without being limited to, the detection levels of confidence and/or the threat level. In a specific example, such functionality may be enabled by displaying a control (not shown on the figures) on the user interface in the form of a pull-down menu providing a user with a set of sorting criteria and allowing the user to select the criteria via a user input device. In response to the user selection, the entries in the list ofentries 1634 are sorting on the basis of the criteria selected by the user. Other manners for providing such functionality will become readily apparent to the person skilled in the art in light of the present description and as such will not be described further here. - Optionally still, functionality is provided to the user for allowing the latter to add and/or remove one or more entries in the list of
entries 1634. Removing an entry may be desirable for example when the screening personnel observes the detection results and decides that the detection was erroneous or, alternatively, that the object detected is not particularly problematic. Adding an entry may be desirable for example when the screening personnel observes the presence of a prohibited object on the image displayed which was not detected. As a variant, when an entry from the list ofentries 1634 is removed/added, the user is prompted to enter information conveying a reason why the entry was removed/added from/to the list of entries. The information may be entered using any suitable user input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen. - The graphical user interface module enables a user to select one or more entries from the plurality of entries in the prohibited
object list 1634 for which third information is to be displayed in thesecond viewing window 1630. In a specific implementation, the user can select one or more entries entry from the prohibitedobject list 1634 by using an input device such as, for example, a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen. The user selection is received atuser input 1550. In response to receiving a signal conveying the user selection atuser input 1550, information associated with the one or more entries selected in the prohibitedobject list 1634 is displayed in thesecond viewing window 1630. - Optionally, in addition to the
control 1608, the user interface module is adapted for displaying a second control (not shown in the figures) for allowing a user to cause the second information to be removed from the user interface module. - Optionally still, in addition to the
control 1608, the user interface module is adapted for displayingadditional controls 1636 for allowing a user to modify the configuration of the user interface. In accordance with a specific implementation, the user interface module displays a control window of the type depicted inFIG. 5 d, in response to actuation ofcontrol button 1680, allowing a user to select screening options. In the specific example depicted, the user is enabled to select between the following screening options: -
- Generate a report data 1652: this option allows a report to be generated detailing information associated to the screening of the receptacle. In the example depicted, this is done by providing a control in the form of a button that can be toggled between an “ON” state and an “OFF” state. It will be readily apparent that other suitable forms of controls may also be used without detracting from the spirit of the invention. The information generated in the report may include, without being limited to, time of the screening, identification of the security personnel operating the screening system, identification of the receptacle and/or receptacle owner (e.g. passport number in the case of a customs screening), locations information, identification of the prohibited object detected and description of the handling that took place and the results of the handling amongst others. Advantageously, this report allows a tracking of the screening operation.
- Highlight prohibited object 1664: this option allows a user to cause the second information to be removed from or displayed on the user interface module. In the example depicted, this is done by provided a control in the form of a button that can be toggled between an “ON” state and an “OFF” state. It will be readily apparent that other suitable forms of controls may also be used without detracting from the spirit of the invention.
- Display warning window 1666: this option allows a user to cause a visual indicator in the form of a warning window to be removed from or displayed on the user interface module when a prohibited object is detected in a receptacle.
- Set threshold sensitivity/confidence level 1660: this option allows a user to modify the detection sensitivity level of the screening system. In specific implementations, this may be done by providing a control in the form of a text box, sliding ruler (as shown in the figure), selection menu or other suitable type of control allowing the user to select between a range of detection sensitivity levels. It will be readily apparent that other suitable forms of controls may also be used without detracting from the spirit of the invention.
- It will be readily appreciated by the person skilled in the art in light of the present description that other options may be provided to the user and that certain options may be omitted from certain implementations without detracting from the spirit of the invention. As a variant, certain options may be selectively provided to certain users or, alternatively, may require a password to be modified. For example, the setting threshold sensitivity/
confidence level 1660 may only be made available to user having certain privileges (examples screening supervisors or security directors). As such, the user interface module may include some type of user identification functionality, such as a login process, to identify the user of the screening system. Alternatively, the user interface module, upon selection by the user of the setting threshold sensitivity/confidence level 1660 option, may prompt the user to enter a password for allowing the user to modify the detection sensitivity level of the screening system. - Optionally still, the user interface module is adapted to allow the user to add complementary information to the information being displayed on the user interface. In a specific example of implementation, the user is enabled to insert markings in the form of text and/or visual indicators in the image displayed on the user interface. The markings may be used, for example, to emphasize certain portions of the receptacle. The marked-up image may then be transmitted to a third party location, such as a checking station, so that the checking station is alerted to verify the marked portion of the receptacle to locate a prohibited object. In such an implementation, the
user input 1550 receives signals from a user input device, the signals conveying commands for marking the image displayed in the user interface. Any suitable user input device for providing user commands may be used such as, for example, a mouse, keyboard, pointing device, speech recognition unit or touch sensitive screen. The specific manner in which the functionality for marking the image is provided is not critical to the present invention and as such will not be described further here. - Optionally still, the user interface module is adapted to store a history of the image signals received at
first input 1512 conveying information related to the contents of previously screened receptacles. The image signals may be stored in association with the corresponding detection signals received atinput 1502 and any corresponding user input received atinput 1550. The history of prior images may be accessed through a suitable control (not shown in the figures) provided on the user interface. The control may be actuated to cause a list for prior images to be displayed to the user. The user may then be enabled to select one or more entries in the list of prior images. In specific examples of implementation, the selection may be effected on the basis of the images themselves or by allowing the user to specify either a time or time period associated to the images in the history of prior images. In response to a user selection, the one or more images from the history of prior images may then be displayed to the user along with information regarding the prohibited objects detected in those images. When multiple images are selected, the selected images may be displayed concurrently with another or may be displayed separately. - Optionally still, the user interface module is adapted to assign a classification to the receptacle depending upon the detection signal received at
second input 1502 and optionally on the basis of information associated with the prohibited objects conveyed by the detection signal. The classification criteria may vary from one implementation to the other and may be further conditioned on the basis of external factors such as national security levels. The classification may be a two level classification, such as an “ACCEPTED/REJECTED” type of classification or alternatively may be a multi-level classification. An example of a multi-level classification is a three level classification where the receptacles are classified as “LOW/MEDIUM/ HIGH RISK”. The classifications may then be associated to respective handling procedures. For example, receptacles classified as “REJECT” may be automatically assigned to be manually inspected while receptacles classified as “ACCEPTED” may proceed without such an inspection. In a specific example of implementation, each class is associated to a set of criteria. Examples of criteria may include, without being limited to: a threshold confidence level associated to the detection process, the level of risk associated with the prohibited object detection and whether a prohibited object was detected. It will be readily apparent to the person skilled in the art in light of the specification that other criteria may be used without detracting from the spirit of the invention. - The
apparatus 200 for implementing a user interface then releases a signal for causing theoutput device 202, in the form of a display, to convey the user interface to a user of the system. - In a second specific example of implementation, the
output controller unit 200 is adapted to cause an audio unit to convey information related to the certain target object in thereceptacle 104. In a specific non-limiting example of implementation, theoutput controller unit 200 generates audio data conveying the presence of the certain target object in thereceptacle 104, the location of the certain target object in thereceptacle 104 and the characteristics of the target object. -
Apparatus 106 - The
apparatus 106 will now be described in greater detail with reference toFIG. 6 . As depicted, theapparatus 106 includes afirst input 310, asecond input 314, anoutput 312 and a processing unit, generally comprising apre-processing module 300, animage comparison module 302 and a detectionsignal generator module 306. - The
first input 310 is for receiving an image signal associated with a receptacle from the image generation device 102 (shown inFIG. 1 ). - The
second input 314 is for receiving target images from the database oftarget images 110. It will be appreciated that in embodiments where the database oftarget images 110 is part ofapparatus 106, thesecond input 314 may be omitted. - The
output 312 is for releasing a detection signal conveying the presence of a target object in thereceptacle 104 for transmittal tooutput module 108. - The processing unit of the
apparatus 106 receives the image signal associated with thereceptacle 104 from thefirst input 310 and processes that image signal in combination with a plurality of target images associated with target objects received atinput 314 to detect a presence of a target object in thereceptacle 104. In response to detection of the presence of at least one target object in thereceptacle 104, the processing unit of theapparatus 106 generates and releases at output 312 a detection signal conveying the presence of the target object in thereceptacle 104. - The process implemented by the various functional elements of the processing unit of the
apparatus 106 is depicted inFIG. 7 of the drawings. Atstep 500, thepre-processing module 300 receives an image signal associated with thereceptacle 104 is received via thefirst input 310. Atstep 501, thepre-processing module 300 processes the image signal in order to enhance the image, remove extraneous information therefrom and remove noise artefacts in order to obtain more accurate comparison results. The complexity of the requisite level of pre-processing and the related tradeoffs between speed and accuracy depend on the application. Examples of pre-processing may include, without being limited to, brightness and contrast manipulation, histogram modification, noise removal and filtering amongst others. It will be appreciated that all or part of the functionality of thepre-processing module 300 may actually be external to theapparatus 106, e.g., it may be integrated as part of theimage generation device 102 or as an external component. It will also be appreciated that the pre-processing module 300 (and hence step 501 ) may be omitted in certain embodiments of the present invention without detracting from the spirit of the invention. As part ofstep 501, thepre-processing module 300 releases a modified image signal for processing by theimage comparison module 302. - At
step 502, theimage comparison module 302 verifies whether there remain any unprocessed target images in the database oftarget images 110. In the affirmative, theimage comparison module 302 proceeds to step 503 where the next target image is accessed and theimage comparison module 302 then proceeds to step 504. If atstep 502 all target images in the database oftarget images 110 have been processed, theimage comparison module 302 proceeds to step 508 and the process is completed. - At
step 504, theimage comparison module 302 compares the image signal associated with thereceptacle 104 against the target image accessed atstep 503 to determine whether a match exists. The comparison may be effected using any image processing algorithm suitable for comparing two images. Examples of algorithms that can be used to perform image processing and comparison include without being limited to: - A—Image Enhancement
-
- Brightness and contrast manipulation
- Histogram modification
- Noise removal
- Filtering
- B—Image Segmentation
-
- Thresholding
- Binary or multilevel
- Hysteresis based
- Statistics/histogram analysis
- Clustering
- Region growing
- Splitting and merging
- Texture analysis
- Watershed
- Blob labeling
- Thresholding
- C—General Detection
-
- Template matching
- Matched filtering
- Image registration
- Image correlation
- Hough transform
- D—Edge Detection
-
- Gradient
- Laplacian
- E—Morphological Image Processing
-
- Binary
- Grayscale
- F—Frequency Analysis
-
- Fourier Transform
- Wavelets
- G—Shape Analysis and Representations
-
- Geometric attributes (e.g. perimeter, area, euler number, compactness)
- Spatial moments (invariance)
- Fourier descriptors
- B-splines
- Chain codes
- Polygons
- Quad tree decomposition
- H—Feature Representation and Classification
-
- Bayesian classifier
- Principal component analysis
- Binary tree
- Graphs
- Neural networks
- Genetic algorithms
- Markov random fields
- The above algorithms are well known in the field of image processing and as such will not be described further here.
- In a specific example of implementation, the
image comparison module 302 includes an edge detector to perform part of the comparison atstep 504. In another specific example of implementation, the comparison performed atstep 504 includes effecting a correlation operation between data derived from the image signal and the target images in thedatabase 110. In a specific example of implementation, the correlation operation is performed by an optical correlator. A specific example of implementation of an optical correlator suitable for use in comparing two images will be described later on in the specification. In an alternative example of implementation, the correlation operation is performed by a digital correlator. - The
image comparison module 302 then proceeds to step 506 where the result of the comparison effected atstep 504 is processed to determine whether a match exists between the image signal associated with thereceptacle 104 and the target image. In the absence of a match, theimage comparison module 302 returns to step 502. In response to detection of a match, theimage comparison module 302 triggers the detectionsignal generation module 306 to executestep 510. Then, theimage comparison module 302 returns to step 502 to continue processing with respect to the next target image. - At
step 510, the detectionsignal generation module 306 generates a detection signal conveying the presence of the target object in thereceptacle 104, and the detection signal is released atoutput 312. The detection signal may simply convey the fact that a target object has been detected as present in thereceptacle 104, without necessarily specifying the identity of the target object. Alternatively, the detection signal may convey the actual identity of the detected target object detected as being present in thereceptacle 104. As previously indicated, the detection signal may include information related to the positioning of the target object within thereceptacle 104 and optionally a target object identifier data element associated to the target object determined to be a potential match. - Specific Example of
Image Comparison Module 302 Including an Optical Correlator - As mentioned above, in a specific implementation of the
image comparison module 302,step 504, which involves a comparison between the image signal associated with thereceptacle 104 and the target images from the database oftarget images 110, is performed using a correlation operation. The correlation operation multiplies together the Fourier transform of the image signal associated with thereceptacle 104 with the Fourier transform complex conjugate of a target image. The result of the correlation operation provides a measure of the degree of similarity between the two images. - In a specific implementation, the
image comparison module 302 includes an optical correlator unit for computing the correlation between the image signal associated with thereceptacle 104 and a target image from the database oftarget images 110. Specific examples of implementation of the optical correlator include a joint transform correlator (JTC) and a focal plane correlator (FPC). - The optical correlator multiplies together the Fourier transform of the image signal associated with the
receptacle 104 with the Fourier transform complex conjugate of a target image and records the result with a camera. An energy peak measured with that camera indicates a match between the image signal associated with thereceptacle 104 and the target image. - Advantageously, an optical correlator performs the correlation operation physically through light-based computation, rather than by using software running on a silicon-based computer, which allows computations to be performed at a higher speed than is possible with a software implementation and thus provides for improved real-time performance.
- It will be appreciated that the correlation computation may also be implemented using a digital correlator. The correlation operation is computationally intensive and, in certain implementations requiring real-time performance, the use of a digital correlator may not provide a suitable performance. In such implementations, an optical correlator will be preferred.
- As described above, the correlation computation is performed between an images associated with the
receptacle 104 and the target images from the database oftarget images 110, which includes a plurality of target images associated to objects, which thesystem 100 is designed to detect. It will be appreciated that the content and format of the database oftarget images 110 may vary from one implementation to the next. The next paragraphs describe manners in which thedatabase 110 can be generated when a correlation computation is used to effect a comparison between an image associated with thereceptacle 104 and the target images from the database oftarget images 110. The skilled person in the art will readily appreciate in fight of the present description that other manners for generating thedatabase 110 may be used without detracting from the spirit of the invention. - In a specific example of implementation, the database of
target images 110 includes data indicative of the Fourier transform of the target image. This data will herein be referred to as a template or filter. In non-limiting examples of implementation, the Fourier transform of the target image is digitally pre-computed such as to improve the speed of the correlation operation when the system is in use. Image processing and enhancement can be performed on an original image of a target object to obtain better matching performance depending on the environment and application. - In a non-limiting example of implementation, the generation of the reference template or filter is performed in a few steps. First, the background is removed from the target image.
- In other words the target image is extracted from the background and the background is replaced by a black background. The resulting image is then processed through a Fourier transform function. The result of this transform is a complex image. A phase only filter (POF) for example will only contain the complex conjugate of the phase information (between zero and 2 pi) which is mapped to a 0 to 255 range values. These 256 values correspond in fact to the 256 levels of gray of an image. The person skilled in the art, in light of the present specification, will readily appreciate that various types of templates or filters can be generated. Many methods for generating Fourier filters are known in the art and a few such methods will be described later on in the specification. The reader is invited to refer to the following document for additional information regarding phase only filters (POF): “Phase-Only Matched Filtering”, Joseph L. Horner and Peter D. Gianino, Appl. Opt. Vol. 23 no. 6, 15 Mar. 1994, pp. 812-816. The contents of this document are incorporated herein by reference.
- As a variant, in order to reduce the amount of data needed to represent the whole range of 3D orientations that a single target object can take, a MACE (Minimum Average Correlation Energy) filter is used to generate a template or filter for a given target object. Typically, the MACE filter combines several different 2D projections of a given object and encodes them in a single MACE filter instead of having one 2D projection per filter. One of the benefits of using MACE filters is that the resulting database of
target images 110 would take less space since it would include fewer items. Also, since the number of correlation operations needed to identify a single target object would be reduced, the total processing time to determine whether a given object is present would also be reduced. The reader is invited to refer to the following document for additional information regarding MACE filters: Mahalanobis, A., B. V. K. Vijaya Kumar, and D. Casasent (1987); Minimum average correlation energy filters, Appl. Opt. 26 no. 17, 3633-3640. The contents of this document are incorporated herein by reference. - Another way of reducing the processing time of the correlation computation is to take advantage of the linear properties of the Fourier transform. By dividing the target image into several sub-images, a composite image can be formed, herein referred to as a mosaic. When a mosaic is displayed at the input of the correlator, the correlation is computed simultaneously on all the sub-images without incurring any substantial time penalty. A mosaic may contain several different target objects or several different orientations of the same target object or a combination of both.
FIG. 9 of the drawings depicts a mosaic including a target object in various orientations and scales. Theparallel processing 5 capabilities of a mosaic effectively increase the throughput of an optical correlator. The reader is invited to refer to the following document for additional information regarding the use of a mosaic in an optical correlator: Method and apparatus for evaluating a scale factor and a rotation angle in image processing, Alain Bergeron et al., U.S. Pat. No. 6,549,683, Apr. 15, 2003. The contents of this document are incorporated herein by reference. -
FIG. 10 depicts a high level functional block diagram of a receptacle screening system using an optical correlator as part of theimage comparison module 302. As shown, animage 800 associated with a receptacle is generated by theimage generation device 102 and provided as input to thepre-processing module 300. Thepre-processing module 300 performs image acquisition and pre-processing operations and forwards the pre-processed signal to the optical correlator, which is part of theimage comparison module 302. At the optical correlator, the pre-processed image undergoes anoptical Fourier transformation 840. The result of the transformation is multiplied 820 by the (previously computed) Fourier transform complex conjugate of atarget image 804 obtained from the database oftarget images 110. The optical correlator then processes the result of the multiplication of the two Fourier transforms by applying anotheroptical Fourier transform 822. The resulting signal is captured by a camera at what is referred to as the correlation plane, which yields the correlation output. The correlation output is released for transmission to thedetection signal generator 306 where it is analyzed. A peak in the correlation output indicates a match between theimage 800 associated with thereceptacle 104 and thetarget image 804. The result of the processing is then conveyed to the user byoutput module 108. - In a non-limiting example of implementation of an optical correlator, the Fourier transform of the
image 800 associated with thereceptacle 104 is performed as follows: The image is displayed internally on a small Liquid Crystal Display (LCD). A collimated coherent light beam projects the image through a lens that performs the equivalent of a Fourier transform on the image. Themultiplication 820 of the Fourier transform of theimage 800 by the (previously computed) Fourier transform complex conjugate of atarget image 804 is performed by projecting the Fourier transform of theimage 800 on a second LCD screen on which is displayed the template or filter associated to thetarget image 804. The two multiplied Fourier transforms are then processed through a second Fourier lens, which forces the light beam image to a CCD (camera) at the correlation plane. The CCD output is then sent to the detectionsignal generator module 306. In a specific implementation, the detectionsignal generator module 306 includes a frame grabber implemented by a digital computer. The digital computer is programmed to detect correlation peaks captured by the CCD. - The inner workings of the aforementioned non-limiting example optical correlator are illustrated in
FIG. 11 . On the left hand side appears alaser source 900 that generates a coherent light beam used to project images across the correlator. The light beam is directed first through a small set oflenses 902 used to expand its diameter in order to illuminate, in parallel, the whole surface of afirst LCD screen 904. Theimage 800 associated with thereceptacle 104 is displayed on thefirst LCD screen 904 either through a direct camera interface or provided as a VGA image by a computing device. Thefirst LCD screen 904 is illuminated by the light beam and the image is propagated through the correlator. In the illustrated example, theimage 800 captured by the camera is that of a gun on a conveyor belt. - The light beam modulated by the first image on the
first LCD screen 904 is then propagated through a second set oflenses 906, referred to as a Fourier lens since it performs the equivalent of the Fourier transform mathematical operation. The inherent properties of light are used to physically perform the appropriate calculations. Specifically, the propagation of light is a function which corresponds to the kernel of the Fourier transform operation, thus the propagation of light along the axis of a Fourier lens represents a sufficiently strong approximation of this natural phenomenon to assert that the light beam undergoes a Fourier transform. Otherwise stated, a lens has the inherent property of performing a Fourier transform on images observed at its front focal plane, provided that this image is displayed at its back focal plane. The Fourier transform, which can normally be rather computation-intensive when calculated by a digital computer, is performed in the optical correlator simply by the propagation of the light. The mathematics behind this optical realization is equivalent to the exact Fourier transform function and can be modeled with standard fast Fourier algorithms. For more information regarding Fourier transforms, the reader is invited to consider B. V. K. Vijaya Kumar, Marios Savvides, Krithika Venkataramani,and Chunyan Xie , “Spatial frequency domain image processing for biometric recognition”, Biometrics ICIP Conference 2002 or alternatively J. W. Goodman, Introduction to Fourier Optics, 2nd Edition, McGraw-Hill, 1996. The contents of these documents are incorporated herein by reference. - After going through the
Fourier lens 906, the signal is projected on asecond LCD screen 908 on which is displayed the target template, i.e., Fourier transform of the target image. When the Fourier transform of the image associated with the receptacle goes through thesecond LCD screen 908 on which the target template is displayed, the light beam crosses asecond Fourier lens 910 which, again, optically computes the equivalent of a Fourier transform multiplication. This operation corresponds to a correlation in the spatial domain. The target image displayed on thesecond LCD screen 908 in fact induces a phase variation on the incoming light beam. Each pixel can potentially induce a phase change whose magnitude is equivalent to its grey level. As such the Fourier transform displayed on thefirst LCD screen 904 is multiplied with the Fourier transform of the target image, which is equivalent to performing a correlation. - The
second Fourier lens 910 finally concentrates the light beam on a small area camera orCCD 912 where the result of the correlation is measured, so to speak. The CCD (camera) 912 in fact measures energy peaks on the correlation plane. The position of a correlation peak corresponds in fact to the location of the target object center in theimage 800 associated with the receptacle. - Referring back to
FIG. 10 , the CCD (or camera) communicates the signal from the optical correlator to the detectionsignal generator module 306. In this specific implementation, the detectionsignal generator module 306 is a computing unit including a frame grabber and software. The software is adapted to processing the signal received from the correlator to detect energy peaks as gray level video signals varying between 0 and 255. A strong intensity peak on the correlation plane indicates a match between theimage 800 associated with the receptacle and thetarget image 804. The location of the energy peak also indicates the location of the center of the target image in theimage 800 associated with the receptacle. - The detection
signal generator module 306 generates a detection signal. The detection signal may provide, for example, information about the level of the peak(s) and, optionally, the position of the peak(s). The detection signal may also include data allowing identifying the target object for which the level of the peak(s) and, optionally, the position of the peak(s) is being provided. - Fourier Transform and Spatial Frequencies
- The Fourier transform as applied to images will now be described in general terms. The Fourier transform is a mathematical tool used to convert the information present within an object's image into its frequency representation. In short, an image can be seen as a superposition of various spatial frequencies and the Fourier transform is a mathematical operation used to compute the intensity of each of these frequencies within the original image. The spatial frequencies represent the rate of variation of image intensity in space. Consequently, a smooth or uniform pattern mainly contains low frequencies. Sharply contoured patterns, by contrast, exhibit a higher frequency content.
- The Fourier transform of an image f(x,y) is given by:
F(u,v)=∫∫f(x,y)e −j2π(ux+vy) dxdy (1)
where u, v are the coordinates in the frequency domain. Thus, the Fourier transform is a global operator: changing a single frequency of the Fourier transform affects the whole object in the spatial domain. - A correlation operation can be mathematically described by:
where ε and ξ represent the pixel coordinates in the correlation plane, C(ε,ξ) stands for the correlation, x and y identify the pixel coordinates of the input image, f(x, y) is the original input image and h *(ε,ξ) is the complex conjugate of the correlation filter. - In the frequency domain the same expression takes a slightly different form:
C(ε, ξ)=ℑ −1(F(u,v)H*(u,v)) (3)
where ℑ is the Fourier transform operator, u and v are the pixel coordinates in the Fourier plane, F(u,v) is the Fourier transform complex conjugate of the image acquired with the camera f(x,y) and H*(u,v) is the Fourier transform of the filter of the reference template. Thus, the correlation between an input image and a target template is equivalent, in mathematical terms, to the multiplication of their respective Fourier transform, provided that the complex conjugate of the filter is used. Consequently, the correlation can be defined in the spatial domain as the search for a given pattern (template), or in the frequency domain, as filtering operation with a specially designed matched filter. - Advantageously, the use of optics for computing a correlation operation allows the computation to be performed in a shorter time than by using a digital implementation of the correlation. It turns out that an optical lens properly positioned (i.e. input and output images are located on the lens's focal planes) automatically computes the Fourier transform of the input image. In order to speed up the computation of the correlation, the Fourier transform of a target image can be computed beforehand and submitted to the correlator as a mask or template. The target template (or filter in short) is generated by computing the Fourier transform of the reference template. This type of filter is called a matched filter.
-
FIG. 10 depicts the Fourier transform of the spatial domain image of a ‘2’. It can be seen that most of the energy (bright areas) is contained in the central portion of the Fourier transform image which correspond to low spatial frequencies (the images are centred on the origin of the Fourier plane). The energy is somewhat more dispersed in the medium frequencies and is concentrated in orientations representative of the shape of the input image. Finally, little energy is contained in the upper frequencies. The right-hand-side image shows the phase content of the Fourier transform. The phase is coded from black (0°) to white ( 360°). - Generation of Filters from Target Images
- Matched filters, as their name implies, are specifically adapted to respond to one image in particular: they are optimized to respond to an object with respect to its energy content. Generally, the contour of an object corresponds to its high frequency content. This can be easily understood as the contour represent areas where the intensity varies rapidly (hence a high frequency).
- In order to emphasize the contour of an object, the matched filter can be divided by its module (the image is normalized), over the whole Fourier transform image. The resulting filter is called a Phase-Only Filter (POF) and is defined by:
- The reader is invited to refer to the following document for additional information regarding phase only filters (POF): “Phase-Only Matched Filtering”, Joseph L. Homer and Peter D. Gianino, Appl. Opt. Vol. 23 no. 6, 15 Mar. 1994, pp. 812-816. The contents of this document are incorporated herein by reference.
- Because these filters are defined in the frequency domain, normalizing over the whole spectrum of frequencies implies that each of the frequency components is considered with the same weight. In the spatial domain (e.g. usual real-world domain), this means that the emphasis is given to the contours (or edges) of the object. As such, the POF filter provides a higher degree of discrimination, sharper correlation peaks and higher energy efficiency.
- The discrimination provided by the POF filter, however, has some disadvantages. It turns out that, although the optical correlator is somewhat insensitive to the size of the objects to be recognized, the images are expected to be properly sized, otherwise the features might not be registered properly. To understand this requirement, imagine a filter defined out of a given instance of a ‘2’. If that filter is applied to a second instance of a ‘2’ whose contour is slightly different, the correlation peak will be significantly reduced as a result of the great sensitivity of the filter to the original shape. A different type of filter, termed a composite filter, was introduced to overcome these limitations. The reader is invited to refer to the following document for additional information regarding this different type of composite filter: H. J. Caufield and W. T. Maloney, Improved discrimination in optical character recognition, Appl. Opt., 8, 2354, 1969. The contents of this document are incorporated herein by reference.
- In accordance with specific implementations, filters can be designed by:
-
- Appropriately choosing one specific instance (because it represents characteristics which are, on average, common to all symbols of a given class) of a symbol and calculating from that image the filter against which all instances of that class of symbols will be compared; or
- Averaging many instances of a given to create a generic or ‘template’ image from which the filter is calculated. The computed filter is then called a composite filter since it incorporates the properties of many images (note that it is irrelevant whether the images are averaged before or after the Fourier transform operator is applied, provided that in the latter case, the additions are performed taking the Fourier domain phase into account).
- The latter procedure forms the basis for the generation of composite filters. Thus composite filters are composed of the response of individual POF filters to the same symbol. Mathematically, this can be expressed by:
h comp(x,y)=αaha(x,y)+αbhb(x,y)+K+α xhx(x,y) (5) - A filter generated in this fashion is likely to be more robust to minor signature variations as the irrelevant high frequency features will be averaged out. In short, the net effect is an equalization of the response of the filter to the different instances of a given symbol.
- Composite filters can also be used to reduce the response of the filter to the other classes of symbols. In equation (5) above, if the coefficient b, for example, is set to a negative value, then the filter response to a symbol of class b will be significantly reduced. In other words, the correlation peak will be high if ha(x,y) is at the input image, and low if hb(x,y) is present at the input. A typical implementation of composite filters is described in: Optical character recognition (OCR) in uncontrolled environments using optical correlators, Andre Morin, Alain Bergeron, Donald Prevost, and Ernst A. Radloff Proc. SPIE Int. Soc. Opt. Eng. 3715, 346 (1999). The contents of this document are incorporated herein by reference.
- Second Embodiment—Screening of Persons
- Although the above-described screening system was described in connection with screening of receptacles, the concepts described above can also be applied to the screening of people.
- For example, in an alternative embodiment, a system for screening people is provided. The system includes components similar to those described in connection with the system depicted in
FIG. 1 . In a specific example of implementation, theimage generation device 102 is configured to scan a person and possibly to scan the person along various axes and/or views to generate multiple images associated to the person. The image or images associated with the person convey information related to the objects carried by the person.FIG. 13 depicts two images associated with a person suitable for use in connection with a specific implementation of the system. Each image is then processed in accordance with the method described in the present specification to detect the presence of prohibited objects on the person. - Optionally, in the case of a system for screening people, the database of target objects 110 may further include entries associated to non-prohibited objects and/or objects that do not represent a potential threat. Such entries may be used to detect objects commonly carried by people such as cell-phones, watches and rings, for example, which are not prohibited and not threatening. Advantageously, by identifying such objects and indicating to the screeners that such objects are not prohibited and/or do not represent a potential and as such can be ignored, unnecessary manual verifications can be avoided.
- Specific Physical Implementation
- Certain portions of the
apparatus 1510 for implementing a user interface (shown inFIG. 3 ) can be implemented on a general purposedigital computer 1300, of the type depicted inFIG. 14 , including aprocessing unit 1302 and amemory 1304 connected by a communication bus. The memory includesdata 1308 andprogram instructions 1306. Theprocessing unit 1302 is adapted to process thedata 1308 and theprogram instructions 1306 in order to implement the functional blocks described in the specification and depicted in the drawings. Thedigital computer 1300 may also comprise an I/O interface 1310 for receiving or sending data elements to external devices. - Similarly, certain portions of the
image processing apparatus 106 can also be implemented on a general purpose digital computer having a similar structure as that described in connection withFIG. 14 . Certain portions of theimage processing apparatus 106 and of theapparatus 1510 for implementing a user interface may be implemented on a same general purpose digital computer without detracting from the spirit of the invention. - Alternatively, the above-described
image processing apparatus 106 can be implemented on a dedicated hardware platform where electrical/optical components implement the functional blocks described in the specification and depicted in the drawings. Specific implementations may be realized using ICs, ASICs, DSPs, FPGA, an optical correlator, digital correlator or other suitable hardware platform. - Other alternative implementations of the
image processing apparatus 106 and theapparatus 1510 for implementing a user interface can be implemented as a combination of dedicated hardware and software such asapparatus 1200 of the type depicted inFIG. 15 . As shown, such an implementation comprises anoptical correlator 1208 or other dedicated image processing hardware and a generalpurpose computing unit 1206 including aCPU 1212 and amemory 1214 connected by a communication bus. The memory includesdata 1218 andprogram instructions 1216. TheCPU 1212 is adapted to process thedata 1218 and theprogram instructions 1216 in order to implement the functional blocks described in the specification and depicted in the drawings. TheCPU 1212 is also adapted to exchange data with theoptical correlator 1208 overcommunication link 1210 to make use of the optical correlator's image processing capabilities. Theapparatus 1202 may also comprise I/O interfaces 1202 1204 for receiving or sending data elements to external devices. - In a variant, a single
optical correlator 1208 can be shared by multiple generalpurpose computing units 1206. In such a variant, conventional parallel processing techniques can be used for sharing a common hardware resource. - In a specific example of implementation, the optical correlator suitable for use in the system described includes two video inputs. The video inputs are suitable for receiving a signal derived from an image generation device and a signal derived from a database of target images. In a specific implementation, the video inputs are suitable for receiving a signal in an NTSC compatible format or a VGA compatible format. It will be appreciated that either one of the video inputs may be adapted for receiving signals of lower or higher resolution than the VGA compatible format signal. Similarly, it will also be appreciated that the video input suitable for receiving a signal in an NTSC compatible format may be adapted for receiving signals in suitable formats such as, but not limited to, PAL and SECAM. In a non-limiting implementation, the optical correlator is adapted to process an image received at the video input having an area of 640×480 pixels. However, it will be readily apparent that, by providing suitable interfaces, larger or smaller images can be handled since the optical correlator's processing capability is independent of the size of the image, as opposed to digital systems that require more processing time and power as images get larger.
- It will be appreciated that the system for screening a receptacle 100 (depicted in
FIG. 1 ) may also be of a distributed nature where the image signals associated with the receptacles are obtained at one location or more locations and transmitted over a network to a server unit implementing the method described above. The server unit may then transmit a signal for causing an output unit to display information to the user. The output unit may be located in the same location where the image signal associated with the receptacle was obtained or in the same location as the server unit or in yet another location. In a non-limiting implementation, the output unit is part of a centralized receptacle screening facility.FIG. 16 illustrates a network-based client-server system 1300 for system for screening receptacles. The client-server system 1300 includes a plurality ofclient systems server system 1310 throughnetwork 1312. Thecommunication links 1314 between theclient systems server system 1310 can be metallic conductors, optical fibres or wireless, without departing from the spirit of the invention. Thenetwork 1312 may be any suitable network including but not limited to a global public network such as the Internet, a private network and a wireless network. Theserver 1310 may be adapted to process and issue signals concurrently using suitable methods known in the computer related arts. - The
server system 1310 includes aprogram element 1316 for execution by a CPU.Program element 1316 includes functionality to implement the methods described above, including a method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects during security screening of a receptacle, and includes the necessary networking functionality to allow theserver system 1310 to communicate with theclient systems network 1312. In a specific implementation, theclient systems server system 1310 for displaying user interface module implementation by theserver system 1310. Optionally,server system 1310 also includes an optical correlator unit. - Although the present invention has been described in considerable detail with reference to certain preferred embodiments thereof, variations and refinements are possible without departing from the spirit of the invention. Therefore, the scope of the invention should be limited only by the appended claims and their equivalents.
Claims (80)
1. A computer readable storage medium storing a program element suitable for execution by a CPU, said program element implementing a graphical user interface module for use in detecting the presence of one or more prohibited objects in a receptacle, said graphical user interface module being adapted for:
a) displaying first information conveying an image associated with the receptacle, said image conveying information related to the contents of the receptacle;
b) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information;
c) providing a control allowing a user to cause third information to be displayed, said third information conveying at least one characteristic associated to the at least one prohibited object.
2. A computer readable storage medium as defined in claim 1 , wherein said third information conveys an image of the at least one prohibited object, the image being extracted from a database of prohibited objects.
3. A computer readable storage medium as defined in claim 1 , wherein said third information conveys a risk level associated to the at least one prohibited object.
4. A computer readable storage medium as defined in claim 1 , wherein said third information conveys a level of confidence associated to the presence of the at least one prohibited object in the receptacle.
5. A computer readable storage medium as defined in claim 1 , wherein said first information and second information are displayed in a first viewing window and said third information is displayed in a second viewing window.
6. A computer readable storage medium as defined in claim 5 , wherein said graphical user interface module being adapted for:
a) displaying a prohibited object list, the prohibited object list including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected;
b) enabling a user to select an entry from said plurality of entries in said prohibited object list;
c) displaying information associated with the entry selected in b).
7. A computer readable storage medium as defined in claim 6 , wherein at least one entry in said prohibited object list conveys a level of confidence associated to detection of the corresponding prohibited object in the receptacle.
8. A computer readable storage medium as defined in claim 6 , wherein said information associated to the selected entry includes an information item selected from the set consisting of an image of a prohibited object and a risk level associated to the prohibited object.
9. A computer readable storage medium as defined in claim 1 , wherein said control is a first control, said graphical user interface module being adapted for providing a second control allowing a user to cause said second information to be removed from the user interface module.
10. A computer readable storage medium as defined in claim 1 , wherein the image associated with the receptacle conveyed by said first information is selected from the set consisting of x-ray, gamma-ray, computed tomography (CT scans), TeraHertz and millimeter wave images.
11. A computer readable storage medium as defined in claim 1 , wherein the image associated with the receptacle conveyed by said first information includes an image generated on the basis of emitted radiation.
12. A computer readable storage medium as defined in claim 1 , wherein said control allows the user to cause said third information to be displayed by using an input device selected from the set consisting of a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
13. A computer readable storage medium as defined in claim 12 , wherein said control includes a selection box.
14. A computer readable storage medium as defined in claim 1 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in textual format.
15. A computer readable storage medium as defined in claim 1 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in graphical format.
16. A computer readable storage medium as defined in claim 15 , wherein the second information conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
17. A computer readable storage medium as defined in claim 1 , wherein said graphical user interface module is adapted for:
a) receiving a detection signal conveying presence of at least one prohibited object in the receptacle;
b) processing the detection signal to derive said second information.
18. A computer readable storage medium as defined in claim 17 , wherein said detection signal includes a data element conveying a level of confidence associated to the presence of at least one prohibited object in the receptacle.
19. A computer readable storage medium as defined in claim 18 , wherein said graphical user interface module is operative for:
a) processing the data element indicative of the level of confidence in combination with a detection sensitivity level;
b) when the level of confidence associated to the presence of at least one prohibited object in the receptacle conveyed by the data element in the detection signal is below the detection sensitivity level, causing said second information to be omitted from the user interface module.
20. A computer readable storage medium as defined in claim 19 , wherein said control is a first control, said user interface is adapted for displaying a second control allowing a user to modify the detection sensitivity level.
21. A computer readable storage medium as defined in claim 17 , wherein the detection signal conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
22. A computer readable storage medium as defined in claim 17 , wherein the detection signal enables identification of the at least one prohibited object whose presence in the receptacle was detected.
23. A computer readable storage medium as defined in claim 1 , wherein the receptacle is a luggage item.
24. A computer readable storage medium as defined in claim 1 , wherein the receptacle is a cargo container.
25. A computer readable storage medium as defined in claim 1 , wherein the receptacle is a mail parcel.
26. An apparatus for implementing a user interface module for use in screening receptacles to detect the presence of one or more prohibited objects, said apparatus comprising:
a) a first input for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle;
b) a second input for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle;
c) a processing unit in communication with said first input and second input, said processing unit being operative for implementing the user interface module, said user interface module being adapted for:
i) displaying first information conveying an image associated with the receptacle on the basis of said image signal;
ii) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal;
iii) providing a control allowing a user to cause third information to be displayed, said third information conveying at least one characteristic associated to the at least one prohibited object;
d) an output for releasing a signal adapted to cause a display module to display said user interface module.
27. An apparatus as defined in claim 26 , wherein said third information conveys an image of the at least one prohibited object, the image being extracted from a database of prohibited objects.
28. An apparatus as defined in claim 26 , wherein said third information conveys a risk level associated to the at least one prohibited object.
29. An apparatus as defined in claim 26 , wherein said third information conveys a level of confidence associated to the presence of the at least one prohibited object in the receptacle.
30. An apparatus as defined in claim 26 , wherein said first information and second information are displayed in a first viewing window and said third information is displayed in a second viewing window.
31. An apparatus as defined in claim 30 , wherein said graphical user interface module being adapted for:
a) displaying a prohibited object list, the prohibited object list including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected;
b) enabling a user to select an entry from said plurality of entries in said prohibited object list;
c) displaying information associated with the entry selected in b).
32. An apparatus as defined in claim 31 , wherein at least one entry in said prohibited object list conveys a level of confidence associated to detection of the corresponding prohibited object in the receptacle.
33. An apparatus as defined in claim 31 , wherein said information associated to the selected entry includes an information item selected from the set consisting of an image of a prohibited object and a risk level associated to the prohibited object.
34. An apparatus as defined in claim 26 , wherein said control is a first control, said graphical user interface module being adapted for displaying a second control allowing a user to cause said second information to be removed from the user interface module.
35. An apparatus as defined in claim 26 , wherein the image associated with the receptacle conveyed by said first information is selected from the set consisting of x-ray, gamma-ray, computed tomography (CT scans), TeraHertz and millimeter wave images.
36. An apparatus as defined in claim 26 , wherein the image associated with the receptacle conveyed by said first information includes an image generated on the basis of emitted radiation.
37. An apparatus as defined in claim 26 , wherein said control allows the user to cause said third information to be displayed by using an input device selected from the set consisting of a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
38. An apparatus as defined in claim 37 , wherein said control includes a selection box.
39. An apparatus as defined in claim 26 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in textual format.
40. An apparatus as defined in claim 26 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in graphical format.
41. An apparatus as defined in claim 40 , wherein the second information conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
42. An apparatus as defined in claim 26 , wherein said detection signal includes a data element conveying a level of confidence associated to the presence of at least one prohibited object in the receptacle.
43. An apparatus as defined in claim 42 , wherein said graphical user interface module is operative for:
a) processing the data element indicative of the level of confidence in combination with a detection sensitivity level;
b) when the level of confidence associated to the presence of at least one prohibited object in the receptacle conveyed by the data element in the detection signal is below the detection sensitivity level, causing said second information to be omitted from the user interface module.
44. An apparatus as defined in claim 43 , wherein said control is a first control, said user interface is adapted for displaying a second control allowing a user to modify the detection sensitivity level.
45. An apparatus as defined in claim 41 , wherein the detection signal conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
46. An apparatus as defined in claim 41 , wherein the detection signal enables identification of the at least one prohibited object whose presence in the receptacle was detected.
47. An apparatus as defined in claim 26 , wherein the receptacle is a luggage item.
48. An apparatus as defined in claim 26 , wherein the receptacle is a cargo container.
49. An apparatus as defined in claim 26 , wherein the receptacle is a mail parcel.
50. A method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening, said method comprising:
a) receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle;
b) displaying first information conveying an image associated with the receptacle on the basis of said image signal;
c) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information;
d) providing a control allowing a user to cause third information to be displayed, said third information conveying at least one characteristic associated to the at least one prohibited object.
51. A method as defined in claim 50 , wherein said third information conveys an image of the at least one prohibited object, the image being extracted from a database of prohibited objects.
52. A method as defined in claim 50 , wherein said third information conveys a risk level associated to the at least one prohibited object.
53. A method as defined in claim 50 , wherein said third information conveys a level of confidence associated to the presence of the at least one prohibited object in the receptacle.
54. A method as defined in claim 50 , wherein said first information and second information are displayed in a first viewing window and said third information is displayed in a second viewing window.
55. A method as defined in claim 54 , wherein said graphical user interface module being adapted for:
a) displaying a prohibited object list, the prohibited object list including a plurality of entries, each entry being associated to a corresponding prohibited object whose presence in the receptacle was detected;
b) enabling a user to select an entry from said plurality of entries in said prohibited object list;
c) displaying information associated with the entry selected in b).
56. A method as defined in claim 55 , wherein at least one entry in said prohibited object list conveys a level of confidence associated to detection of the corresponding prohibited object in the receptacle.
57. A method as defined in claim 55 , wherein said information associated to the selected entry includes an information item selected from the set consisting of an image of a prohibited object and a risk level associated to the prohibited object.
58. A method as defined in claim 50 , wherein said control is a first control, said graphical user interface module being adapted for displaying a second control allowing a user to cause said second information to be removed from the user interface module.
59. A method as defined in claim 50 , wherein the image associated with the receptacle conveyed by said first information is selected from the set consisting of x-ray, gamma-ray, computed tomography (CT scans), TeraHertz and millimeter wave images.
60. A method as defined in claim 50 , wherein the image associated with the receptacle conveyed by said first information includes an image generated on the basis of emitted radiation.
61. A method as defined in claim 50 , wherein said control allows the user to cause said third information to be displayed by using an input device selected from the set consisting of a mouse, keyboard, pointing device, speech recognition unit and touch sensitive screen.
62. A method as defined in claim 61 , wherein said control includes a selection box.
63. A method as defined in claim 50 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in textual format.
64. A method as defined in claim 50 , wherein the second information conveys the presence of at least one prohibited object in the receptacle in graphical format.
65. A method as defined in claim 64 , wherein the second information conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
66. A method as defined in claim 50 , wherein said method comprises:
a) receiving a detection signal conveying presence of at least one prohibited object in the receptacle;
b) processing the detection signal to derive said second information.
67. A method as defined in claim 66 , wherein said detection signal includes a data element conveying a level of confidence associated to the presence of at least one prohibited object in the receptacle.
68. A method as defined in claim 67 , wherein said graphical user interface module is operative for:
a) processing the data element indicative of the level of confidence in combination with a detection sensitivity level;
b) when the level of confidence associated to the presence of at least one prohibited object in the receptacle conveyed by the data element in the detection signal is below the detection sensitivity level, causing said second information to be omitted from the user interface module.
69. A method as defined in claim 68 , wherein said control is a first control, said user interface is adapted for displaying a second control allowing a user to modify the detection sensitivity level.
70. A method as defined in claim 66 , wherein the detection signal conveys position information related to the at least one prohibited object whose presence in the receptacle was detected.
71. A method as defined in claim 66 , wherein the detection signal enables identification of the at least one prohibited object whose presence in the receptacle was detected.
72. A method as defined in claim 50 , wherein the receptacle is a luggage item.
73. A method as defined in claim 50 , wherein the receptacle is a cargo container.
74. A method as defined in claim 50 , wherein the receptacle is a mail parcel.
75. An apparatus for implementing a user interface module for use in detecting the presence of one or more prohibited objects in a receptacle, said apparatus comprising:
a) means for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle;
b) means for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle;
c) means for implementing the user interface module, said user interface module being adapted for:
i) displaying first information conveying an image associated with the receptacle on the basis of said image signal;
ii) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal;
iii) providing a control allowing a user to cause third information to be displayed, said third information conveying at least one characteristic associated to the at least one prohibited object;
d) means for releasing a signal adapted to cause a display module to display said user interface module.
76. A system for detecting the presence of one or more prohibited objects in a receptacle, comprising:
a) an input for receiving data conveying graphic information regarding the contents of the receptacle;
b) an optical correlator in communication with said input, said optical correlator being operative for processing the graphic information to detect a depiction of the one or more prohibited objects in the receptacle;
c) an apparatus for implementing a user interface module, said apparatus being in communication with said input and with said optical correlator, said apparatus being operative for implementing a user interface module adapted for:
i) displaying first information conveying an image associated with the receptacle on the basis of the data conveying graphic information regarding the contents of the receptacle;
ii) displaying second information conveying a presence of one or more prohibited objects in the receptacle, said second information being displayed simultaneously with said first information;
iii) releasing a signal adapted to cause a display module to display said user interface module.
77. A computer readable storage medium storing a program element suitable for execution by a CPU, said program element implementing a graphical user interface module for use in detecting the presence of one or more prohibited objects in a receptacle, said graphical user interface module being adapted for:
a) displaying first information conveying an image associated with the receptacle, said image conveying information related to the contents of the receptacle;
b) receiving a detection signal conveying a presence of at least one prohibited object in the receptacle;
c) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal.
78. A method for displaying information associated to a receptacle for use in detecting the presence of one or more prohibited objects in the receptacle during security screening, said method comprising:
a) receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle;
b) displaying first information conveying an image associated with the receptacle on the basis of said image signal;
c) receiving a detection signal conveying a presence of at least one prohibited object in the receptacle;
d) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal.
79. An apparatus for implementing a user interface module for use in screening receptacles to detect a presence of one or more prohibited objects, said apparatus comprising:
a) a first input for receiving an image signal associated with a receptacle, the image signal conveying information related to the contents of the receptacle;
b) a second input for receiving a detection signal conveying a presence of at least one prohibited object in the receptacle;
c) a processing unit in communication with said first input and second input, said processing unit being operative for implementing the user interface module, said user interface module being adapted for:
i) displaying first information conveying an image associated with the receptacle on the basis of said image signal;
ii) displaying second information conveying a presence of at least one prohibited object in the receptacle, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal;
d) an output for releasing a signal adapted to cause a display module to display said user interface module.
80. An apparatus for implementing a user interface module for use in screening a person to detect the presence of one or more prohibited objects, said apparatus comprising:
a) a first input for receiving an image signal associated with the person;
b) a second input for receiving a detection signal conveying a presence of at least one prohibited object on the person;
c) a processing unit in communication with said first input and second input, said processing unit being operative for implementing the user interface module, said user interface module being adapted for:
i) displaying first information conveying an image associated with the person on the basis of said image signal;
ii) displaying second information conveying a presence of at least one prohibited object on the person, said second information being displayed simultaneously with said first information, said second information being derived at least in part on the basis of said detection signal;
d) an output for releasing a signal adapted to cause a display module to display said user interface module.
Priority Applications (9)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CA2006/000655 WO2006119609A1 (en) | 2005-05-11 | 2006-04-25 | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same |
CA002608121A CA2608121A1 (en) | 2005-05-11 | 2006-04-25 | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same |
AU2006246250A AU2006246250A1 (en) | 2005-05-11 | 2006-04-25 | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same |
EP06741416A EP1880331A4 (en) | 2005-05-11 | 2006-04-25 | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same |
PCT/CA2006/000751 WO2006119629A1 (en) | 2005-05-11 | 2006-05-11 | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
US11/431,719 US20070041613A1 (en) | 2005-05-11 | 2006-05-11 | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
CA002608124A CA2608124A1 (en) | 2005-05-11 | 2006-05-11 | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
CA002584683A CA2584683A1 (en) | 2006-04-20 | 2007-04-13 | Apparatus, method and system for screening receptacles and persons |
US11/785,116 US20080062262A1 (en) | 2005-05-11 | 2007-04-16 | Apparatus, method and system for screening receptacles and persons |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CA2005/000716 WO2006119603A1 (en) | 2005-05-11 | 2005-05-11 | Method and system for screening luggage items, cargo containers or persons |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CA2005/000716 Continuation-In-Part WO2006119603A1 (en) | 2005-05-11 | 2005-05-11 | Method and system for screening luggage items, cargo containers or persons |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/431,719 Continuation-In-Part US20070041613A1 (en) | 2005-05-11 | 2006-05-11 | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
US11/785,116 Continuation-In-Part US20080062262A1 (en) | 2005-05-11 | 2007-04-16 | Apparatus, method and system for screening receptacles and persons |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070058037A1 true US20070058037A1 (en) | 2007-03-15 |
Family
ID=37396125
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/920,042 Abandoned US20090174554A1 (en) | 2005-05-11 | 2005-05-11 | Method and system for screening luggage items, cargo containers or persons |
US11/268,749 Expired - Fee Related US7734102B2 (en) | 2005-05-11 | 2005-11-08 | Method and system for screening cargo containers |
US11/407,217 Abandoned US20070058037A1 (en) | 2005-05-11 | 2006-04-20 | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/920,042 Abandoned US20090174554A1 (en) | 2005-05-11 | 2005-05-11 | Method and system for screening luggage items, cargo containers or persons |
US11/268,749 Expired - Fee Related US7734102B2 (en) | 2005-05-11 | 2005-11-08 | Method and system for screening cargo containers |
Country Status (4)
Country | Link |
---|---|
US (3) | US20090174554A1 (en) |
EP (1) | EP1886257A1 (en) |
CA (1) | CA2608119A1 (en) |
WO (1) | WO2006119603A1 (en) |
Cited By (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070041613A1 (en) * | 2005-05-11 | 2007-02-22 | Luc Perron | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
US20070041612A1 (en) * | 2005-05-11 | 2007-02-22 | Luc Perron | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality |
WO2008019473A1 (en) * | 2006-08-16 | 2008-02-21 | Optosecurity Inc. | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same |
US20080062262A1 (en) * | 2005-05-11 | 2008-03-13 | Luc Perron | Apparatus, method and system for screening receptacles and persons |
US20080152082A1 (en) * | 2006-08-16 | 2008-06-26 | Michel Bouchard | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same |
WO2008119151A1 (en) * | 2007-03-30 | 2008-10-09 | Optosecurity Inc. | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same |
US20080260199A1 (en) * | 2004-08-13 | 2008-10-23 | Ingenia Technology Limited | Authenticity Verification Methods, Products and Apparatuses |
US20080294900A1 (en) * | 2004-08-13 | 2008-11-27 | Ingenia Technology Limited | Authenticity Verification of Articles Using a Database |
US20090175411A1 (en) * | 2006-07-20 | 2009-07-09 | Dan Gudmundson | Methods and systems for use in security screening, with parallel processing capability |
US20090196396A1 (en) * | 2006-10-02 | 2009-08-06 | Optosecurity Inc. | Tray for assessing the threat status of an article at a security check point |
WO2009114928A1 (en) * | 2008-03-17 | 2009-09-24 | Optosecurity, Inc. | Method and apparatus for assessing characteristics of liquids |
US20100002834A1 (en) * | 2006-09-18 | 2010-01-07 | Optosecurity Inc | Method and apparatus for assessing characteristics of liquids |
EP2165188A2 (en) * | 2007-06-21 | 2010-03-24 | Rapiscan Security Products, Inc. | Systems and methods for improving directed people screening |
US7734102B2 (en) | 2005-05-11 | 2010-06-08 | Optosecurity Inc. | Method and system for screening cargo containers |
US20100166322A1 (en) * | 2008-12-30 | 2010-07-01 | International Business Machines Corporation | Security Screening Image Analysis Simplification Through Object Pattern Identification |
US20100208972A1 (en) * | 2008-09-05 | 2010-08-19 | Optosecurity Inc. | Method and system for performing x-ray inspection of a liquid product at a security checkpoint |
US20100207741A1 (en) * | 2007-10-10 | 2010-08-19 | Optosecurity Inc. | Method, apparatus and system for use in connection with the inspection of liquid merchandise |
US20100316251A1 (en) * | 2005-12-23 | 2010-12-16 | Ingenia Holdings Limited | Optical Authentication |
US20110007870A1 (en) * | 2007-10-01 | 2011-01-13 | Optosecurity Inc. | Method and devices for assessing the threat status of an article at a security check point |
US7899232B2 (en) | 2006-05-11 | 2011-03-01 | Optosecurity Inc. | Method and apparatus for providing threat image projection (TIP) in a luggage screening system, and luggage screening system implementing same |
US20110172972A1 (en) * | 2008-09-15 | 2011-07-14 | Optosecurity Inc. | Method and apparatus for asssessing properties of liquids by using x-rays |
US8078875B2 (en) | 2005-07-27 | 2011-12-13 | Ingenia Holdings Limited | Verification of authenticity |
US20120093367A1 (en) * | 2009-06-15 | 2012-04-19 | Optosecurity Inc. | Method and apparatus for assessing the threat status of luggage |
US8421625B2 (en) | 2004-03-12 | 2013-04-16 | Ingenia Holdings Limited | System and method for article authentication using thumbnail signatures |
US8494210B2 (en) | 2007-03-30 | 2013-07-23 | Optosecurity Inc. | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same |
US8508592B2 (en) | 2009-02-25 | 2013-08-13 | The University Of Memphis Research Foundation | Spatially-selective reflector structures, reflector disks, and systems and methods for use thereof |
US8615475B2 (en) | 2008-12-19 | 2013-12-24 | Ingenia Holdings Limited | Self-calibration |
US8682076B2 (en) | 2008-12-19 | 2014-03-25 | Ingenia Holdings Limited | Signature generation for use in authentication and verification using a non-coherent radiation source |
US8699088B2 (en) | 2004-03-12 | 2014-04-15 | Ingenia Holdings Limited | Methods and apparatuses for creating authenticatable printed articles and subsequently verifying them |
DE102012111201A1 (en) * | 2012-11-21 | 2014-05-22 | Eads Deutschland Gmbh | Sensor system for detecting e.g. weapons hidden in clothing of aircraft passenger in airport, has sensor devices arranged such that monitored area is detected from different viewpoints, and evaluating device evaluating detected data |
US8780345B2 (en) | 2011-04-22 | 2014-07-15 | The University Of Memphis Research Foundation | Spatially-selective disks, submillimeter imaging devices, methods of submillimeter imaging, profiling scanners, spectrometry devices, and methods of spectrometry |
US8831331B2 (en) | 2009-02-10 | 2014-09-09 | Optosecurity Inc. | Method and system for performing X-ray inspection of a product at a security checkpoint using simulation |
US8879791B2 (en) | 2009-07-31 | 2014-11-04 | Optosecurity Inc. | Method, apparatus and system for determining if a piece of luggage contains a liquid product |
US8892556B2 (en) | 2009-11-10 | 2014-11-18 | Ingenia Holdings Limited | Optimisation |
US20150181136A1 (en) * | 2011-03-29 | 2015-06-25 | Thermal Matrix USA, Inc. | Method and system for detecting concealed objects using handheld thermal imager |
US9632206B2 (en) | 2011-09-07 | 2017-04-25 | Rapiscan Systems, Inc. | X-ray inspection system that integrates manifest data with imaging/detection processing |
US9818249B1 (en) | 2002-09-04 | 2017-11-14 | Copilot Ventures Fund Iii Llc | Authentication method and system |
US9891314B2 (en) | 2014-03-07 | 2018-02-13 | Rapiscan Systems, Inc. | Ultra wide band detectors |
US10013750B2 (en) * | 2012-12-27 | 2018-07-03 | Tsinghua University | Object detection methods, display methods and apparatuses |
US10134254B2 (en) | 2014-11-25 | 2018-11-20 | Rapiscan Systems, Inc. | Intelligent security management system |
US10302807B2 (en) | 2016-02-22 | 2019-05-28 | Rapiscan Systems, Inc. | Systems and methods for detecting threats and contraband in cargo |
JP2019111473A (en) * | 2017-12-21 | 2019-07-11 | アンリツインフィビス株式会社 | Article inspection device |
US10949677B2 (en) | 2011-03-29 | 2021-03-16 | Thermal Matrix USA, Inc. | Method and system for detecting concealed objects using handheld thermal imager |
US20210239875A1 (en) * | 2020-01-30 | 2021-08-05 | Hitachi, Ltd. | Alert output timing control apparatus, alert output timing control method, and non-transitory computer readable storage medium |
US11280898B2 (en) | 2014-03-07 | 2022-03-22 | Rapiscan Systems, Inc. | Radar-based baggage and parcel inspection systems |
WO2022062872A1 (en) * | 2020-09-22 | 2022-03-31 | Kyndryl, Inc. | Identifying a distributed threat in a security zone |
Families Citing this family (58)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7963695B2 (en) | 2002-07-23 | 2011-06-21 | Rapiscan Systems, Inc. | Rotatable boom cargo scanning system |
US20080089477A1 (en) * | 2006-05-25 | 2008-04-17 | Eyal Eshed | System and method for security of shipped containers |
JP2008089868A (en) * | 2006-09-29 | 2008-04-17 | Fujifilm Corp | Method and device for acquiring drawing point data and method and device for drawing |
US8081811B2 (en) * | 2007-04-12 | 2011-12-20 | Fujifilm Corporation | Method, apparatus, and program for judging image recognition results, and computer readable medium having the program stored therein |
US7926705B2 (en) * | 2007-04-20 | 2011-04-19 | Morpho Detection, Inc. | Method and system for using a recording device in an inspection system |
US20090003699A1 (en) * | 2007-05-29 | 2009-01-01 | Peter Dugan | User guided object segmentation recognition |
US7692541B1 (en) * | 2007-05-31 | 2010-04-06 | Gianni Arcaini | Method and apparatus for detecting container breach via visual cues |
US20080298546A1 (en) * | 2007-05-31 | 2008-12-04 | General Electric Company | Cargo container inspection method |
US8108055B2 (en) | 2007-12-28 | 2012-01-31 | Larry Wong | Method, system and apparatus for controlling an electrical device |
GB0803644D0 (en) | 2008-02-28 | 2008-04-02 | Rapiscan Security Products Inc | Scanning systems |
US8462206B1 (en) * | 2010-02-25 | 2013-06-11 | Amazon Technologies, Inc. | Image acquisition system |
US8917927B2 (en) * | 2010-08-17 | 2014-12-23 | Telesecurity Sciences, Inc. | Portable backscatter advanced imaging technology scanner with automated target recognition |
US8586955B2 (en) * | 2010-09-22 | 2013-11-19 | Ko Khee Tay | Apparatus and method for attenuating high energy radiation based on detected vehicle type |
DE102011112652A1 (en) * | 2011-05-16 | 2012-11-22 | Eads Deutschland Gmbh | Image analysis for ordnance disposal and security controls |
DK2538394T3 (en) * | 2011-06-24 | 2022-02-14 | Tomra Systems Asa | Method and apparatus for detecting fraud attempts in return vending machines |
KR101305300B1 (en) * | 2012-10-11 | 2013-09-06 | 사단법인 대덕원자력포럼 | Millimeter wave search system and unified search method for freight and passenger using the system |
US9256798B2 (en) * | 2013-01-31 | 2016-02-09 | Aurasma Limited | Document alteration based on native text analysis and OCR |
EP2793164A1 (en) * | 2013-04-19 | 2014-10-22 | Stichting Maastricht Radiation Oncology "Maastro Clinic" | Method and system for determining a phenotype of a neoplasm in a human or animal body |
EP2796902B1 (en) * | 2013-04-23 | 2017-06-14 | Spinner GmbH | Millimeter Wave Scanning Imaging System |
WO2015013359A1 (en) | 2013-07-23 | 2015-01-29 | Rapiscan Systems, Inc. | Methods for improving processing speed for object inspection |
DE102013222098A1 (en) * | 2013-10-30 | 2015-04-30 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Cargo scanning system |
CN104636707B (en) | 2013-11-07 | 2018-03-23 | 同方威视技术股份有限公司 | The method of automatic detection cigarette |
CN104751163B (en) * | 2013-12-27 | 2018-06-19 | 同方威视技术股份有限公司 | The fluoroscopic examination system and method for automatic Classification and Identification are carried out to cargo |
US9558547B2 (en) * | 2014-01-09 | 2017-01-31 | The Boeing Company | System and method for determining whether an apparatus or an assembly process is acceptable |
GB2525228B (en) | 2014-04-16 | 2020-05-06 | Smiths Heimann Sas | Identification or determination of a load based on texture |
WO2016003547A1 (en) | 2014-06-30 | 2016-01-07 | American Science And Engineering, Inc. | Rapidly relocatable modular cargo container scanner |
US20160027303A1 (en) * | 2014-07-22 | 2016-01-28 | Advanced Technology & Research Corp. | Active detection system for low ground clearance vehicles |
CN104112190B (en) * | 2014-07-23 | 2018-07-31 | 清华大学 | Customs is in way supervisory systems and method |
CN105808555B (en) * | 2014-12-30 | 2019-07-26 | 清华大学 | Check the method and system of cargo |
CN105809655B (en) * | 2014-12-30 | 2021-06-29 | 清华大学 | Vehicle inspection method and system |
US10366364B2 (en) * | 2015-04-16 | 2019-07-30 | United Parcel Service Of America, Inc. | Enhanced multi-layer cargo screening system, computer program product, and method of using the same |
US10345479B2 (en) | 2015-09-16 | 2019-07-09 | Rapiscan Systems, Inc. | Portable X-ray scanner |
US9989669B2 (en) * | 2015-10-23 | 2018-06-05 | The United States of America, as Represented by the Secretary of Homeland Security | Intermodal container scanning |
US20170230620A1 (en) * | 2016-02-04 | 2017-08-10 | Panasonic Intellectual Property Management Co., Ltd. | Container use state determining device |
WO2017139481A1 (en) * | 2016-02-11 | 2017-08-17 | American Rebel, Inc. | Firearm detecting luggage |
KR101778503B1 (en) | 2016-03-25 | 2017-09-14 | 한국해양과학기술원 | Method for providing information of cargo inspection, apparatus performing the same and storage media storing the same |
EP3321845A1 (en) * | 2016-11-14 | 2018-05-16 | Nederlandse Organisatie voor toegepast- natuurwetenschappelijk onderzoek TNO | Ship identification system and method |
CN110199373B (en) | 2017-01-31 | 2021-09-28 | 拉皮斯坎系统股份有限公司 | High power X-ray source and method of operation |
FR3064069B1 (en) * | 2017-03-20 | 2019-06-07 | Safran | NON-DESTRUCTIVE CONTROL OF AERONAUTICAL PARTS BY A RADIOGRAPHY SYSTEM |
CN108734183A (en) * | 2017-04-14 | 2018-11-02 | 清华大学 | Inspection method and inspection equipment |
CN108876230A (en) * | 2017-05-12 | 2018-11-23 | 菜鸟智能物流控股有限公司 | Boxing method, boxing scheme generation method, boxing system and server |
US10572963B1 (en) | 2017-07-14 | 2020-02-25 | Synapse Technology Corporation | Detection of items |
US10210631B1 (en) | 2017-08-18 | 2019-02-19 | Synapse Technology Corporation | Generating synthetic image data |
US11144523B2 (en) * | 2017-11-17 | 2021-10-12 | Battelle Memorial Institute | Methods and data structures for efficient cross-referencing of physical-asset spatial identifiers |
KR101969022B1 (en) * | 2017-12-29 | 2019-04-15 | (주)제이엘케이인스펙션 | Image analysis apparatus and method |
US10452959B1 (en) | 2018-07-20 | 2019-10-22 | Synapse Tehnology Corporation | Multi-perspective detection of objects |
JP2020035086A (en) * | 2018-08-28 | 2020-03-05 | 富士ゼロックス株式会社 | Information processing system, information processing apparatus and program |
CN110020586B (en) * | 2018-12-29 | 2021-01-29 | 北京品恩科技股份有限公司 | Personnel security check management system based on big data |
US11847606B2 (en) | 2019-04-30 | 2023-12-19 | Blackberry Limited | System and method for cargo transportation unit tracking and monitoring device verification |
JP7332351B2 (en) * | 2019-06-06 | 2023-08-23 | 株式会社東芝 | Inspection equipment and inspection program |
US11010605B2 (en) | 2019-07-30 | 2021-05-18 | Rapiscan Laboratories, Inc. | Multi-model detection of objects |
WO2021130931A1 (en) * | 2019-12-25 | 2021-07-01 | Nec Corporation | Object detection apparatus, system, and method, data conversion unit, and non-transitory computer readable medium |
US11212902B2 (en) | 2020-02-25 | 2021-12-28 | Rapiscan Systems, Inc. | Multiplexed drive systems and methods for a multi-emitter X-ray source |
CN113643360A (en) * | 2020-05-11 | 2021-11-12 | 同方威视技术股份有限公司 | Target object positioning method, apparatus, device, medium, and program product |
US11193898B1 (en) | 2020-06-01 | 2021-12-07 | American Science And Engineering, Inc. | Systems and methods for controlling image contrast in an X-ray system |
CN113830510B (en) * | 2020-06-23 | 2023-05-23 | 同方威视技术股份有限公司 | Conveying device and inspection system |
US11796489B2 (en) | 2021-02-23 | 2023-10-24 | Rapiscan Systems, Inc. | Systems and methods for eliminating cross-talk signals in one or more scanning systems having multiple X-ray sources |
CN113093308B (en) * | 2021-05-08 | 2024-04-16 | 佳都科技集团股份有限公司 | Correction method, device, equipment and storage medium for X-ray baggage inspection equipment |
Citations (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4573198A (en) * | 1982-12-27 | 1986-02-25 | Litton Systems, Inc. | Optical image processing/pattern recognition system |
US4637056A (en) * | 1983-10-13 | 1987-01-13 | Battelle Development Corporation | Optical correlator using electronic image preprocessing |
US4722096A (en) * | 1985-03-04 | 1988-01-26 | Heimann Gmbh | Apparatus for transradiating objects on a conveyor path |
US4724543A (en) * | 1985-09-10 | 1988-02-09 | Beckman Research Institute, City Of Hope | Method and apparatus for automatic digital image analysis |
US4725733A (en) * | 1983-07-18 | 1988-02-16 | The United States Of America As Represented By The Secretary Of The Navy | Apparatus and method for remotely detecting the presence of chemical warfare nerve agents in an air-released thermal cloud |
US4795253A (en) * | 1987-04-24 | 1989-01-03 | Mobay Corporation | Remote sensing gas analyzer |
US5079698A (en) * | 1989-05-03 | 1992-01-07 | Advanced Light Imaging Technologies Ltd. | Transillumination method apparatus for the diagnosis of breast tumors and other breast lesions by normalization of an electronic image of the breast |
US5091924A (en) * | 1989-08-09 | 1992-02-25 | Heimann Gmbh | Apparatus for the transillumination of articles with a fan-shaped radiation beam |
US5179581A (en) * | 1989-09-13 | 1993-01-12 | American Science And Engineering, Inc. | Automatic threat detection based on illumination by penetrating radiant energy |
US5181234A (en) * | 1990-08-06 | 1993-01-19 | Irt Corporation | X-ray backscatter detection system |
US5283641A (en) * | 1954-12-24 | 1994-02-01 | Lemelson Jerome H | Apparatus and methods for automated analysis |
US5379336A (en) * | 1992-05-21 | 1995-01-03 | Hughes Aircraft Company | Hybridized semiconductor pixel detector arrays for use in digital radiography |
US5379334A (en) * | 1992-07-20 | 1995-01-03 | Heimann Systems Gmbh & Co. Kg | Object testing system |
US5481622A (en) * | 1994-03-01 | 1996-01-02 | Rensselaer Polytechnic Institute | Eye tracking apparatus and method employing grayscale threshold values |
US5481584A (en) * | 1994-11-23 | 1996-01-02 | Tang; Jihong | Device for material separation using nondestructive inspection imaging |
US5483569A (en) * | 1991-10-25 | 1996-01-09 | American Science And Engineering | Inspection system with no intervening belt |
US5485312A (en) * | 1993-09-14 | 1996-01-16 | The United States Of America As Represented By The Secretary Of The Air Force | Optical pattern recognition system and method for verifying the authenticity of a person, product or thing |
US5490218A (en) * | 1990-08-10 | 1996-02-06 | Vivid Technologies, Inc. | Device and method for inspection of baggage and other objects |
US5493444A (en) * | 1994-04-28 | 1996-02-20 | The United States Of America As Represented By The Secretary Of The Air Force | Photorefractive two-beam coupling nonlinear joint transform correlator |
US5595767A (en) * | 1992-03-27 | 1997-01-21 | Universite Joseph Fourier | Three-dimensional energy distribution device |
US5600485A (en) * | 1991-04-23 | 1997-02-04 | Seiko Instruments Inc. | Optical pattern recognition system method of ferroelectric liquid crystal spatial light modulator |
US5600303A (en) * | 1993-01-15 | 1997-02-04 | Technology International Incorporated | Detection of concealed explosives and contraband |
US5600700A (en) * | 1995-09-25 | 1997-02-04 | Vivid Technologies, Inc. | Detecting explosives or other contraband by employing transmitted and scattered X-rays |
US5604634A (en) * | 1993-09-20 | 1997-02-18 | The United States Of America As Represented By The Secretary Of The Air Force | All optical nonlinear joint fourier transform correlator |
US5706816A (en) * | 1995-07-17 | 1998-01-13 | Aloka Co., Ltd. | Image processing apparatus and image processing method for use in the image processing apparatus |
US5862198A (en) * | 1997-09-30 | 1999-01-19 | Siemens Corporate Research, Inc. | Pre-calculated hitlist for reducing run-time processing of an exact cone beam reconstruction algorithm |
US5862258A (en) * | 1996-11-04 | 1999-01-19 | The United States Of America As Represented By The Secretary Of The Army | Method for distinguishing between objects using mace filters |
US5864598A (en) * | 1997-04-21 | 1999-01-26 | General Electric Company | Methods and apparatus for scanning an object in a computed tomography system |
US5866907A (en) * | 1993-10-12 | 1999-02-02 | Biotraces, Inc. | Ultralow background multiple photon detector |
US6011620A (en) * | 1998-04-06 | 2000-01-04 | Northrop Grumman Corporation | Method and apparatus for the automatic inspection of optically transmissive planar objects |
US6018561A (en) * | 1998-07-27 | 2000-01-25 | Siemens Corporate Research, Inc. | Mask boundary correction in a cone beam imaging system using simplified filtered backprojection image reconstruction |
US6018562A (en) * | 1995-11-13 | 2000-01-25 | The United States Of America As Represented By The Secretary Of The Army | Apparatus and method for automatic recognition of concealed objects using multiple energy computed tomography |
US6031890A (en) * | 1993-04-05 | 2000-02-29 | Heimann Systems Gmbh & Co. Kg | Monitoring installation for containers and trucks |
US6175613B1 (en) * | 1997-07-11 | 2001-01-16 | Ge Medical Systems S.A. | Method for processing a sequence of radiological images of an object |
US6175417B1 (en) * | 1998-02-13 | 2001-01-16 | Micron Technology, Inc. | Method and apparatus for detecting defects in the manufacture of an electronic device |
US6185272B1 (en) * | 1999-03-15 | 2001-02-06 | Analogic Corporation | Architecture for CT scanning system |
US6188747B1 (en) * | 1998-01-24 | 2001-02-13 | Heimann Systems Gmbh | X-ray generator |
US6195444B1 (en) * | 1999-01-12 | 2001-02-27 | Analogic Corporation | Apparatus and method for detecting concealed objects in computed tomography data |
US6195413B1 (en) * | 1998-06-12 | 2001-02-27 | Heimann Systems Gmbh | Method and arrangement for detecting X-rays |
US6335742B1 (en) * | 1997-07-24 | 2002-01-01 | Ricoh Company, Ltd. | Apparatus for file management and manipulation using graphical displays and textual descriptions |
US20020001366A1 (en) * | 2000-03-31 | 2002-01-03 | Toshikazu Tamura | Imaging apparatus, imaging method, and storage medium |
US20020015475A1 (en) * | 1998-09-11 | 2002-02-07 | Kazuhiro Matsumoto | A grid holding frame, which provides grid information to-ray image processing apparatus |
US20020016546A1 (en) * | 2000-06-22 | 2002-02-07 | Marino Cerofolini | Method and apparatus for ultrasound imaging, particularly for three-dimensional imaging |
US20020018199A1 (en) * | 1999-11-04 | 2002-02-14 | Martin Blumenfeld | Imaging of biological samples using electronic light detector |
US20020017620A1 (en) * | 2000-08-04 | 2002-02-14 | Nikon Corporation | Surface inspection apparatus |
US20020024016A1 (en) * | 2000-07-28 | 2002-02-28 | Tadao Endo | Photoelectric conversion device, radiation detection apparatus, image processing system, and driving method thereof |
US6502984B2 (en) * | 1997-01-17 | 2003-01-07 | Canon Kabushiki Kaisha | Radiographic apparatus |
US6507025B1 (en) * | 1995-10-23 | 2003-01-14 | Science Applications International Corporation | Density detection using real time discrete photon counting for fast moving targets |
US6507278B1 (en) * | 2000-06-28 | 2003-01-14 | Adt Security Services, Inc. | Ingress/egress control system for airport concourses and other access controlled areas |
US20030012420A1 (en) * | 2001-06-12 | 2003-01-16 | Applied Imaging Corporation | Automated scanning method for pathology samples |
US20030023592A1 (en) * | 2001-07-27 | 2003-01-30 | Rapiscan Security Products (Usa), Inc. | Method and system for certifying operators of x-ray inspection systems |
US20030024315A1 (en) * | 2000-08-31 | 2003-02-06 | Harald Merkel | Device, method and system for measuring the distribution of selected properties in a material |
US20030031289A1 (en) * | 2001-07-18 | 2003-02-13 | Jiang Hsieh | Methods and apparatus for FOV-dependent aliasing artifact reduction |
US20030031291A1 (en) * | 2000-04-18 | 2003-02-13 | Yoshimichi Yamamoto | X-ray apparatus |
US20030036006A1 (en) * | 2001-03-26 | 2003-02-20 | Shipley Company, L.L.C. | Methods for monitoring photoresists |
US6525331B1 (en) * | 1999-12-01 | 2003-02-25 | Nanyang Technological University | Ball grid array (BGA) package on-line non-contact inspection method and system |
US6526120B1 (en) * | 1998-05-06 | 2003-02-25 | Joseph N. Gray | X-ray flow rate measurement system for materials, including agricultural materials and food products |
US20030038945A1 (en) * | 2001-08-17 | 2003-02-27 | Bernward Mahner | Method and apparatus for testing objects |
US20040012853A1 (en) * | 2002-05-13 | 2004-01-22 | Garcia Juan Manuel Bueno | Method and apparatus for imaging using polarimetry and matrix based image reconstruction |
US20040013239A1 (en) * | 2002-03-13 | 2004-01-22 | Breakaway Imaging, Llc | Systems and methods for quasi-simultaneous multi-planar x-ray imaging |
US20040017883A1 (en) * | 2002-07-22 | 2004-01-29 | Tarou Takagi | CT apparatus, CT imaging method and method of providing service using the same |
US20040016271A1 (en) * | 2002-07-23 | 2004-01-29 | Kirti Shah | Portable inspection containers |
US20040017888A1 (en) * | 2002-07-24 | 2004-01-29 | Seppi Edward J. | Radiation scanning of objects for contraband |
US20040017935A1 (en) * | 2002-07-25 | 2004-01-29 | Avinash Gopal B. | Temporal image comparison method |
US20040017882A1 (en) * | 2002-03-06 | 2004-01-29 | National Institute Of Advanced Industrial Science And Technology | Oblique view cone beam CT system |
US20040022425A1 (en) * | 2002-07-31 | 2004-02-05 | Ge Medical Systems Global Technology Company Llc | Temporal image comparison method |
US20040027127A1 (en) * | 2000-08-22 | 2004-02-12 | Mills Randell L | 4 dimensinal magnetic resonance imaging |
US20040037462A1 (en) * | 1998-08-24 | 2004-02-26 | Lewis Meirion F. | Pattern recognition and other inventions |
US6839403B1 (en) * | 2000-07-24 | 2005-01-04 | Rapiscan Security Products (Usa), Inc. | Generation and distribution of annotation overlays of digital X-ray images for security systems |
US6839406B2 (en) * | 1999-11-13 | 2005-01-04 | Smiths Heimann Gmbh | Apparatus and method for detecting items in objects |
US6837422B1 (en) * | 2000-09-01 | 2005-01-04 | Heimann Systems Gmbh | Service unit for an X-ray examining device |
US20050008119A1 (en) * | 2001-04-03 | 2005-01-13 | L-3 Communications Security And Detections Systems | Remote baggage screening system, software and method |
US20050008203A1 (en) * | 2003-07-09 | 2005-01-13 | General Electric Company | System and method for real-time processing and display of digital medical images |
US6843599B2 (en) * | 2002-07-23 | 2005-01-18 | Rapiscan, Inc. | Self-contained, portable inspection system and method |
US20050018812A1 (en) * | 2001-12-11 | 2005-01-27 | Wolfs Peter Bas Anton | X-ray examination apparatus and method |
US20050017181A1 (en) * | 1999-04-16 | 2005-01-27 | The Regents Of The University Of Michigan | Method and system for high-speed, 3D imaging of optically-invisible radiation and detector and array of such detectors for use therein |
US20050025280A1 (en) * | 2002-12-10 | 2005-02-03 | Robert Schulte | Volumetric 3D x-ray imaging system for baggage inspection including the detection of explosives |
US20050025350A1 (en) * | 2003-07-28 | 2005-02-03 | Engelbart Roger W. | Systems and method for identifying foreign objects and debris (FOD) and defects during fabrication of a composite structure |
US20050031069A1 (en) * | 2003-08-07 | 2005-02-10 | General Electric Company | System and method for detecting an object by dynamically adjusting computational load |
US6856272B2 (en) * | 2002-08-28 | 2005-02-15 | Personnel Protection Technoloties Llc | Methods and apparatus for detecting threats in different areas |
US6982643B2 (en) * | 2002-10-08 | 2006-01-03 | Freight Glove Technologies, Llc | Cargo security method and apparatus |
US20060000911A1 (en) * | 2002-05-07 | 2006-01-05 | Amit Stekel | Automatic certification, identification and tracking of remote objects in relative motion |
US20060002504A1 (en) * | 2004-06-30 | 2006-01-05 | Bruno Kristiaan Bernard De Man | System and method for boundary estimation using CT metrology |
US20060009269A1 (en) * | 2004-07-08 | 2006-01-12 | Hoskinson Reed L | Method and apparatus for monitoring characteristics of a flow path having solid components flowing therethrough |
US20060008054A1 (en) * | 2004-07-12 | 2006-01-12 | Hiromu Ohara | Radiographic image capturing system and radiographic image capturing method |
US20060013464A1 (en) * | 2004-05-26 | 2006-01-19 | Guardian Technologies International, Inc. | System and method for identifying objects of interest in image data |
US20060013455A1 (en) * | 2002-12-17 | 2006-01-19 | Qinetiq Limited | Image analysis |
US6990171B2 (en) * | 2003-10-27 | 2006-01-24 | General Electric Company | System and method of determining a user-defined region-of-interest of an imaging subject for x-ray flux management control |
US20060018434A1 (en) * | 2004-07-20 | 2006-01-26 | Jacobs Alan M | Radiography by selective detection of scatter field velocity components |
US20060018517A1 (en) * | 2000-09-15 | 2006-01-26 | Canon Kabushiki Kaisha | Image processing methods and apparatus for detecting human eyes, human face, and other objects in an image |
US20060017605A1 (en) * | 2003-08-12 | 2006-01-26 | Trex Enterprises Corp. | Millimeter wave portal imaging system |
US20060019409A1 (en) * | 2000-10-13 | 2006-01-26 | Chemimage Corporation | Spectroscopic methods for component particle analysis |
US20060034503A1 (en) * | 2004-08-12 | 2006-02-16 | Fuji Photo Film Co., Ltd. | Medical image processing system |
US20060036167A1 (en) * | 2004-07-03 | 2006-02-16 | Shina Systems Ltd. | Vascular image processing |
US20060056584A1 (en) * | 2002-07-23 | 2006-03-16 | Bryan Allman | Self-contained mobile inspection system and method |
US20070003122A1 (en) * | 2005-06-29 | 2007-01-04 | General Electric Company | Method for quantifying an object in a larger structure using a reconstructed image |
US7164750B2 (en) * | 2003-03-26 | 2007-01-16 | Smiths Detection, Inc. | Non-destructive inspection of material in container |
US7183906B2 (en) * | 2004-03-19 | 2007-02-27 | Lockheed Martin Corporation | Threat scanning machine management system |
US7882141B2 (en) * | 2004-02-20 | 2011-02-01 | Fujifilm Corporation | Digital pictorial book sytstem, pictorial book searching method, and machine readable medium storing thereon pictorial book searching method |
Family Cites Families (507)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4338626A (en) | 1963-03-11 | 1982-07-06 | Lemelson Jerome H | Scanning apparatus and method |
DE3019435A1 (en) | 1980-05-21 | 1981-11-26 | SIEMENS AG AAAAA, 1000 Berlin und 8000 München | METHOD FOR PROCESSING ULTRASONIC ECHOSIGNALS FROM OBJECTS REFLECTING BOTH DIRECTLY AS WELL AS NON-DIRECTIVE, IN PARTICULAR FOR ULTRASONIC IMAGE PROCESSING IN THE FIELD OF FABRIC OR TISSUE EXAMINATION |
US4379348A (en) * | 1980-09-23 | 1983-04-05 | North American Philips Corporation | X-Ray security screening system having magnification |
JPS6058504B2 (en) | 1980-11-17 | 1985-12-20 | 株式会社東芝 | reconstruction device |
US4383327A (en) | 1980-12-01 | 1983-05-10 | University Of Utah | Radiographic systems employing multi-linear arrays of electronic radiation detectors |
US4509075A (en) | 1981-06-15 | 1985-04-02 | Oxbridge, Inc. | Automatic optical inspection apparatus |
US4480899A (en) | 1981-10-08 | 1984-11-06 | Xerox Corporation | Two dimensional electro-optic modulator and applications therefor |
US5297222A (en) | 1982-05-04 | 1994-03-22 | Hitachi, Ltd. | Image processing apparatus |
US4470303A (en) | 1982-09-20 | 1984-09-11 | General Electric Company | Quantitative volume backscatter imaging |
EP0113676A3 (en) | 1983-01-08 | 1989-02-08 | Fuji Photo Film Co., Ltd. | Signal processing method in autoradiography |
EP0113677A3 (en) | 1983-01-08 | 1989-02-01 | Fuji Photo Film Co., Ltd. | Signal processing method in autoradiography |
US4612666A (en) | 1984-07-05 | 1986-09-16 | The United States Of America As Represented By The Secretary Of The Navy | Automatic pattern recognition apparatus |
US4653109A (en) * | 1984-07-30 | 1987-03-24 | Lemelson Jerome H | Image analysis system and method |
JPS6145737A (en) | 1984-08-08 | 1986-03-05 | 株式会社東芝 | X-ray examination apparatus |
GB8424084D0 (en) | 1984-09-24 | 1984-10-31 | Sira Ltd | Inspection apparatus |
US4819188A (en) | 1984-11-22 | 1989-04-04 | Kabushiki Kaisha Toshiba | Image processing apparatus |
US4651297A (en) | 1984-11-28 | 1987-03-17 | General Dynamics, Pomona Division | Two-dimensional image correlator |
CN1003542B (en) | 1985-03-04 | 1989-03-08 | 海曼股份公司 | X-ray scanner |
CN85107860A (en) | 1985-04-03 | 1986-10-01 | 海曼股份公司 | The X-ray scanner |
DE3650496T2 (en) * | 1985-04-17 | 1996-10-31 | Hitachi Ltd | Image processing device |
US4651957A (en) * | 1985-06-20 | 1987-03-24 | The United States Of America As Represented By The Secretary Of The Air Force | Large capacity tow/release apparatus |
DE3530955A1 (en) | 1985-08-29 | 1987-03-05 | Heimann Gmbh | LUGGAGE TEST SYSTEM |
DE3530938A1 (en) | 1985-08-29 | 1987-03-12 | Heimann Gmbh | LUGGAGE TEST SYSTEM |
DE3544182A1 (en) | 1985-12-13 | 1987-06-19 | Heimann Gmbh | CONTACT IMAGE SENSOR LINE |
EP0247491B1 (en) | 1986-05-28 | 1990-08-16 | Heimann GmbH | X-ray scanning system |
US4901084A (en) | 1988-04-19 | 1990-02-13 | Millitech Corporation | Object detection and location system |
US4916722A (en) | 1986-06-26 | 1990-04-10 | Kabushiki Kaisha Toshiba | X-ray image processing apparatus |
EP0253060B1 (en) | 1986-07-14 | 1990-05-30 | Heimann GmbH | X-ray scanning system |
DE8703674U1 (en) | 1987-03-11 | 1988-07-14 | Heimann Gmbh, 6200 Wiesbaden, De | |
US5063602A (en) | 1987-04-14 | 1991-11-05 | Nippon Sheet Glass Co., Ltd. | Image correlation calculation apparatus |
GB8713819D0 (en) | 1987-06-12 | 1987-12-16 | Smiths Industries Plc | Information processing systems |
US4955060A (en) | 1987-07-02 | 1990-09-04 | Nippon Sheet Glass Co., Ltd. | Image recognition apparatus |
US4849912A (en) | 1987-08-03 | 1989-07-18 | Vexcel Corporation | Stereo mensuration method |
US4838644A (en) | 1987-09-15 | 1989-06-13 | The United States Of America As Represented By The United States Department Of Energy | Position, rotation, and intensity invariant recognizing method |
DE8717508U1 (en) | 1987-10-19 | 1989-01-05 | Heimann Gmbh, 6200 Wiesbaden, De | |
US4832447A (en) | 1987-12-04 | 1989-05-23 | Board Of Trustees Operating Michigan State University | Joint transform image correlation using a nonlinear spatial light modulator at the fourier plane |
US4887899A (en) | 1987-12-07 | 1989-12-19 | Hung Yau Y | Apparatus and method for electronic analysis of test objects |
JP2633607B2 (en) | 1988-03-08 | 1997-07-23 | 株式会社東芝 | X-ray CT scanner |
US4869574A (en) | 1988-05-13 | 1989-09-26 | The United States Of America As Represented By The Secretary Of The Army | Hybrid optical correlator |
US5109276A (en) | 1988-05-27 | 1992-04-28 | The University Of Connecticut | Multi-dimensional multi-spectral imaging system |
IL87570A (en) | 1988-08-26 | 1991-12-12 | Israel Atomic Energy Comm | Method and apparatus for the detection and imaging of heavy metals |
JPH0830830B2 (en) | 1988-09-07 | 1996-03-27 | セイコー電子工業株式会社 | Optical correlation processor |
US5020111A (en) | 1988-10-14 | 1991-05-28 | The United States Of America As Represented By The Secretary Of The Army | Spatial symmetry cueing image processing method and apparatus |
GB8826550D0 (en) | 1988-11-14 | 1989-05-17 | Smiths Industries Plc | Image processing apparatus and methods |
US5353799A (en) | 1991-01-22 | 1994-10-11 | Non Invasive Technology, Inc. | Examination of subjects using photon migration with high directionality techniques |
ATE152339T1 (en) | 1989-02-23 | 1997-05-15 | Teijin Ltd | METHOD FOR MEASURING A BONE SHAPE, DEVICE THEREFOR AND SYSTEM FOR BONE EVALUATION |
US5132998A (en) | 1989-03-03 | 1992-07-21 | Matsushita Electric Industrial Co., Ltd. | Radiographic image processing method and photographic imaging apparatus therefor |
JP2527807B2 (en) | 1989-05-09 | 1996-08-28 | 住友大阪セメント株式会社 | Optical associative identification device |
EP0403305B1 (en) | 1989-06-16 | 1996-09-18 | Seiko Instruments Inc. | Optical pattern recognition apparatus |
US5850465A (en) | 1989-06-26 | 1998-12-15 | Fuji Photo Film Co., Ltd. | Abnormnal pattern detecting or judging apparatus, circular pattern judging apparatus, and image finding apparatus |
US5132842A (en) | 1989-07-21 | 1992-07-21 | Rockwell International Corporation | Optical image transformation system |
EP0412190B1 (en) | 1989-08-09 | 1993-10-27 | Heimann Systems GmbH & Co. KG | Device for transmitting fan-shaped radiation through objects |
US5132811A (en) | 1989-08-10 | 1992-07-21 | Seiko Instruments Inc. | Holographic operating optical apparatus |
US5022062A (en) | 1989-09-13 | 1991-06-04 | American Science And Engineering, Inc. | Automatic threat detection based on illumination by penetrating radiant energy using histogram processing |
US5198669A (en) * | 1989-09-20 | 1993-03-30 | Fujitsu Limited | Digital X-ray image processing apparatus |
US5257322A (en) | 1989-09-20 | 1993-10-26 | Agency Of Industrial Science And Technology | Method and apparatus for pattern recognition and display with optical correlator |
US5056130A (en) | 1989-10-06 | 1991-10-08 | The United States Of America As Represented By The Administrator, National Aeronautics And Space Administration | Computerized tomography calibrator |
JPH0670798B2 (en) | 1989-11-20 | 1994-09-07 | 住友セメント株式会社 | Optical pattern recognition method |
US5107351A (en) | 1990-02-16 | 1992-04-21 | Grumman Aerospace Corporation | Image enhanced optical correlator system |
US5323472A (en) | 1990-03-27 | 1994-06-21 | The Boeing Company | Optical image analyzer using optical correlation and opto-electronic feedback |
JP2551497B2 (en) | 1990-06-08 | 1996-11-06 | シャープ株式会社 | Image processing device |
US5138167A (en) | 1991-01-23 | 1992-08-11 | University Of Alabama - Birmingham | Split energy radiation detection |
US5257085A (en) | 1991-04-24 | 1993-10-26 | Kaman Aerospace Corporation | Spectrally dispersive imaging lidar system |
US5528702A (en) | 1991-05-31 | 1996-06-18 | Seiko Instruments Inc. | Optical pattern recognition apparatus with coordinate conversion function |
US5365560A (en) | 1991-07-29 | 1994-11-15 | General Electric Company | Method and apparatus for acquiring a uniform distribution of radon data sufficiently dense to constitute a complete set for exact image reconstruction of an object irradiated by a cone beam source |
JP3146534B2 (en) | 1991-08-02 | 2001-03-19 | 株式会社日立製作所 | Nuclear magnetic resonance equipment |
US5680525A (en) | 1991-08-08 | 1997-10-21 | Hitachi, Ltd. | Three-dimensional graphic system with an editor for generating a textrue mapping image |
US5367552A (en) | 1991-10-03 | 1994-11-22 | In Vision Technologies, Inc. | Automatic concealed object detection system having a pre-scan stage |
GB9122843D0 (en) * | 1991-10-28 | 1991-12-11 | Imperial College | Method and apparatus for image processing |
JP2849964B2 (en) | 1991-12-26 | 1999-01-27 | 富士写真フイルム株式会社 | Image processing method and apparatus |
GB9200828D0 (en) | 1992-01-15 | 1992-03-11 | Image Research Ltd | Improvements in and relating to material identification using x-rays |
EP0664038B1 (en) | 1992-02-18 | 2000-10-11 | Neopath, Inc. | Method for identifying objects using data processing techniques |
US5375156A (en) | 1992-03-31 | 1994-12-20 | Siemens Medical Systems, Inc. | Method and apparatus for 3-D computer tomography |
US5371542A (en) | 1992-06-23 | 1994-12-06 | The United States Of America As Represented By The Secretary Of The Navy | Dual waveband signal processing system |
JP2940317B2 (en) | 1992-06-24 | 1999-08-25 | 三菱電機株式会社 | Image processing device |
US5268967A (en) | 1992-06-29 | 1993-12-07 | Eastman Kodak Company | Method for automatic foreground and background detection in digital radiographic images |
US5329310A (en) | 1992-06-30 | 1994-07-12 | The Walt Disney Company | Method and apparatus for controlling distortion of a projected image |
US5757981A (en) | 1992-08-20 | 1998-05-26 | Toyo Ink Mfg. Co., Ltd. | Image inspection device |
US5327286A (en) | 1992-08-31 | 1994-07-05 | Texas Instruments Incorporated | Real time optical correlation system |
US5345081A (en) | 1992-09-10 | 1994-09-06 | Penetect, Inc. | Pit detector and method |
US6005916A (en) | 1992-10-14 | 1999-12-21 | Techniscan, Inc. | Apparatus and method for imaging with wavefields using inverse scattering techniques |
US5953452A (en) | 1992-11-05 | 1999-09-14 | The Johns Hopkins University | Optical-digital method and processor for pattern recognition |
US5376796A (en) | 1992-11-25 | 1994-12-27 | Adac Laboratories, Inc. | Proximity detector for body contouring system of a medical camera |
US5430787A (en) | 1992-12-03 | 1995-07-04 | The United States Of America As Represented By The Secretary Of Commerce | Compton scattering tomography |
US5311359A (en) | 1992-12-24 | 1994-05-10 | Litton Systems, Inc. | Reflective optical correlator with a folded asymmetrical optical axis |
US7241569B2 (en) | 1993-01-18 | 2007-07-10 | Olympus Corporation | Method and a device for the evaluation of biopolymer fitness |
DE4306187A1 (en) | 1993-02-27 | 1994-09-01 | Industrieanlagen Betriebsges | Method and arrangement of devices for protecting buildings and people against violent offenders |
US5777742A (en) | 1993-03-11 | 1998-07-07 | Environmental Research Institute Of Michigan | System and method for holographic imaging with discernible image of an object |
US5754621A (en) | 1993-03-15 | 1998-05-19 | Hitachi, Ltd. | X-ray inspection method and apparatus, prepreg inspecting method, and method for fabricating multi-layer printed circuit board |
CH693968A5 (en) | 1993-04-21 | 2004-05-14 | Fraunhofer Ges Forschung | Method and apparatus for the Topographiepruefung of surfaces. |
EP0696236B1 (en) | 1993-04-30 | 1996-10-30 | Robert Prof. Dr. Massen | Process and device for sorting materials |
FR2705786B1 (en) | 1993-05-28 | 1995-08-25 | Schlumberger Ind Sa | Method and device for recognizing certain materials in the composition of an object. |
US5420788A (en) | 1993-08-09 | 1995-05-30 | General Electric Company | Method for processing data in 2D and 3D computed X-ray tomography |
JP3326902B2 (en) | 1993-09-10 | 2002-09-24 | 株式会社日立製作所 | Pattern detection method, pattern detection apparatus, and projection exposure apparatus using the same |
CA2132138C (en) | 1993-09-29 | 2004-01-06 | Shih-Ping Wang | Computer-aided diagnosis system and method |
US5619596A (en) | 1993-10-06 | 1997-04-08 | Seiko Instruments Inc. | Method and apparatus for optical pattern recognition |
AU700802B2 (en) | 1993-10-29 | 1999-01-14 | Trustees Of The University Of Pennsylvania, The | Object imaging using diffuse light |
US6049381A (en) | 1993-10-29 | 2000-04-11 | The United States Of America As Represented By The Secretary Of The Navy | Real time suspended particle monitor |
US5910765A (en) | 1993-11-02 | 1999-06-08 | Advanced Optical Controls, Inc. | Sensor module |
DE4410603C1 (en) | 1994-03-26 | 1995-06-14 | Jenoptik Technologie Gmbh | Detecting faults during inspection of masks, LCDs, circuit boards and semiconductor wafers |
US5428657A (en) | 1994-03-22 | 1995-06-27 | Georgia Tech Research Corporation | X-ray monitoring system |
US5580471A (en) | 1994-03-30 | 1996-12-03 | Panasonic Technologies, Inc. | Apparatus and method for material treatment and inspection using fiber-coupled laser diode |
US5418380A (en) | 1994-04-12 | 1995-05-23 | Martin Marietta Corporation | Optical correlator using ferroelectric liquid crystal spatial light modulators and Fourier transform lenses |
JP3159869B2 (en) | 1994-05-09 | 2001-04-23 | 松下電器産業株式会社 | Thermal object detection device and measurement method |
US5546189A (en) | 1994-05-19 | 1996-08-13 | View Engineering, Inc. | Triangulation-based 3D imaging and processing method and system |
US6800452B1 (en) | 1994-08-08 | 2004-10-05 | Science Applications International Corporation | Automated methods for simultaneously performing a plurality of signal-based assays |
US6060677A (en) | 1994-08-19 | 2000-05-09 | Tiedemanns-Jon H. Andresen Ans | Determination of characteristics of material |
JP3449791B2 (en) | 1994-08-23 | 2003-09-22 | 理化学研究所 | Imaging method and imaging apparatus |
US5815264A (en) | 1994-09-21 | 1998-09-29 | Laser Sensor Technology, Inc | System for acquiring an image of a multi-phase fluid by measuring backscattered light |
US5519225A (en) | 1994-10-03 | 1996-05-21 | General Electric Company | System and method for using a dual modality detector for inspecting objects |
US5841907A (en) | 1994-11-14 | 1998-11-24 | The University Of Connecticut | Spatial integrating optical correlator for verifying the authenticity of a person, product or thing |
US6487307B1 (en) | 1994-11-30 | 2002-11-26 | Isoa, Inc. | System and method of optically inspecting structures on an object |
US5506880A (en) | 1994-12-16 | 1996-04-09 | Morton International, Inc. | X-ray inspection system and method |
WO1996020412A1 (en) | 1994-12-23 | 1996-07-04 | Digirad | Semiconductor gamma-ray camera and medical imaging system |
DE69629732T2 (en) | 1995-01-23 | 2004-07-15 | Fuji Photo Film Co., Ltd., Minami-Ashigara | Device for computer-aided diagnosis |
US6070583A (en) | 1995-02-21 | 2000-06-06 | Massachusetts Institute Of Technology | Optical imaging of tissue using inelastically scattered light |
DE19509007C2 (en) | 1995-03-13 | 2001-07-05 | Siemens Ag | C-arm X-ray diagnostic device for creating slice images |
DE19510168C2 (en) | 1995-03-21 | 2001-09-13 | Heimann Systems Gmbh & Co | Method and device for determining crystalline and polycrystalline materials in an examination area |
US5756875A (en) | 1995-04-14 | 1998-05-26 | The University Of Rochester | Thin film phantoms and phantom systems |
US5703921A (en) | 1995-05-30 | 1997-12-30 | Kabushiki Kaisha Toshiba | X-ray computed tomography apparatus |
CN1253636A (en) | 1995-06-22 | 2000-05-17 | 3Dv系统有限公司 | Telecentric stop 3-D camera and its method |
JP3140664B2 (en) | 1995-06-30 | 2001-03-05 | 松下電器産業株式会社 | Foreign matter inspection method and apparatus |
DE19532965C2 (en) | 1995-09-07 | 1998-07-16 | Heimann Systems Gmbh & Co | X-ray inspection system for large-volume goods |
US5642393A (en) | 1995-09-26 | 1997-06-24 | Vivid Technologies, Inc. | Detecting contraband by employing interactive multiprobe tomography |
JP3604467B2 (en) | 1995-09-27 | 2004-12-22 | 株式会社東芝 | Myocardial twist correction method |
US7045787B1 (en) | 1995-10-23 | 2006-05-16 | Science Applications International Corporation | Density detection using real time discrete photon counting for fast moving targets |
EP0774730B1 (en) | 1995-11-01 | 2005-08-24 | Canon Kabushiki Kaisha | Object extraction method, and image sensing apparatus using the method |
JPH09146228A (en) | 1995-11-20 | 1997-06-06 | Konica Corp | Radiograph forming method |
JP3441578B2 (en) | 1995-11-22 | 2003-09-02 | 株式会社モリタ製作所 | Dental panoramic radiography device |
JPH09149902A (en) | 1995-12-01 | 1997-06-10 | Hitachi Medical Corp | Tomography and tomograph |
US5796868A (en) | 1995-12-28 | 1998-08-18 | Cognex Corporation | Object edge point filtering system for machine vision |
US5809171A (en) | 1996-01-05 | 1998-09-15 | Mcdonnell Douglas Corporation | Image processing method and apparatus for correlating a test image with a template |
US6058159A (en) * | 1996-01-11 | 2000-05-02 | Control Screening, L.L.C. | Compact scanner apparatus and method |
US5981949A (en) | 1996-01-18 | 1999-11-09 | The United States Of America As Represented By The Secretary Of The Air Force | Locating defects in solid material |
US5778046A (en) | 1996-01-19 | 1998-07-07 | The Regents Of The University Of California | Automatic X-ray Beam Equalizer |
AUPN800796A0 (en) | 1996-02-09 | 1996-03-07 | Unisearch Limited | Visual inspection system for leather hide |
US5764683B1 (en) | 1996-02-12 | 2000-11-21 | American Science & Eng Inc | Mobile x-ray inspection system for large objects |
CN1218377A (en) | 1996-03-13 | 1999-06-02 | 模拟技术有限公司 | Computed tomography ring suppression filter |
BR9708028A (en) | 1996-03-13 | 1999-08-03 | Analogic Corp | Self-calibrating ring suppression filter for computed tomography |
US6044170A (en) | 1996-03-21 | 2000-03-28 | Real-Time Geometry Corporation | System and method for rapid shape digitizing and adaptive mesh generation |
US5739539A (en) | 1996-03-28 | 1998-04-14 | Adac Laboratories | Use of body boundary information to perform iterative reconstruction in medical imaging system |
US5893095A (en) * | 1996-03-29 | 1999-04-06 | Virage, Inc. | Similarity engine for content-based retrieval of images |
US5732529A (en) | 1996-03-29 | 1998-03-31 | Ethicon, Inc. | Apparatus for feeding foil stock in a process for making sealed sterile packages |
US5911139A (en) * | 1996-03-29 | 1999-06-08 | Virage, Inc. | Visual image database search engine which allows for different schema |
US5642394A (en) * | 1996-04-03 | 1997-06-24 | American Science And Engineering, Inc. | Sidescatter X-ray detection system |
US6586193B2 (en) | 1996-04-25 | 2003-07-01 | Genicon Sciences Corporation | Analyte assay using particulate labels |
JP3562902B2 (en) | 1996-04-26 | 2004-09-08 | 株式会社日立メディコ | RF probe for magnetic resonance imaging equipment |
US5699400A (en) * | 1996-05-08 | 1997-12-16 | Vivid Technologies, Inc. | Operator console for article inspection systems |
DE19619925C2 (en) | 1996-05-17 | 1999-09-09 | Sirona Dental Systems Gmbh | X-ray diagnostic device for tomosynthesis |
JPH09311109A (en) | 1996-05-22 | 1997-12-02 | Matsushita Electric Ind Co Ltd | Defect inspection method and device utilizing light |
US5815198A (en) | 1996-05-31 | 1998-09-29 | Vachtsevanos; George J. | Method and apparatus for analyzing an image to detect and identify defects |
US5930314A (en) | 1996-05-31 | 1999-07-27 | Massachusetts Institute Of Technology | Coded aperture imaging |
US5799100A (en) | 1996-06-03 | 1998-08-25 | University Of South Florida | Computer-assisted method and apparatus for analysis of x-ray images using wavelet transforms |
US6043870A (en) | 1996-07-01 | 2000-03-28 | Cybernet Systems Corporation | Compact fiber optic electronic laser speckle pattern interferometer |
US5638420A (en) | 1996-07-03 | 1997-06-10 | Advanced Research And Applications Corporation | Straddle inspection system |
US5838759A (en) * | 1996-07-03 | 1998-11-17 | Advanced Research And Applications Corporation | Single beam photoneutron probe and X-ray imaging system for contraband detection and identification |
EP1015913A1 (en) | 1996-07-22 | 2000-07-05 | American Science & Engineering, Inc. | System for rapid x-ray inspection of enclosures |
US5943388A (en) | 1996-07-30 | 1999-08-24 | Nova R & D, Inc. | Radiation detector and non-destructive inspection |
US5960104A (en) | 1996-08-16 | 1999-09-28 | Virginia Polytechnic & State University | Defect detection system for lumber |
EP0825457A3 (en) | 1996-08-19 | 2002-02-13 | Analogic Corporation | Multiple angle pre-screening tomographic systems and methods |
US5974111A (en) * | 1996-09-24 | 1999-10-26 | Vivid Technologies, Inc. | Identifying explosives or other contraband by employing transmitted or scattered X-rays |
WO1998016817A1 (en) | 1996-10-16 | 1998-04-23 | Illinois Institute Of Technology | Method for detecting an image of an object |
US5668846A (en) | 1996-10-18 | 1997-09-16 | General Electric Company | Methods and apparatus for scanning an object and displaying an image in a computed tomography system |
WO1998020366A1 (en) | 1996-11-08 | 1998-05-14 | American Science And Engineering, Inc. | Coded aperture x-ray imaging system |
JPH10164437A (en) | 1996-11-26 | 1998-06-19 | Canon Inc | X-ray image-pickup device and drive method for x-ray image-pickup element |
KR100188712B1 (en) | 1996-11-28 | 1999-06-01 | 이종구 | Cross-sectional image apparatus and method for obtaining cross-sectional image using the same |
IL119850A (en) | 1996-12-17 | 2000-11-21 | Prolaser Ltd | Optical method and apparatus for detecting low frequency defects |
US5748697A (en) | 1996-12-20 | 1998-05-05 | Siemens Corporate Research, Inc. | Method and apparatus for eliminating boundary errors in cone beam imaging |
US6057761A (en) | 1997-01-21 | 2000-05-02 | Spatial Dynamics, Ltd. | Security system and method |
US6480285B1 (en) | 1997-01-28 | 2002-11-12 | Zetetic Institute | Multiple layer confocal interference microscopy using wavenumber domain reflectometry and background amplitude reduction and compensation |
US5802134A (en) | 1997-04-09 | 1998-09-01 | Analogic Corporation | Nutating slice CT image reconstruction apparatus and method |
AUPO625497A0 (en) | 1997-04-16 | 1997-05-15 | Macquarie Research Limited | Analysis of molecules |
EP0979450B1 (en) | 1997-04-30 | 2002-06-19 | The Foxboro Company | Methods and systems for synchronizing processes executing on a digital data processing system |
US6008496A (en) | 1997-05-05 | 1999-12-28 | University Of Florida | High resolution resonance ionization imaging detector and method |
US5909285A (en) | 1997-05-05 | 1999-06-01 | Beaty; Elwin M. | Three dimensional inspection system |
US5779641A (en) | 1997-05-07 | 1998-07-14 | General Electric Company | Method and apparatus for three-dimensional ultrasound imaging by projecting filtered pixel data |
US5877849A (en) | 1997-05-12 | 1999-03-02 | Advanced Optical Technologies, Llc | Object detection system |
US5926568A (en) | 1997-06-30 | 1999-07-20 | The University Of North Carolina At Chapel Hill | Image object matching using core analysis and deformable shape loci |
US6292260B1 (en) | 1997-07-03 | 2001-09-18 | Isoa, Inc. | System and method of optically inspecting surface structures on an object |
JPH1156828A (en) | 1997-08-27 | 1999-03-02 | Fuji Photo Film Co Ltd | Abnormal shadow candidate detecting method and its device |
EP1012586A2 (en) | 1997-09-09 | 2000-06-28 | American Science & Engineering, Inc. | A tomographic inspection system |
US7028899B2 (en) * | 1999-06-07 | 2006-04-18 | Metrologic Instruments, Inc. | Method of speckle-noise pattern reduction and apparatus therefore based on reducing the temporal-coherence of the planar laser illumination beam before it illuminates the target object by applying temporal phase modulation techniques during the transmission of the plib towards the target |
US5901196A (en) | 1997-09-30 | 1999-05-04 | Siemens Corporate Research, Inc. | Reduction of hitlist size in spiral cone beam CT by use of local radon origins |
US5901198A (en) | 1997-10-10 | 1999-05-04 | Analogic Corporation | Computed tomography scanning target detection using target surface normals |
US6256404B1 (en) | 1997-10-10 | 2001-07-03 | Analogic Corporation | Computed tomography scanning apparatus and method using adaptive reconstruction window |
US6149300A (en) | 1997-10-17 | 2000-11-21 | Welch Allyn, Inc. | Intraoral imaging system |
GB9726386D0 (en) | 1997-12-12 | 1998-02-11 | Univ Cambridge Tech | Optical correlator |
US6442288B1 (en) | 1997-12-17 | 2002-08-27 | Siemens Aktiengesellschaft | Method for reconstructing a three-dimensional image of an object scanned in the context of a tomosynthesis, and apparatus for tomosynthesis |
JP4497570B2 (en) | 1998-01-22 | 2010-07-07 | 株式会社東芝 | Diagnostic imaging equipment |
US6035014A (en) | 1998-02-11 | 2000-03-07 | Analogic Corporation | Multiple-stage apparatus and method for detecting objects in computed tomography data |
US20030165213A1 (en) | 1998-02-18 | 2003-09-04 | Maglich Bogdan C. | Method and apparatus for neutron microscopy with stoichiometric imaging |
JP3385994B2 (en) * | 1998-02-27 | 2003-03-10 | 株式会社ニコン | Image detector |
US6289235B1 (en) | 1998-03-05 | 2001-09-11 | Wake Forest University | Method and system for creating three-dimensional images using tomosynthetic computed tomography |
US6388788B1 (en) | 1998-03-16 | 2002-05-14 | Praelux, Inc. | Method and apparatus for screening chemical compounds |
DE19812055C2 (en) | 1998-03-19 | 2002-08-08 | Heimann Systems Gmbh & Co | Image processing for material detection using X-rays |
US6218943B1 (en) * | 1998-03-27 | 2001-04-17 | Vivid Technologies, Inc. | Contraband detection and article reclaim system |
US5881123A (en) | 1998-03-31 | 1999-03-09 | Siemens Corporate Research, Inc. | Simplified cone beam image reconstruction using 3D backprojection |
US6009142A (en) | 1998-03-31 | 1999-12-28 | Siemens Corporate Research, Inc. | Practical cone beam image reconstruction using local regions-of-interest |
US6094472A (en) | 1998-04-14 | 2000-07-25 | Rapiscan Security Products, Inc. | X-ray backscatter imaging system including moving body tracking assembly |
US6153873A (en) | 1998-05-20 | 2000-11-28 | E. I. Dupont De Numours And Company | Optical probe having an imaging apparatus |
US6088423A (en) | 1998-06-05 | 2000-07-11 | Vivid Technologies, Inc. | Multiview x-ray based system for detecting contraband such as in baggage |
US6621888B2 (en) | 1998-06-18 | 2003-09-16 | American Science And Engineering, Inc. | X-ray inspection by coherent-scattering from variably disposed scatterers identified as suspect objects |
JP4078716B2 (en) | 1998-06-29 | 2008-04-23 | ソニー株式会社 | Image processing apparatus and method, and providing medium |
US6899540B1 (en) | 1998-07-30 | 2005-05-31 | The United States Of America As Represented By The Secretary Of Transportation | Threat image projection system |
US6080994A (en) | 1998-07-30 | 2000-06-27 | Litton Systems, Inc. | High output reflective optical correlator having a folded optical axis using ferro-electric liquid crystal spatial light modulators |
US6163403A (en) | 1998-07-30 | 2000-12-19 | Litton Systems, Inc. | High output reflective optical correlator having a folded optical axis using grayscale spatial light modulators |
JP3299193B2 (en) | 1998-08-21 | 2002-07-08 | 日本電気株式会社 | Bump inspection method / apparatus, information storage medium |
JP2000069243A (en) | 1998-08-24 | 2000-03-03 | Matsushita Electric Ind Co Ltd | Image processing method and image processor |
US6714623B2 (en) | 1998-08-31 | 2004-03-30 | Canon Kabushiki Kaisha | Image collecting system |
US6850597B2 (en) | 1998-09-11 | 2005-02-01 | Canon Kabushiki Kaisha | X-ray image photographing apparatus and grid device |
US6078638A (en) | 1998-09-30 | 2000-06-20 | Siemens Corporate Research, Inc. | Pixel grouping for filtering cone beam detector data during 3D image reconstruction |
JP3619027B2 (en) | 1998-09-30 | 2005-02-09 | キヤノン株式会社 | X-ray imaging apparatus, X-ray imaging system, X-ray image processing method, and computer-readable recording medium |
DE19848765C2 (en) | 1998-10-22 | 2000-12-21 | Brainlab Med Computersyst Gmbh | Position verification in camera images |
DE19849764B4 (en) | 1998-10-28 | 2004-10-21 | Siemens Ag | Imaging diagnostic facility |
US6887710B2 (en) | 1998-11-13 | 2005-05-03 | Mesosystems Technology, Inc. | Robust system for screening mail for biological agents |
US6278760B1 (en) | 1998-11-13 | 2001-08-21 | Fuji Photo Film Co., Ltd. | Radiation image forming method and apparatus |
US6236708B1 (en) | 1998-11-25 | 2001-05-22 | Picker International, Inc. | 2D and 3D tomographic X-ray imaging using flat panel detectors |
US6222902B1 (en) | 1998-11-25 | 2001-04-24 | Picker International, Inc. | Real-time tomographic system with flat panel detectors |
DE19855250A1 (en) | 1998-11-30 | 2000-06-08 | Heimann Systems Gmbh & Co | Process for optimizing an X-ray image |
US6421420B1 (en) | 1998-12-01 | 2002-07-16 | American Science & Engineering, Inc. | Method and apparatus for generating sequential beams of penetrating radiation |
WO2000035254A1 (en) | 1998-12-08 | 2000-06-15 | Koninklijke Philips Electronics N.V. | An x-ray examination apparatus having an object absorption dependent brightness control |
US6754374B1 (en) | 1998-12-16 | 2004-06-22 | Surgical Navigation Technologies, Inc. | Method and apparatus for processing images with regions representing target objects |
US6229872B1 (en) | 1998-12-22 | 2001-05-08 | United Technologies Corporation | Method and apparatus for use in inspection of objects |
US6373970B1 (en) | 1998-12-29 | 2002-04-16 | General Electric Company | Image registration using fourier phase matching |
DE69936769T2 (en) | 1998-12-30 | 2008-04-30 | General Electric Co. | IMAGE THICKNESS ELECTRICATION FOR MULTI-LAYER IMAGE DEVICE |
JP2000197627A (en) | 1999-01-05 | 2000-07-18 | Hitachi Medical Corp | X-ray ct device |
US6249341B1 (en) | 1999-01-25 | 2001-06-19 | Amnis Corporation | Imaging and analyzing parameters of small moving objects such as cells |
US6373979B1 (en) * | 1999-01-29 | 2002-04-16 | Lg Electronics, Inc. | System and method for determining a level of similarity among more than one image and a segmented data structure for enabling such determination |
TW418342B (en) | 1999-02-02 | 2001-01-11 | Samsung Electronics Co Ltd | Sectional image photography system and method thereof |
JP3866431B2 (en) | 1999-02-17 | 2007-01-10 | 株式会社東芝 | X-ray CT system |
US6370222B1 (en) | 1999-02-17 | 2002-04-09 | Ccvs, Llc | Container contents verification |
US6272204B1 (en) | 1999-02-23 | 2001-08-07 | Cr Technology, Inc. | Integrated X-ray and visual inspection systems |
US7068825B2 (en) | 1999-03-08 | 2006-06-27 | Orametrix, Inc. | Scanning system and calibration method for capturing precise three-dimensional information of objects |
DE19910615C1 (en) | 1999-03-10 | 2000-06-21 | Heimann Systems Gmbh & Co | Ergonomic operating station for x-ray equipment includes seat, monitor and control panel with individual adjustments allowing user of any size to customize settings for ideal posture, sitting or standing |
FR2790851B1 (en) | 1999-03-12 | 2001-06-08 | Ge Medical Syst Sa | METHOD FOR IMPROVING THE DETECTION OF ELEMENTS OF INTEREST IN A DIGITAL RADIOGRAPHIC IMAGE |
US6448545B1 (en) | 2000-01-18 | 2002-09-10 | Syncrotronics Corp. | Precision endoscopic imaging system |
US6650779B2 (en) | 1999-03-26 | 2003-11-18 | Georgia Tech Research Corp. | Method and apparatus for analyzing an image to detect and identify patterns |
US6288974B1 (en) | 1999-03-30 | 2001-09-11 | The United States Of Americas As Represented By The Secretary Of The Navy | System and method for enhancing detection of objects through an obscuring medium |
DE19916664A1 (en) | 1999-04-14 | 2000-10-19 | Heimann Systems Gmbh & Co | X-ray image processing method, involves checking coincidence of matching sides of image markings and ratio of common areas to achieve greater clarity by forming final summation marking |
US6256370B1 (en) | 2000-01-24 | 2001-07-03 | General Electric Company | Method and apparatus for performing tomosynthesis |
JP4159701B2 (en) | 1999-04-20 | 2008-10-01 | 富士フイルム株式会社 | Evaluation method and apparatus for digital radiographic image |
US6292530B1 (en) | 1999-04-29 | 2001-09-18 | General Electric Company | Method and apparatus for reconstructing image data acquired by a tomosynthesis x-ray imaging system |
US6403960B1 (en) | 1999-04-29 | 2002-06-11 | Koninklijijke Philips Electronics N.V. | Correction for spatial variations of deadtime of a monolithic scintillator based detector in a medical imaging system |
DE10022468A1 (en) | 1999-05-10 | 2001-10-11 | Shimadzu Corp | X-ray computerized tomography apparatus for medical diagnosis, selects scanning type based on comparison result of visual field and image pick-up ranges of radiated X-ray |
US6731819B1 (en) | 1999-05-21 | 2004-05-04 | Olympus Optical Co., Ltd. | Optical information processing apparatus capable of various types of filtering and image processing |
US6236704B1 (en) | 1999-06-30 | 2001-05-22 | Siemens Corporate Research, Inc. | Method and apparatus using a virtual detector for three-dimensional reconstruction from x-ray images |
US6919959B2 (en) | 1999-06-30 | 2005-07-19 | Masten Opto-Diagnostics Co. | Digital spectral identifier-controller and related methods |
US6813395B1 (en) * | 1999-07-14 | 2004-11-02 | Fuji Photo Film Co., Ltd. | Image searching method and image processing method |
US6233303B1 (en) | 1999-07-21 | 2001-05-15 | Siemens Corporate Research, Inc. | Method and apparatus for reducing X-ray dosage in a spiral scan cone beam CT imaging system |
US6785410B2 (en) * | 1999-08-09 | 2004-08-31 | Wake Forest University Health Sciences | Image reporting method and system |
US6744909B1 (en) | 1999-08-19 | 2004-06-01 | Physical Optics Corporation | Authentication system and method |
DE19943183A1 (en) | 1999-09-09 | 2001-03-15 | Heimann Systems Gmbh & Co | Method for color matching an image, in particular an X-ray image |
DE19959081A1 (en) | 1999-09-09 | 2001-03-22 | Heimann Systems Gmbh & Co | Process for image management of X-ray images |
US6473489B2 (en) | 1999-09-30 | 2002-10-29 | Siemens Corporate Research, Inc | Apparatus for superimposition of X-ray and video images |
DE19954663B4 (en) | 1999-11-13 | 2006-06-08 | Smiths Heimann Gmbh | Method and device for determining a material of a detected object |
US6542578B2 (en) | 1999-11-13 | 2003-04-01 | Heimann Systems Gmbh | Apparatus for determining the crystalline and polycrystalline materials of an item |
US6751349B2 (en) | 1999-11-30 | 2004-06-15 | Fuji Photo Film Co., Ltd. | Image processing system |
US7050646B2 (en) | 1999-12-10 | 2006-05-23 | British Telecommunications Public Limited Company | Image processing system and method for image segmentation using intensity contrast and depth contrast values |
FR2803069B1 (en) | 1999-12-28 | 2002-12-13 | Ge Medical Syst Sa | METHOD AND SYSTEM FOR COMPENSATING THE THICKNESS OF AN ORGAN |
US7108970B2 (en) | 2000-01-07 | 2006-09-19 | Transform Pharmaceuticals, Inc. | Rapid identification of conditions, compounds, or compositions that inhibit, prevent, induce, modify, or reverse transitions of physical state |
US6418189B1 (en) | 2000-01-24 | 2002-07-09 | Analogic Corporation | Explosive material detection apparatus and method using dual energy information of a scan |
US6570177B1 (en) | 2000-01-31 | 2003-05-27 | Dcs Corporation | System, apparatus, and method for detecting defects in particles |
JP2001212144A (en) | 2000-01-31 | 2001-08-07 | Toshiba Corp | Ultrasonic diagnostic apparatus and ultrasonic imaging method |
US6744848B2 (en) | 2000-02-11 | 2004-06-01 | Brandeis University | Method and system for low-dose three-dimensional imaging of a scene |
US6735274B1 (en) | 2000-02-15 | 2004-05-11 | Koninklijke Philips Electronics N.V. | Clinical screening CT systems |
JP3643745B2 (en) | 2000-02-21 | 2005-04-27 | 株式会社モリタ製作所 | X-ray imaging detector and X-ray imaging apparatus |
GB2372930B (en) | 2000-03-03 | 2003-03-19 | Teraview Ltd | Apparatus and method for investigating a sample |
AU2001249138A1 (en) | 2000-03-10 | 2001-09-24 | Cyrano Sciences, Inc. | Measuring and analyzing multi-dimensional sensory information for identificationpurposes |
US6865509B1 (en) | 2000-03-10 | 2005-03-08 | Smiths Detection - Pasadena, Inc. | System for providing control to an industrial process using one or more multidimensional variables |
DE10011318A1 (en) | 2000-03-13 | 2001-09-20 | Visicontrol Ges Fuer Elektroni | Image processing apparatus for inspection and/or measurement of objects, includes infrared light source |
US6628982B1 (en) | 2000-03-30 | 2003-09-30 | The Regents Of The University Of Michigan | Internal marker device for identification of biological substances |
JP3707347B2 (en) | 2000-04-07 | 2005-10-19 | 株式会社島津製作所 | Image processing method for X-ray CT apparatus, X-ray CT apparatus, and recording medium for X-ray CT imaging |
US7274810B2 (en) | 2000-04-11 | 2007-09-25 | Cornell Research Foundation, Inc. | System and method for three-dimensional image rendering and analysis |
US6577708B2 (en) | 2000-04-17 | 2003-06-10 | Leroy Dean Chapman | Diffraction enhanced x-ray imaging of articular cartilage |
AU5063301A (en) | 2000-04-20 | 2001-11-12 | Greenvision Systems Ltd. | Method for generating intra-particle morphological concentration / density maps and histograms of a chemically pure particulate substance |
US6980681B1 (en) | 2000-04-24 | 2005-12-27 | Ge Medical Systems Global Technology Company, Llc | Methods and apparatus for helical reconstruction for multislice CT scan |
US6353673B1 (en) | 2000-04-27 | 2002-03-05 | Physical Optics Corporation | Real-time opto-electronic image processor |
JP2001299733A (en) | 2000-04-27 | 2001-10-30 | Konica Corp | Pci radiation beam image processing apparatus, pci radiation beam image detecting and processing apparatus, pci radiation beam image outputting apparatus and pci radiation beam image diagnosis supporting apparatus |
US7359541B2 (en) | 2000-04-28 | 2008-04-15 | Konica Corporation | Radiation image processing apparatus |
US6408046B2 (en) | 2000-04-29 | 2002-06-18 | General Electric Company | Method for dynamic range management of digital radiographic images |
CA2307439C (en) | 2000-05-02 | 2008-07-08 | Institut National D'optique | Method and apparatus for evaluating a scale factor and a rotation angle in image processing |
US6549683B1 (en) | 2000-05-02 | 2003-04-15 | Institut National D'optique | Method and apparatus for evaluating a scale factor and a rotation angle in image processing |
JP4812048B2 (en) | 2000-05-09 | 2011-11-09 | 株式会社日立メディコ | Ultrasonic diagnostic equipment |
US6583420B1 (en) | 2000-06-07 | 2003-06-24 | Robert S. Nelson | Device and system for improved imaging in nuclear medicine and mammography |
JP2001346790A (en) | 2000-06-12 | 2001-12-18 | Hitachi Medical Corp | X-ray detector, and x-ray ct device using the same |
JP3517388B2 (en) | 2000-06-14 | 2004-04-12 | 新光電気工業株式会社 | Bump inspection method and bump inspection device |
JP2002042169A (en) | 2000-07-06 | 2002-02-08 | Infiniteface.Com Inc | Three-dimensional image providing system, its method, morphing image providing system, and its method |
US7193706B2 (en) | 2000-08-02 | 2007-03-20 | Arizona Board Of Regents, Acting On Behalf Of Arizona State University | Computer interfaced scanning fluorescence lifetime microscope applied to directed evolution |
US6381297B1 (en) | 2000-08-15 | 2002-04-30 | Ge Medical Systems Global Technology Company, Llc | High pitch reconstruction of multislice CT scans |
DE10044357A1 (en) | 2000-09-07 | 2002-03-21 | Heimann Systems Gmbh & Co | Detector arrangement for the detection of X-rays |
US7819003B2 (en) * | 2002-06-11 | 2010-10-26 | Intelligent Technologies International, Inc. | Remote monitoring of fluid storage tanks |
DE10046091C1 (en) | 2000-09-18 | 2002-01-17 | Siemens Ag | Computer tomography device has optical imaging device providing visual images of area of patient examined via X-ray beam |
CA2319898C (en) | 2000-09-18 | 2006-05-16 | Institut National D'optique | Position encoding optical device and method |
CA2319958C (en) | 2000-09-18 | 2007-03-13 | Institut National D'optique | Image processing apparatus and method with locking feature |
JP2002095655A (en) | 2000-09-26 | 2002-04-02 | Shimadzu Corp | Ct apparatus |
WO2002028154A1 (en) | 2000-09-29 | 2002-04-04 | Analogic Corporation | Method of and system for improving the signal to noise characteristics of images from a digital |
US6720874B2 (en) | 2000-09-29 | 2004-04-13 | Ids Systems, Inc. | Portal intrusion detection apparatus and method |
US6324245B1 (en) | 2000-10-17 | 2001-11-27 | Siemens Corporation Research, Inc. | Method and apparatus for identifying and correcting image inaccuracies caused by simplified processing of masked cone beam projection data |
US6628983B1 (en) | 2000-10-25 | 2003-09-30 | Koninklijke Philips Electronics N.V. | Nuclear imaging systems and methods with feature-enhanced transmission imaging |
US6763148B1 (en) * | 2000-11-13 | 2004-07-13 | Visual Key, Inc. | Image recognition methods |
AU2002219847A1 (en) | 2000-11-15 | 2002-05-27 | Real Time Metrology, Inc. | Optical method and apparatus for inspecting large area planar objects |
KR100406113B1 (en) | 2000-11-27 | 2003-11-15 | 학교법인 동국학원 | searching apparatus using X-band and millimeter-wave signal |
US6671541B2 (en) | 2000-12-01 | 2003-12-30 | Neomed Technologies, Inc. | Cardiovascular imaging and functional analysis system |
DE10062214B4 (en) | 2000-12-13 | 2013-01-24 | Smiths Heimann Gmbh | Devices for transilluminating objects |
US7171038B2 (en) | 2000-12-15 | 2007-01-30 | Kla-Tencor Technologies Corporation | Method and apparatus for inspecting a substrate |
NL1016916C2 (en) | 2000-12-15 | 2002-07-02 | Univ Delft Tech | Method and device for analyzing and separating material flows. |
US6473488B2 (en) | 2000-12-20 | 2002-10-29 | Cedara Software Corp. | Three dimensional image reconstruction from single plane X-ray fluorograms |
US6463181B2 (en) * | 2000-12-22 | 2002-10-08 | The United States Of America As Represented By The Secretary Of The Navy | Method for optimizing visual display of enhanced digital images |
US6463118B2 (en) | 2000-12-29 | 2002-10-08 | Ge Medical Systems Global Technology Company, Llc | Computed tomography (CT) weighting for high quality image recontruction |
US6418184B1 (en) | 2000-12-29 | 2002-07-09 | Ge Medical Systems Global Technology Company, Llc | Helical rowwise view weighting of computed tomographic images |
US20020094119A1 (en) | 2001-01-18 | 2002-07-18 | Velayudhan Sahadevan | High resolution digitized image analysis of chest x-rays for diagnosis of difficult to visualize evolving very ealrly stage lung cancer, pnumoconiosis and pulmonary diseases |
US20020127586A1 (en) | 2001-01-31 | 2002-09-12 | Novozymes A/S | Method of analyzing granular composition by fluorescene analysis |
WO2002062201A2 (en) | 2001-02-02 | 2002-08-15 | Dana-Farber Cancer Institute, Inc. | Rare event detection system |
US6587575B1 (en) | 2001-02-09 | 2003-07-01 | The United States Of America As Represented By The Secretary Of Agriculture | Method and system for contaminant detection during food processing |
US6477221B1 (en) | 2001-02-16 | 2002-11-05 | University Of Rochester | System and method for fast parallel cone-beam reconstruction using one or more microprocessors |
DE10108295B4 (en) | 2001-02-21 | 2004-01-29 | Sirona Dental Systems Gmbh | Tooth identification on digital x-rays and assignment of information to digital x-rays |
US6876322B2 (en) | 2003-06-26 | 2005-04-05 | Battelle Memorial Institute | Concealed object detection |
US6366638B1 (en) | 2001-03-16 | 2002-04-02 | Ge Medical Systems Global Technology Company, Llc | Methods and apparatus for CT scout image processing |
US6522775B2 (en) | 2001-03-28 | 2003-02-18 | Alan C. Nelson | Apparatus and method for imaging small objects in a flow stream using optical tomography |
WO2002080100A2 (en) | 2001-03-29 | 2002-10-10 | Koninklijke Philips Electronics N.V. | A method of reducing artifacts in object images |
US6968034B2 (en) | 2001-04-03 | 2005-11-22 | L-3 Communications Security And Detection Systems, Inc. | X-ray inspection system |
DE10122279A1 (en) | 2001-05-08 | 2002-12-12 | Heimann Systems Gmbh & Co | X-ray system |
US6778681B2 (en) | 2001-05-09 | 2004-08-17 | Invision Technologies, Inc. | Analysis and presentation of internal features of logs |
US20020172409A1 (en) | 2001-05-18 | 2002-11-21 | Motoaki Saito | Displaying three-dimensional medical images |
US6597760B2 (en) | 2001-05-23 | 2003-07-22 | Heimann Systems Gmbh | Inspection device |
US6580778B2 (en) | 2001-05-23 | 2003-06-17 | Heimann Systems Gmbh | Inspection device |
US7462852B2 (en) * | 2001-12-17 | 2008-12-09 | Tecomet, Inc. | Devices, methods, and systems involving cast collimators |
US6721387B1 (en) | 2001-06-13 | 2004-04-13 | Analogic Corporation | Method of and system for reducing metal artifacts in images generated by x-ray scanning devices |
US6408042B1 (en) | 2001-06-15 | 2002-06-18 | Ge Medical Systems Global Technology Company, Llc | Methods and apparatus for cone beam artifact suppression in scanning imaging systems |
JP2003072186A (en) | 2001-06-19 | 2003-03-12 | Canon Inc | Imaging apparatus, imaging method, program and recording medium |
WO2002103338A1 (en) | 2001-06-20 | 2002-12-27 | Koninklijke Philips Electronics N.V. | X-ray device for medical examination and method of improving the image quality thereof |
CN1287736C (en) | 2001-06-22 | 2006-12-06 | 株式会社日立医药 | X-ray image diagnosing device and x-ray image data correcting method |
US6663280B2 (en) | 2001-06-28 | 2003-12-16 | Heimann Systems Gmbh | Inspection unit |
US6611575B1 (en) | 2001-07-27 | 2003-08-26 | General Electric Company | Method and system for high resolution 3D visualization of mammography images |
DE10139672A1 (en) | 2001-08-11 | 2003-03-06 | Heimann Systems Gmbh & Co | Method and installation for inspecting an object, in particular a piece of luggage |
US7042977B2 (en) | 2001-09-05 | 2006-05-09 | Koninklijke Philips Electronics N.V. | Dose control in CT-images |
CA2460212C (en) | 2001-09-06 | 2013-01-22 | Genomic Profiling Systems, Inc. | Rapid and sensitive detection of cells and viruses |
US7218775B2 (en) | 2001-09-17 | 2007-05-15 | Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of Agriculture And Agrifood | Method and apparatus for identifying and quantifying characteristics of seeds and other small objects |
US6610977B2 (en) * | 2001-10-01 | 2003-08-26 | Lockheed Martin Corporation | Security system for NBC-safe building |
US20060274916A1 (en) | 2001-10-01 | 2006-12-07 | L-3 Communications Security And Detection Systems | Remote data access |
US20060115109A1 (en) * | 2001-10-01 | 2006-06-01 | L-3 Communications Security And Detection Systems, Inc. | Ensuring airline safety while safeguarding personal passenger information |
US6559769B2 (en) * | 2001-10-01 | 2003-05-06 | Eric Anthony | Early warning real-time security system |
DE10149254B4 (en) | 2001-10-05 | 2006-04-20 | Smiths Heimann Gmbh | Method and device for detecting a specific material in an object by means of electromagnetic radiation |
US6621887B2 (en) | 2001-10-15 | 2003-09-16 | General Electric Company | Method and apparatus for processing a fluoroscopic image |
US6792070B2 (en) | 2001-10-16 | 2004-09-14 | Fuji Photo Film Co., Ltd. | Radiation image recording method and apparatus |
US6661867B2 (en) * | 2001-10-19 | 2003-12-09 | Control Screening, Llc | Tomographic scanning X-ray inspection system using transmitted and compton scattered radiation |
US6574297B2 (en) | 2001-10-30 | 2003-06-03 | Siemens Corporate Research, Inc. | System and method for image reconstruction in a cone beam imaging system |
JP3948249B2 (en) * | 2001-10-30 | 2007-07-25 | 日本電気株式会社 | Similarity determination apparatus, similarity determination method, and program |
AU2002348143A1 (en) | 2001-10-31 | 2003-05-12 | Vrex, Inc. | 3d stereoscopic x-ray system with two different object trajectories |
US6946655B2 (en) | 2001-11-07 | 2005-09-20 | Applied Materials, Inc. | Spot grid array electron imaging system |
US6618465B2 (en) | 2001-11-12 | 2003-09-09 | General Electric Company | X-ray shielding system and shielded digital radiographic inspection system and method |
WO2003043498A2 (en) | 2001-11-17 | 2003-05-30 | Council For The Central Laboratory Of The Research Councils | Method and apparatus for obtaining simultaneously absorption and refraction images by use of a monochromator with integrated radiation detector |
US7058210B2 (en) | 2001-11-20 | 2006-06-06 | General Electric Company | Method and system for lung disease detection |
US20040041082A1 (en) | 2001-11-27 | 2004-03-04 | Harmon Gary R. | Molecular sensing array |
US20030210139A1 (en) | 2001-12-03 | 2003-11-13 | Stephen Brooks | Method and system for improved security |
US6873261B2 (en) * | 2001-12-07 | 2005-03-29 | Eric Anthony | Early warning near-real-time security system |
JP2003175027A (en) | 2001-12-10 | 2003-06-24 | Hitachi Medical Corp | X-ray ct system |
DE10163351C1 (en) | 2001-12-14 | 2003-06-26 | Heimann Biometric Systems Gmbh | Method and arrangement for the low-distortion recording of intensity patterns created on a contact surface by disturbed total reflection |
US7012256B1 (en) * | 2001-12-21 | 2006-03-14 | National Recovery Technologies, Inc. | Computer assisted bag screening system |
US6542580B1 (en) | 2002-01-15 | 2003-04-01 | Rapiscan Security Products (Usa), Inc. | Relocatable X-ray imaging system and method for inspecting vehicles and containers |
US20030138147A1 (en) | 2002-01-17 | 2003-07-24 | Yandi Ongkojoyo | Object recognition system for screening device |
ES2409268T3 (en) | 2002-01-19 | 2013-06-26 | Roche Pvt Gmbh | Procedure for classifying color images of serum samples from centrifuged body fluids |
US6603536B1 (en) | 2002-01-24 | 2003-08-05 | Textron Systems Corporation | Hybrid optical correlator/digital processor for target detection and discrimination |
US9092841B2 (en) | 2004-06-09 | 2015-07-28 | Cognex Technology And Investment Llc | Method and apparatus for visual detection and inspection of objects |
US7738995B2 (en) | 2002-02-12 | 2010-06-15 | United Parcel Service Of America, Inc. | Systems and methods for use in electronic processing of foreign origin and export shipments and/or passengers and baggage at security check points |
GB0203530D0 (en) | 2002-02-14 | 2002-04-03 | Lettington Alan H | Scanning apparatus |
US6618466B1 (en) | 2002-02-21 | 2003-09-09 | University Of Rochester | Apparatus and method for x-ray scatter reduction and correction for fan beam CT and cone beam volume CT |
US6654443B1 (en) | 2002-02-25 | 2003-11-25 | Ge Medical Systems Global Technology Co., Llc | Thermal sensing detector cell for a computed tomography system and method of manufacturing same |
US6459755B1 (en) | 2002-02-26 | 2002-10-01 | Ge Medical Systems Global Technology Co. Llc | Method and apparatus for administering low dose CT scans |
US6803997B2 (en) | 2002-03-08 | 2004-10-12 | Anzus, Inc. | Gridlocking and correlation methods and arrangements |
US6665373B1 (en) | 2002-03-12 | 2003-12-16 | Rapiscan Security Products (Usa), Inc. | X-ray imaging system with active detector |
EP1347309A3 (en) | 2002-03-20 | 2012-04-18 | Hitachi, Ltd. | Radiological imaging apparatus and method |
US7050616B2 (en) * | 2002-04-01 | 2006-05-23 | Ge Medical Systems Global Technology Company, Llc | Data transmission scheme and system for image reconstruction |
JP4149189B2 (en) | 2002-04-04 | 2008-09-10 | 株式会社日立メディコ | X-ray CT system |
US7043474B2 (en) * | 2002-04-15 | 2006-05-09 | International Business Machines Corporation | System and method for measuring image similarity based on semantic meaning |
US7218766B2 (en) | 2002-04-15 | 2007-05-15 | General Electric Company | Computer aided detection (CAD) for 3D digital mammography |
US7811825B2 (en) | 2002-04-19 | 2010-10-12 | University Of Washington | System and method for processing specimens and images for optical tomography |
US7071823B2 (en) * | 2002-04-24 | 2006-07-04 | Brian Boesch | System and method for centralized security screening |
JP3785576B2 (en) | 2002-04-24 | 2006-06-14 | 株式会社モリタ製作所 | Subject blur correction means and medical X-ray imaging apparatus using the same |
US6570951B1 (en) | 2002-05-14 | 2003-05-27 | Ge Medical Systems Global Technology Company, Llc | Image space compensation scheme for reducing artifacts |
US7280696B2 (en) * | 2002-05-20 | 2007-10-09 | Simmonds Precision Products, Inc. | Video detection/verification system |
US7174033B2 (en) | 2002-05-22 | 2007-02-06 | A4Vision | Methods and systems for detecting and recognizing an object based on 3D image data |
US6868138B2 (en) | 2002-05-29 | 2005-03-15 | The Regents Of The University Of Michigan | Method, processor and computed tomography (CT) machine for generating images utilizing high and low sensitivity data collected from a flat panel detector having an extended dynamic range |
US7272252B2 (en) | 2002-06-12 | 2007-09-18 | Clarient, Inc. | Automated system for combining bright field and fluorescent microscopy |
IL150251A (en) | 2002-06-16 | 2007-06-03 | Securelogic Ltd | Screening system for objects in transit |
AU2003281427A1 (en) | 2002-07-03 | 2004-01-23 | Security Point Media | Advertising trays for security screening |
US20040247171A1 (en) | 2002-07-26 | 2004-12-09 | Yoshihito Hashimoto | Image processing method for appearance inspection |
USH2110H1 (en) * | 2002-07-30 | 2004-10-05 | The United States Of America As Represented By The Secretary Of The Air Force | Automated security scanning process |
US6938488B2 (en) * | 2002-08-21 | 2005-09-06 | Battelle Memorial Institute | Acoustic inspection device |
DE10339784B4 (en) | 2002-08-28 | 2021-09-16 | Carl Zeiss Meditec Ag | Microscopy system and microscopy method |
US20050153356A1 (en) | 2002-09-09 | 2005-07-14 | Olympus Corporation | Image processing method for biochemical test |
US20040051030A1 (en) | 2002-09-17 | 2004-03-18 | Artur Olszak | Method and apparatus for acquiring images from a multiple axis imaging system |
US7359042B2 (en) | 2002-09-23 | 2008-04-15 | S.T.I. Security Technology Integration Ltd | Inspection system for limited access spaces |
US6961404B2 (en) | 2002-09-26 | 2005-11-01 | Eastman Kodak Company | Method and system for reconstructing an image from projection data acquired by a cone beam computed tomography system |
US6946671B2 (en) | 2002-09-26 | 2005-09-20 | Honeywell Federal Manufacturing & Technologies, Llc | System and method for identifying, reporting, and evaluating presence of substance |
DE10245676B4 (en) | 2002-09-30 | 2008-01-17 | Siemens Ag | Phase-contrast x-ray machine with line focus for creating a phase-contrast image of an object and method for producing the phase-contrast image |
US6970531B2 (en) | 2002-10-07 | 2005-11-29 | General Electric Company | Continuous scan RAD tomosynthesis system and method |
US6940943B2 (en) | 2002-10-07 | 2005-09-06 | General Electric Company | Continuous scan tomosynthesis system and method |
US6906330B2 (en) | 2002-10-22 | 2005-06-14 | Elgems Ltd. | Gamma camera |
US7633518B2 (en) | 2002-10-25 | 2009-12-15 | Quantum Magnetics, Inc. | Object detection portal with video display overlay |
US7289841B2 (en) | 2002-10-25 | 2007-10-30 | Koninklijke Philips Electronics N.V. | Method and apparatus for volumetric cardiac computed tomography imaging |
US6904117B2 (en) | 2002-10-30 | 2005-06-07 | Toshiba Corporation | Tilted gantry helical cone-beam Feldkamp reconstruction for multislice CT |
US6737102B1 (en) | 2002-10-31 | 2004-05-18 | Nordson Corporation | Apparatus and methods for applying viscous material in a pattern onto one or more moving strands |
US7233817B2 (en) | 2002-11-01 | 2007-06-19 | Brian Yen | Apparatus and method for pattern delivery of radiation and biological characteristic analysis |
US7505556B2 (en) * | 2002-11-06 | 2009-03-17 | American Science And Engineering, Inc. | X-ray backscatter detection imaging modules |
JP2004166923A (en) | 2002-11-19 | 2004-06-17 | Canon Inc | X-ray computed tomography imaging apparatus |
JP2004177138A (en) * | 2002-11-25 | 2004-06-24 | Hitachi Ltd | Dangerous object detector and dangerous object detection method |
JP2004177377A (en) | 2002-11-29 | 2004-06-24 | Hitachi Ltd | Inspecting method and inspecting apparatus |
US6738450B1 (en) * | 2002-12-10 | 2004-05-18 | Agilent Technologies, Inc. | System and method for cost-effective classification of an object under inspection |
EP1579202B1 (en) | 2002-12-10 | 2007-04-18 | Commonwealth Scientific And Industrial Research Organisation | Radiographic equipment |
US7634138B2 (en) * | 2002-12-20 | 2009-12-15 | Eastman Kodak Company | Method for generating an image of a detected subject |
US7286634B2 (en) * | 2002-12-23 | 2007-10-23 | Select Technologies, Llc | Method and apparatus for improving baggage screening examination |
ATE341275T1 (en) | 2003-01-20 | 2006-10-15 | Koninkl Philips Electronics Nv | MEDICAL IMAGING DEVICE HAVING MEANS FOR MAINTAINING THE ORIENTATION OF THE DETECTOR AND THE ORIENTATION OF THE PLAYBACK DEVICE |
JP3760918B2 (en) | 2003-01-21 | 2006-03-29 | 株式会社日立製作所 | Security system |
US6735279B1 (en) | 2003-01-21 | 2004-05-11 | University Of Florida | Snapshot backscatter radiography system and protocol |
WO2004065990A1 (en) * | 2003-01-23 | 2004-08-05 | Reveal Imaging Technologies, Inc. | System and method for ct scanning of baggage |
EP1593093B1 (en) | 2003-01-31 | 2006-08-16 | Philips Intellectual Property & Standards GmbH | Method for the reconstruction of three-dimensional objects |
GB2398379A (en) | 2003-02-11 | 2004-08-18 | Qinetiq Ltd | Automated digital image analysis |
EP1597611A1 (en) | 2003-02-13 | 2005-11-23 | Philips Intellectual Property & Standards GmbH | Method and device for examining an object |
US6936828B2 (en) | 2003-02-14 | 2005-08-30 | Honeywell International Inc. | Particle detection system and method |
US7212661B2 (en) * | 2003-02-14 | 2007-05-01 | Ge Medical Systems Information Technologies. Inc. | Image data navigation method and apparatus |
US7103301B2 (en) | 2003-02-18 | 2006-09-05 | Ricoh Company, Ltd. | Image forming apparatus using a contact or a proximity type of charging system including a protection substance on a moveable body to be charged |
CN100339727C (en) * | 2003-02-24 | 2007-09-26 | 皇家飞利浦电子股份有限公司 | Automatic material discrimination by using computer tomography |
US7019650B2 (en) * | 2003-03-03 | 2006-03-28 | Caducys, L.L.C. | Interrogator and interrogation system employing the same |
US7065175B2 (en) | 2003-03-03 | 2006-06-20 | Varian Medical Systems Technologies, Inc. | X-ray diffraction-based scanning system |
US7991454B2 (en) | 2003-03-04 | 2011-08-02 | Academia Sinica | Imaging of biological structures |
EP1599828A1 (en) | 2003-03-06 | 2005-11-30 | Animetrics, Inc. | Viewpoint-invariant image matching and generation of three-dimensional models from two-dimensional imagery |
US6895072B2 (en) | 2003-03-26 | 2005-05-17 | Heimann Systems Corp. | Apparatus and method for non-destructive inspection of material in containers |
US20050190882A1 (en) | 2003-04-04 | 2005-09-01 | Mcguire Edward L. | Multi-spectral x-ray image processing |
JP2004321310A (en) | 2003-04-22 | 2004-11-18 | Canon Inc | Radiation imaging apparatus |
US7369642B2 (en) * | 2003-04-23 | 2008-05-06 | L-3 Communications and Security Detection Systems Inc. | X-ray imaging technique |
US20050058242A1 (en) | 2003-09-15 | 2005-03-17 | Peschmann Kristian R. | Methods and systems for the rapid detection of concealed objects |
US6920198B2 (en) | 2003-05-02 | 2005-07-19 | Ge Medical Systems Global Technology Company, Llc | Methods and apparatus for processing a fluoroscopic image |
US20040225222A1 (en) | 2003-05-08 | 2004-11-11 | Haishan Zeng | Real-time contemporaneous multimodal imaging and spectroscopy uses thereof |
US20040228766A1 (en) | 2003-05-14 | 2004-11-18 | Witty Thomas R. | Point of care diagnostic platform |
JP4245976B2 (en) | 2003-05-16 | 2009-04-02 | オリンパス株式会社 | Ultrasonic image processing device |
US7092485B2 (en) * | 2003-05-27 | 2006-08-15 | Control Screening, Llc | X-ray inspection system for detecting explosives and other contraband |
US7317390B2 (en) | 2003-06-11 | 2008-01-08 | Quantum Magnetics, Inc. | Screening checkpoint for passengers and baggage |
US7119553B2 (en) | 2003-06-11 | 2006-10-10 | Konsulteurope Limited Limited Joint Stock Company | Security scanners with capacitance and magnetic sensor arrays |
US6952163B2 (en) * | 2003-06-11 | 2005-10-04 | Quantum Magnetics, Inc. | Combined systems user interface for centralized monitoring of a screening checkpoint for passengers and baggage |
US20040253660A1 (en) | 2003-06-12 | 2004-12-16 | Giles Scientific, Inc. | Automated microbiological testing apparatus and method |
US6947521B2 (en) | 2003-06-17 | 2005-09-20 | Illinois Institute Of Technology | Imaging method based on attenuation, refraction and ultra-small-angle-scattering of x-rays |
US20060142662A1 (en) | 2003-06-19 | 2006-06-29 | Van Beek Michael C | Analysis apparatus and method comprising auto-focusing means |
US6928141B2 (en) | 2003-06-20 | 2005-08-09 | Rapiscan, Inc. | Relocatable X-ray imaging system and method for inspecting commercial vehicles and cargo containers |
US6888924B2 (en) | 2003-06-25 | 2005-05-03 | General Electric Company | Method, apparatus, and medium for calibration of tomosynthesis system geometry using fiducial markers with non-determined position |
US6950493B2 (en) | 2003-06-25 | 2005-09-27 | Besson Guy M | Dynamic multi-spectral CT imaging |
CN100445767C (en) | 2003-07-08 | 2008-12-24 | 通用电气家园保护有限公司 | Security checkpoint |
US7248749B2 (en) * | 2003-07-29 | 2007-07-24 | Ge Medical Systems Global Technology Company, Llc | Method and apparatus for signal-to-noise ratio dependent image processing |
US7596286B2 (en) | 2003-08-06 | 2009-09-29 | Sony Corporation | Image processing apparatus, image processing system, imaging apparatus and image processing method |
US7366282B2 (en) | 2003-09-15 | 2008-04-29 | Rapiscan Security Products, Inc. | Methods and systems for rapid detection of concealed objects using fluorescence |
US20050117700A1 (en) | 2003-08-08 | 2005-06-02 | Peschmann Kristian R. | Methods and systems for the rapid detection of concealed objects |
US7099432B2 (en) | 2003-08-27 | 2006-08-29 | Matsushita Electric Industrial Co., Ltd. | X-ray inspection apparatus and X-ray inspection method |
WO2005086616A2 (en) | 2003-09-10 | 2005-09-22 | Sommer Jr Edward J | Method and apparatus for improving baggage screening examination |
US20050058350A1 (en) | 2003-09-15 | 2005-03-17 | Lockheed Martin Corporation | System and method for object identification |
US7145148B2 (en) | 2003-09-25 | 2006-12-05 | Alfano Robert R | Systems and methods for non-destructively detecting material abnormalities beneath a coated surface |
US7889113B2 (en) | 2003-10-10 | 2011-02-15 | L-3 Communications Security and Detection Systems Inc. | Mmw contraband screening system |
US7672431B2 (en) | 2003-10-10 | 2010-03-02 | Bayer Schering Pharma Aktiengesellschaft | X-ray arrangement and x-ray contrast process for imaging an object under examination that contains at least one radiopaque element as well as use of the x-ray arrangement |
US7068751B2 (en) * | 2003-10-27 | 2006-06-27 | General Electric Company | System and method of determining a center of mass of an imaging subject for x-ray flux management control |
US20050094856A1 (en) | 2003-11-03 | 2005-05-05 | Warren Scott R. | Systems and methods for detecting target focus and tilt errors during genetic analysis |
US20050128069A1 (en) | 2003-11-12 | 2005-06-16 | Sondre Skatter | System and method for detecting contraband |
EP1689640A2 (en) * | 2003-11-19 | 2006-08-16 | L-3 Communications Security and Detection Systems Corporation | Security system with distributed computing |
US7492931B2 (en) * | 2003-11-26 | 2009-02-17 | Ge Medical Systems Global Technology Company, Llc | Image temporal change detection and display method and apparatus |
US7555144B2 (en) | 2003-12-03 | 2009-06-30 | Centre National De La Recherche Scientifique (Cnrs) | Optical scanning device for liquid biological samples, process of operation and computer program for a computer connected to said device |
US20050133708A1 (en) | 2003-12-18 | 2005-06-23 | General Electric Company | Method and system for three dimensional tomosynthesis imaging |
US7027553B2 (en) | 2003-12-29 | 2006-04-11 | Ge Medical Systems Global Technology Company, Llc | Systems and methods for generating images by using monochromatic x-rays |
US7177389B2 (en) | 2004-01-09 | 2007-02-13 | Adelphi Technology | X-ray tomography and laminography |
US7105828B2 (en) | 2004-02-10 | 2006-09-12 | Ge Medical Systems Global Technology Company, Llc | Hybrid x-ray detector |
US7246745B2 (en) * | 2004-02-27 | 2007-07-24 | Evolution Robotics Retail, Inc. | Method of merchandising for checkout lanes |
US20050207655A1 (en) | 2004-03-22 | 2005-09-22 | Nasreen Chopra | Inspection system and method for providing feedback |
DE102004015806A1 (en) | 2004-03-29 | 2005-10-27 | Smiths Heimann Biometrics Gmbh | Method and device for recording areas of interest of moving objects |
US7142633B2 (en) * | 2004-03-31 | 2006-11-28 | General Electric Company | Enhanced X-ray imaging system and method |
DE102004016586A1 (en) | 2004-03-31 | 2005-11-03 | Siemens Ag | Image reconstruction device for an X-ray device and method for local 3D reconstruction of an object region |
JP2005283527A (en) | 2004-03-31 | 2005-10-13 | Hitachi High-Technologies Corp | Apparatus for detecting foreign substance |
WO2005104728A2 (en) | 2004-04-26 | 2005-11-10 | L-3 Communications Security & Detection Systems, Inc. | User interface for inspection system with isoluminant regions |
EP1741038A4 (en) * | 2004-04-26 | 2011-06-15 | Creo Inc | Systems and methods for comparing documents containing graphic elements |
US20050251397A1 (en) | 2004-05-04 | 2005-11-10 | Lockheed Martin Corporation | Passenger and item tracking with predictive analysis |
US7212113B2 (en) | 2004-05-04 | 2007-05-01 | Lockheed Martin Corporation | Passenger and item tracking with system alerts |
US20050251398A1 (en) | 2004-05-04 | 2005-11-10 | Lockheed Martin Corporation | Threat scanning with pooled operators |
US7356174B2 (en) * | 2004-05-07 | 2008-04-08 | General Electric Company | Contraband detection system and method using variance data |
US7313270B2 (en) | 2004-05-19 | 2007-12-25 | Applied Vision Company, Llc | Vision system and method for process monitoring |
US20050276443A1 (en) | 2004-05-28 | 2005-12-15 | Slamani Mohamed A | Method and apparatus for recognizing an object within an image |
US7285779B2 (en) | 2004-06-21 | 2007-10-23 | Applied Materials Israel, Ltd. | Methods of scanning an object that includes multiple regions of interest using an array of scanning beams |
US7604595B2 (en) | 2004-06-22 | 2009-10-20 | General Electric Company | Method and system for performing real time navigation of ultrasound volumetric data |
DE102004031130A1 (en) * | 2004-06-28 | 2006-01-19 | Yxlon International Security Gmbh | Method for checking a piece of luggage by means of an X-ray diffraction method |
US20080063140A1 (en) | 2004-07-20 | 2008-03-13 | William Awad | System and Method for Detecting the Presence of a Threat in a Package |
US7233682B2 (en) * | 2004-08-02 | 2007-06-19 | Levine Michael C | Security screening system and method |
US7570786B2 (en) | 2004-08-30 | 2009-08-04 | Antoun Ateya | Automatic digital object counting and verification system and associated method |
DE102004042491B4 (en) | 2004-08-31 | 2009-07-09 | Siemens Ag | A method for generating tomographic slice images of an examination subject with at least two angularly offset beams and computed tomography device for performing this method |
US20060064246A1 (en) | 2004-09-20 | 2006-03-23 | Medberry Scott L | Automated Processing of chemical arrays and systems therefore |
US20060065844A1 (en) | 2004-09-30 | 2006-03-30 | Zelakiewicz Scott S | Systems and methods for dynamic optimization of image |
US7330530B2 (en) | 2004-10-04 | 2008-02-12 | Illinois Institute Of Technology | Diffraction enhanced imaging method using a line x-ray source |
WO2006044652A1 (en) | 2004-10-16 | 2006-04-27 | Identix Incorporated | Diffractive imaging system for the reading and analysis of skin topology |
US7356116B2 (en) * | 2004-12-03 | 2008-04-08 | Eg&G Middle East | Container inspection system |
US7583821B2 (en) | 2004-12-21 | 2009-09-01 | Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. | Apparatus for classifying a material by analyzing the material's surface, and related systems and method |
US7729523B2 (en) * | 2004-12-21 | 2010-06-01 | General Electric Company | Method and system for viewing image data |
US20060173268A1 (en) * | 2005-01-28 | 2006-08-03 | General Electric Company | Methods and systems for controlling acquisition of images |
US7308142B2 (en) | 2005-03-10 | 2007-12-11 | Applied Vision Company, Llc | System and methods for rogue can detection |
US20060269135A1 (en) * | 2005-03-15 | 2006-11-30 | Ramsay Thomas E | System and method for identifying objects of interest in image data |
CA2608119A1 (en) | 2005-05-11 | 2006-11-16 | Optosecurity Inc. | Method and system for screening luggage items, cargo containers or persons |
US20060262902A1 (en) | 2005-05-19 | 2006-11-23 | The Regents Of The University Of California | Security X-ray screening system |
US20060282886A1 (en) | 2005-06-09 | 2006-12-14 | Lockheed Martin Corporation | Service oriented security device management network |
US20070168467A1 (en) * | 2006-01-15 | 2007-07-19 | Telesecurity Sciences Inc. | Method and system for providing remote access to baggage scanned images |
US7358733B2 (en) * | 2006-02-28 | 2008-04-15 | Ge Security, Inc. | High performance security inspection system with physically isolated detection sensors |
US7466790B2 (en) * | 2006-03-02 | 2008-12-16 | General Electric Company | Systems and methods for improving a resolution of an image |
US7899232B2 (en) * | 2006-05-11 | 2011-03-01 | Optosecurity Inc. | Method and apparatus for providing threat image projection (TIP) in a luggage screening system, and luggage screening system implementing same |
US7720194B2 (en) * | 2007-02-16 | 2010-05-18 | L-3 Communications Security and Detection Systems Inc. | High throughput baggage inspection system |
US7769132B1 (en) * | 2007-03-13 | 2010-08-03 | L-3 Communications Security And Detection Systems, Inc. | Material analysis based on imaging effective atomic numbers |
-
2005
- 2005-05-11 CA CA002608119A patent/CA2608119A1/en not_active Abandoned
- 2005-05-11 US US11/920,042 patent/US20090174554A1/en not_active Abandoned
- 2005-05-11 WO PCT/CA2005/000716 patent/WO2006119603A1/en active Application Filing
- 2005-05-11 EP EP05742546A patent/EP1886257A1/en not_active Withdrawn
- 2005-11-08 US US11/268,749 patent/US7734102B2/en not_active Expired - Fee Related
-
2006
- 2006-04-20 US US11/407,217 patent/US20070058037A1/en not_active Abandoned
Patent Citations (101)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5283641A (en) * | 1954-12-24 | 1994-02-01 | Lemelson Jerome H | Apparatus and methods for automated analysis |
US4573198A (en) * | 1982-12-27 | 1986-02-25 | Litton Systems, Inc. | Optical image processing/pattern recognition system |
US4725733A (en) * | 1983-07-18 | 1988-02-16 | The United States Of America As Represented By The Secretary Of The Navy | Apparatus and method for remotely detecting the presence of chemical warfare nerve agents in an air-released thermal cloud |
US4637056A (en) * | 1983-10-13 | 1987-01-13 | Battelle Development Corporation | Optical correlator using electronic image preprocessing |
US4722096A (en) * | 1985-03-04 | 1988-01-26 | Heimann Gmbh | Apparatus for transradiating objects on a conveyor path |
US4724543A (en) * | 1985-09-10 | 1988-02-09 | Beckman Research Institute, City Of Hope | Method and apparatus for automatic digital image analysis |
US4795253A (en) * | 1987-04-24 | 1989-01-03 | Mobay Corporation | Remote sensing gas analyzer |
US5079698A (en) * | 1989-05-03 | 1992-01-07 | Advanced Light Imaging Technologies Ltd. | Transillumination method apparatus for the diagnosis of breast tumors and other breast lesions by normalization of an electronic image of the breast |
US5091924A (en) * | 1989-08-09 | 1992-02-25 | Heimann Gmbh | Apparatus for the transillumination of articles with a fan-shaped radiation beam |
US5179581A (en) * | 1989-09-13 | 1993-01-12 | American Science And Engineering, Inc. | Automatic threat detection based on illumination by penetrating radiant energy |
US5181234A (en) * | 1990-08-06 | 1993-01-19 | Irt Corporation | X-ray backscatter detection system |
US5181234B1 (en) * | 1990-08-06 | 2000-01-04 | Rapiscan Security Products Inc | X-ray backscatter detection system |
US5490218A (en) * | 1990-08-10 | 1996-02-06 | Vivid Technologies, Inc. | Device and method for inspection of baggage and other objects |
US5600485A (en) * | 1991-04-23 | 1997-02-04 | Seiko Instruments Inc. | Optical pattern recognition system method of ferroelectric liquid crystal spatial light modulator |
US5483569A (en) * | 1991-10-25 | 1996-01-09 | American Science And Engineering | Inspection system with no intervening belt |
US5595767A (en) * | 1992-03-27 | 1997-01-21 | Universite Joseph Fourier | Three-dimensional energy distribution device |
US5379336A (en) * | 1992-05-21 | 1995-01-03 | Hughes Aircraft Company | Hybridized semiconductor pixel detector arrays for use in digital radiography |
US5379334A (en) * | 1992-07-20 | 1995-01-03 | Heimann Systems Gmbh & Co. Kg | Object testing system |
US5600303A (en) * | 1993-01-15 | 1997-02-04 | Technology International Incorporated | Detection of concealed explosives and contraband |
US6031890A (en) * | 1993-04-05 | 2000-02-29 | Heimann Systems Gmbh & Co. Kg | Monitoring installation for containers and trucks |
US5485312A (en) * | 1993-09-14 | 1996-01-16 | The United States Of America As Represented By The Secretary Of The Air Force | Optical pattern recognition system and method for verifying the authenticity of a person, product or thing |
US5604634A (en) * | 1993-09-20 | 1997-02-18 | The United States Of America As Represented By The Secretary Of The Air Force | All optical nonlinear joint fourier transform correlator |
US5866907A (en) * | 1993-10-12 | 1999-02-02 | Biotraces, Inc. | Ultralow background multiple photon detector |
US5481622A (en) * | 1994-03-01 | 1996-01-02 | Rensselaer Polytechnic Institute | Eye tracking apparatus and method employing grayscale threshold values |
US5493444A (en) * | 1994-04-28 | 1996-02-20 | The United States Of America As Represented By The Secretary Of The Air Force | Photorefractive two-beam coupling nonlinear joint transform correlator |
US5481584A (en) * | 1994-11-23 | 1996-01-02 | Tang; Jihong | Device for material separation using nondestructive inspection imaging |
US5706816A (en) * | 1995-07-17 | 1998-01-13 | Aloka Co., Ltd. | Image processing apparatus and image processing method for use in the image processing apparatus |
US5600700A (en) * | 1995-09-25 | 1997-02-04 | Vivid Technologies, Inc. | Detecting explosives or other contraband by employing transmitted and scattered X-rays |
US6507025B1 (en) * | 1995-10-23 | 2003-01-14 | Science Applications International Corporation | Density detection using real time discrete photon counting for fast moving targets |
US6018562A (en) * | 1995-11-13 | 2000-01-25 | The United States Of America As Represented By The Secretary Of The Army | Apparatus and method for automatic recognition of concealed objects using multiple energy computed tomography |
US5862258A (en) * | 1996-11-04 | 1999-01-19 | The United States Of America As Represented By The Secretary Of The Army | Method for distinguishing between objects using mace filters |
US6502984B2 (en) * | 1997-01-17 | 2003-01-07 | Canon Kabushiki Kaisha | Radiographic apparatus |
US5864598A (en) * | 1997-04-21 | 1999-01-26 | General Electric Company | Methods and apparatus for scanning an object in a computed tomography system |
US6175613B1 (en) * | 1997-07-11 | 2001-01-16 | Ge Medical Systems S.A. | Method for processing a sequence of radiological images of an object |
US6335742B1 (en) * | 1997-07-24 | 2002-01-01 | Ricoh Company, Ltd. | Apparatus for file management and manipulation using graphical displays and textual descriptions |
US5862198A (en) * | 1997-09-30 | 1999-01-19 | Siemens Corporate Research, Inc. | Pre-calculated hitlist for reducing run-time processing of an exact cone beam reconstruction algorithm |
US6188747B1 (en) * | 1998-01-24 | 2001-02-13 | Heimann Systems Gmbh | X-ray generator |
US6175417B1 (en) * | 1998-02-13 | 2001-01-16 | Micron Technology, Inc. | Method and apparatus for detecting defects in the manufacture of an electronic device |
US6011620A (en) * | 1998-04-06 | 2000-01-04 | Northrop Grumman Corporation | Method and apparatus for the automatic inspection of optically transmissive planar objects |
US6526120B1 (en) * | 1998-05-06 | 2003-02-25 | Joseph N. Gray | X-ray flow rate measurement system for materials, including agricultural materials and food products |
US6195413B1 (en) * | 1998-06-12 | 2001-02-27 | Heimann Systems Gmbh | Method and arrangement for detecting X-rays |
US6018561A (en) * | 1998-07-27 | 2000-01-25 | Siemens Corporate Research, Inc. | Mask boundary correction in a cone beam imaging system using simplified filtered backprojection image reconstruction |
US20040037462A1 (en) * | 1998-08-24 | 2004-02-26 | Lewis Meirion F. | Pattern recognition and other inventions |
US20020015475A1 (en) * | 1998-09-11 | 2002-02-07 | Kazuhiro Matsumoto | A grid holding frame, which provides grid information to-ray image processing apparatus |
US6195444B1 (en) * | 1999-01-12 | 2001-02-27 | Analogic Corporation | Apparatus and method for detecting concealed objects in computed tomography data |
US6185272B1 (en) * | 1999-03-15 | 2001-02-06 | Analogic Corporation | Architecture for CT scanning system |
US20050017181A1 (en) * | 1999-04-16 | 2005-01-27 | The Regents Of The University Of Michigan | Method and system for high-speed, 3D imaging of optically-invisible radiation and detector and array of such detectors for use therein |
US20020018199A1 (en) * | 1999-11-04 | 2002-02-14 | Martin Blumenfeld | Imaging of biological samples using electronic light detector |
US6839406B2 (en) * | 1999-11-13 | 2005-01-04 | Smiths Heimann Gmbh | Apparatus and method for detecting items in objects |
US6525331B1 (en) * | 1999-12-01 | 2003-02-25 | Nanyang Technological University | Ball grid array (BGA) package on-line non-contact inspection method and system |
US20020001366A1 (en) * | 2000-03-31 | 2002-01-03 | Toshikazu Tamura | Imaging apparatus, imaging method, and storage medium |
US20030031291A1 (en) * | 2000-04-18 | 2003-02-13 | Yoshimichi Yamamoto | X-ray apparatus |
US20020016546A1 (en) * | 2000-06-22 | 2002-02-07 | Marino Cerofolini | Method and apparatus for ultrasound imaging, particularly for three-dimensional imaging |
US6507278B1 (en) * | 2000-06-28 | 2003-01-14 | Adt Security Services, Inc. | Ingress/egress control system for airport concourses and other access controlled areas |
US6839403B1 (en) * | 2000-07-24 | 2005-01-04 | Rapiscan Security Products (Usa), Inc. | Generation and distribution of annotation overlays of digital X-ray images for security systems |
US20020024016A1 (en) * | 2000-07-28 | 2002-02-28 | Tadao Endo | Photoelectric conversion device, radiation detection apparatus, image processing system, and driving method thereof |
US20020017620A1 (en) * | 2000-08-04 | 2002-02-14 | Nikon Corporation | Surface inspection apparatus |
US20040027127A1 (en) * | 2000-08-22 | 2004-02-12 | Mills Randell L | 4 dimensinal magnetic resonance imaging |
US20030024315A1 (en) * | 2000-08-31 | 2003-02-06 | Harald Merkel | Device, method and system for measuring the distribution of selected properties in a material |
US6837422B1 (en) * | 2000-09-01 | 2005-01-04 | Heimann Systems Gmbh | Service unit for an X-ray examining device |
US7000827B2 (en) * | 2000-09-01 | 2006-02-21 | Heimann Systems Gmbh | Operator unit for an X-ray examining apparatus |
US20060018517A1 (en) * | 2000-09-15 | 2006-01-26 | Canon Kabushiki Kaisha | Image processing methods and apparatus for detecting human eyes, human face, and other objects in an image |
US20060019409A1 (en) * | 2000-10-13 | 2006-01-26 | Chemimage Corporation | Spectroscopic methods for component particle analysis |
US20030036006A1 (en) * | 2001-03-26 | 2003-02-20 | Shipley Company, L.L.C. | Methods for monitoring photoresists |
US20050008119A1 (en) * | 2001-04-03 | 2005-01-13 | L-3 Communications Security And Detections Systems | Remote baggage screening system, software and method |
US20030012420A1 (en) * | 2001-06-12 | 2003-01-16 | Applied Imaging Corporation | Automated scanning method for pathology samples |
US20030031289A1 (en) * | 2001-07-18 | 2003-02-13 | Jiang Hsieh | Methods and apparatus for FOV-dependent aliasing artifact reduction |
US20030023592A1 (en) * | 2001-07-27 | 2003-01-30 | Rapiscan Security Products (Usa), Inc. | Method and system for certifying operators of x-ray inspection systems |
US20030038945A1 (en) * | 2001-08-17 | 2003-02-27 | Bernward Mahner | Method and apparatus for testing objects |
US20050018812A1 (en) * | 2001-12-11 | 2005-01-27 | Wolfs Peter Bas Anton | X-ray examination apparatus and method |
US20040017882A1 (en) * | 2002-03-06 | 2004-01-29 | National Institute Of Advanced Industrial Science And Technology | Oblique view cone beam CT system |
US20040013239A1 (en) * | 2002-03-13 | 2004-01-22 | Breakaway Imaging, Llc | Systems and methods for quasi-simultaneous multi-planar x-ray imaging |
US20060000911A1 (en) * | 2002-05-07 | 2006-01-05 | Amit Stekel | Automatic certification, identification and tracking of remote objects in relative motion |
US20040012853A1 (en) * | 2002-05-13 | 2004-01-22 | Garcia Juan Manuel Bueno | Method and apparatus for imaging using polarimetry and matrix based image reconstruction |
US20040017883A1 (en) * | 2002-07-22 | 2004-01-29 | Tarou Takagi | CT apparatus, CT imaging method and method of providing service using the same |
US20040016271A1 (en) * | 2002-07-23 | 2004-01-29 | Kirti Shah | Portable inspection containers |
US6843599B2 (en) * | 2002-07-23 | 2005-01-18 | Rapiscan, Inc. | Self-contained, portable inspection system and method |
US20060056584A1 (en) * | 2002-07-23 | 2006-03-16 | Bryan Allman | Self-contained mobile inspection system and method |
US20040017888A1 (en) * | 2002-07-24 | 2004-01-29 | Seppi Edward J. | Radiation scanning of objects for contraband |
US20040017935A1 (en) * | 2002-07-25 | 2004-01-29 | Avinash Gopal B. | Temporal image comparison method |
US20040022425A1 (en) * | 2002-07-31 | 2004-02-05 | Ge Medical Systems Global Technology Company Llc | Temporal image comparison method |
US6856272B2 (en) * | 2002-08-28 | 2005-02-15 | Personnel Protection Technoloties Llc | Methods and apparatus for detecting threats in different areas |
US6982643B2 (en) * | 2002-10-08 | 2006-01-03 | Freight Glove Technologies, Llc | Cargo security method and apparatus |
US20050025280A1 (en) * | 2002-12-10 | 2005-02-03 | Robert Schulte | Volumetric 3D x-ray imaging system for baggage inspection including the detection of explosives |
US20060013455A1 (en) * | 2002-12-17 | 2006-01-19 | Qinetiq Limited | Image analysis |
US7164750B2 (en) * | 2003-03-26 | 2007-01-16 | Smiths Detection, Inc. | Non-destructive inspection of material in container |
US20050008203A1 (en) * | 2003-07-09 | 2005-01-13 | General Electric Company | System and method for real-time processing and display of digital medical images |
US20050025350A1 (en) * | 2003-07-28 | 2005-02-03 | Engelbart Roger W. | Systems and method for identifying foreign objects and debris (FOD) and defects during fabrication of a composite structure |
US20050031069A1 (en) * | 2003-08-07 | 2005-02-10 | General Electric Company | System and method for detecting an object by dynamically adjusting computational load |
US20060017605A1 (en) * | 2003-08-12 | 2006-01-26 | Trex Enterprises Corp. | Millimeter wave portal imaging system |
US6990171B2 (en) * | 2003-10-27 | 2006-01-24 | General Electric Company | System and method of determining a user-defined region-of-interest of an imaging subject for x-ray flux management control |
US7882141B2 (en) * | 2004-02-20 | 2011-02-01 | Fujifilm Corporation | Digital pictorial book sytstem, pictorial book searching method, and machine readable medium storing thereon pictorial book searching method |
US7183906B2 (en) * | 2004-03-19 | 2007-02-27 | Lockheed Martin Corporation | Threat scanning machine management system |
US20060013464A1 (en) * | 2004-05-26 | 2006-01-19 | Guardian Technologies International, Inc. | System and method for identifying objects of interest in image data |
US20060002504A1 (en) * | 2004-06-30 | 2006-01-05 | Bruno Kristiaan Bernard De Man | System and method for boundary estimation using CT metrology |
US20060036167A1 (en) * | 2004-07-03 | 2006-02-16 | Shina Systems Ltd. | Vascular image processing |
US20060009269A1 (en) * | 2004-07-08 | 2006-01-12 | Hoskinson Reed L | Method and apparatus for monitoring characteristics of a flow path having solid components flowing therethrough |
US20060008054A1 (en) * | 2004-07-12 | 2006-01-12 | Hiromu Ohara | Radiographic image capturing system and radiographic image capturing method |
US20060018434A1 (en) * | 2004-07-20 | 2006-01-26 | Jacobs Alan M | Radiography by selective detection of scatter field velocity components |
US20060034503A1 (en) * | 2004-08-12 | 2006-02-16 | Fuji Photo Film Co., Ltd. | Medical image processing system |
US20070003122A1 (en) * | 2005-06-29 | 2007-01-04 | General Electric Company | Method for quantifying an object in a larger structure using a reconstructed image |
Cited By (81)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9818249B1 (en) | 2002-09-04 | 2017-11-14 | Copilot Ventures Fund Iii Llc | Authentication method and system |
US8421625B2 (en) | 2004-03-12 | 2013-04-16 | Ingenia Holdings Limited | System and method for article authentication using thumbnail signatures |
US8757493B2 (en) | 2004-03-12 | 2014-06-24 | Ingenia Holdings Limited | System and method for article authentication using encoded signatures |
US8749386B2 (en) | 2004-03-12 | 2014-06-10 | Ingenia Holdings Limited | System and method for article authentication using signatures |
US8896885B2 (en) | 2004-03-12 | 2014-11-25 | Ingenia Holdings Limited | Creating authenticatable printed articles and subsequently verifying them based on scattered light caused by surface structure |
US8766800B2 (en) | 2004-03-12 | 2014-07-01 | Ingenia Holdings Limited | Authenticity verification methods, products, and apparatuses |
US9019567B2 (en) | 2004-03-12 | 2015-04-28 | Ingenia Holdings Limited | Methods and apparatuses for creating authenticatable printed articles and subsequently verifying them |
US8502668B2 (en) | 2004-03-12 | 2013-08-06 | Ingenia Holdings Limited | System and method for article authentication using blanket illumination |
US8699088B2 (en) | 2004-03-12 | 2014-04-15 | Ingenia Holdings Limited | Methods and apparatuses for creating authenticatable printed articles and subsequently verifying them |
US20080260199A1 (en) * | 2004-08-13 | 2008-10-23 | Ingenia Technology Limited | Authenticity Verification Methods, Products and Apparatuses |
US20080294900A1 (en) * | 2004-08-13 | 2008-11-27 | Ingenia Technology Limited | Authenticity Verification of Articles Using a Database |
US8103046B2 (en) * | 2004-08-13 | 2012-01-24 | Ingenia Holdings Limited | Authenticity verification of articles using a database |
US7734102B2 (en) | 2005-05-11 | 2010-06-08 | Optosecurity Inc. | Method and system for screening cargo containers |
US7991242B2 (en) | 2005-05-11 | 2011-08-02 | Optosecurity Inc. | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality |
US20070041612A1 (en) * | 2005-05-11 | 2007-02-22 | Luc Perron | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality |
US20070041613A1 (en) * | 2005-05-11 | 2007-02-22 | Luc Perron | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same |
US20080062262A1 (en) * | 2005-05-11 | 2008-03-13 | Luc Perron | Apparatus, method and system for screening receptacles and persons |
US8078875B2 (en) | 2005-07-27 | 2011-12-13 | Ingenia Holdings Limited | Verification of authenticity |
US20100316251A1 (en) * | 2005-12-23 | 2010-12-16 | Ingenia Holdings Limited | Optical Authentication |
US8497983B2 (en) | 2005-12-23 | 2013-07-30 | Ingenia Holdings Limited | Optical authentication |
US7899232B2 (en) | 2006-05-11 | 2011-03-01 | Optosecurity Inc. | Method and apparatus for providing threat image projection (TIP) in a luggage screening system, and luggage screening system implementing same |
US20090175411A1 (en) * | 2006-07-20 | 2009-07-09 | Dan Gudmundson | Methods and systems for use in security screening, with parallel processing capability |
WO2008019473A1 (en) * | 2006-08-16 | 2008-02-21 | Optosecurity Inc. | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same |
US20080152082A1 (en) * | 2006-08-16 | 2008-06-26 | Michel Bouchard | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same |
US8781066B2 (en) | 2006-09-18 | 2014-07-15 | Optosecurity Inc. | Method and apparatus for assessing characteristics of liquids |
US20100002834A1 (en) * | 2006-09-18 | 2010-01-07 | Optosecurity Inc | Method and apparatus for assessing characteristics of liquids |
US8116428B2 (en) | 2006-09-18 | 2012-02-14 | Optosecurity Inc. | Method and apparatus for assessing characteristics of liquids |
US8009800B2 (en) | 2006-10-02 | 2011-08-30 | Optosecurity Inc. | Tray for assessing the threat status of an article at a security check point |
US8009799B2 (en) | 2006-10-02 | 2011-08-30 | Optosecurity Inc. | Tray for use in assessing the threat status of an article at a security check point |
US20100027741A1 (en) * | 2006-10-02 | 2010-02-04 | Aidan Doyle | Tray for assessing the threat status of an article at a security check point |
US20090196396A1 (en) * | 2006-10-02 | 2009-08-06 | Optosecurity Inc. | Tray for assessing the threat status of an article at a security check point |
WO2008119151A1 (en) * | 2007-03-30 | 2008-10-09 | Optosecurity Inc. | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same |
US8494210B2 (en) | 2007-03-30 | 2013-07-23 | Optosecurity Inc. | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same |
EP2165188A4 (en) * | 2007-06-21 | 2014-01-22 | Rapiscan Systems Inc | Systems and methods for improving directed people screening |
EP2165188A2 (en) * | 2007-06-21 | 2010-03-24 | Rapiscan Security Products, Inc. | Systems and methods for improving directed people screening |
US20110007870A1 (en) * | 2007-10-01 | 2011-01-13 | Optosecurity Inc. | Method and devices for assessing the threat status of an article at a security check point |
US8014493B2 (en) | 2007-10-01 | 2011-09-06 | Optosecurity Inc. | Method and devices for assessing the threat status of an article at a security check point |
US20100207741A1 (en) * | 2007-10-10 | 2010-08-19 | Optosecurity Inc. | Method, apparatus and system for use in connection with the inspection of liquid merchandise |
WO2009114928A1 (en) * | 2008-03-17 | 2009-09-24 | Optosecurity, Inc. | Method and apparatus for assessing characteristics of liquids |
US8867816B2 (en) | 2008-09-05 | 2014-10-21 | Optosecurity Inc. | Method and system for performing X-ray inspection of a liquid product at a security checkpoint |
US20100208972A1 (en) * | 2008-09-05 | 2010-08-19 | Optosecurity Inc. | Method and system for performing x-ray inspection of a liquid product at a security checkpoint |
US9170212B2 (en) | 2008-09-05 | 2015-10-27 | Optosecurity Inc. | Method and system for performing inspection of a liquid product at a security checkpoint |
US20110172972A1 (en) * | 2008-09-15 | 2011-07-14 | Optosecurity Inc. | Method and apparatus for asssessing properties of liquids by using x-rays |
US8682076B2 (en) | 2008-12-19 | 2014-03-25 | Ingenia Holdings Limited | Signature generation for use in authentication and verification using a non-coherent radiation source |
US8615475B2 (en) | 2008-12-19 | 2013-12-24 | Ingenia Holdings Limited | Self-calibration |
US8401309B2 (en) * | 2008-12-30 | 2013-03-19 | International Business Machines Corporation | Security screening image analysis simplification through object pattern identification |
US8903180B2 (en) | 2008-12-30 | 2014-12-02 | International Business Machines Corporation | Security screening image analysis simplification through object pattern identification |
US20100166322A1 (en) * | 2008-12-30 | 2010-07-01 | International Business Machines Corporation | Security Screening Image Analysis Simplification Through Object Pattern Identification |
US8831331B2 (en) | 2009-02-10 | 2014-09-09 | Optosecurity Inc. | Method and system for performing X-ray inspection of a product at a security checkpoint using simulation |
US9297693B2 (en) | 2009-02-25 | 2016-03-29 | The University Of Memphis Research Foundation | Spatially-selective reflector structures, reflector disks, and systems and methods for use thereof |
US8508592B2 (en) | 2009-02-25 | 2013-08-13 | The University Of Memphis Research Foundation | Spatially-selective reflector structures, reflector disks, and systems and methods for use thereof |
US9157873B2 (en) * | 2009-06-15 | 2015-10-13 | Optosecurity, Inc. | Method and apparatus for assessing the threat status of luggage |
US20120093367A1 (en) * | 2009-06-15 | 2012-04-19 | Optosecurity Inc. | Method and apparatus for assessing the threat status of luggage |
US8879791B2 (en) | 2009-07-31 | 2014-11-04 | Optosecurity Inc. | Method, apparatus and system for determining if a piece of luggage contains a liquid product |
US9194975B2 (en) | 2009-07-31 | 2015-11-24 | Optosecurity Inc. | Method and system for identifying a liquid product in luggage or other receptacle |
US8892556B2 (en) | 2009-11-10 | 2014-11-18 | Ingenia Holdings Limited | Optimisation |
US20150181136A1 (en) * | 2011-03-29 | 2015-06-25 | Thermal Matrix USA, Inc. | Method and system for detecting concealed objects using handheld thermal imager |
US10949677B2 (en) | 2011-03-29 | 2021-03-16 | Thermal Matrix USA, Inc. | Method and system for detecting concealed objects using handheld thermal imager |
US8780345B2 (en) | 2011-04-22 | 2014-07-15 | The University Of Memphis Research Foundation | Spatially-selective disks, submillimeter imaging devices, methods of submillimeter imaging, profiling scanners, spectrometry devices, and methods of spectrometry |
US10422919B2 (en) | 2011-09-07 | 2019-09-24 | Rapiscan Systems, Inc. | X-ray inspection system that integrates manifest data with imaging/detection processing |
US9632206B2 (en) | 2011-09-07 | 2017-04-25 | Rapiscan Systems, Inc. | X-ray inspection system that integrates manifest data with imaging/detection processing |
US10830920B2 (en) | 2011-09-07 | 2020-11-10 | Rapiscan Systems, Inc. | Distributed analysis X-ray inspection methods and systems |
US11099294B2 (en) | 2011-09-07 | 2021-08-24 | Rapiscan Systems, Inc. | Distributed analysis x-ray inspection methods and systems |
US10509142B2 (en) | 2011-09-07 | 2019-12-17 | Rapiscan Systems, Inc. | Distributed analysis x-ray inspection methods and systems |
DE102012111201A1 (en) * | 2012-11-21 | 2014-05-22 | Eads Deutschland Gmbh | Sensor system for detecting e.g. weapons hidden in clothing of aircraft passenger in airport, has sensor devices arranged such that monitored area is detected from different viewpoints, and evaluating device evaluating detected data |
DE102012111201B4 (en) * | 2012-11-21 | 2014-07-17 | Eads Deutschland Gmbh | Sensor system and sensor device for it |
US10013750B2 (en) * | 2012-12-27 | 2018-07-03 | Tsinghua University | Object detection methods, display methods and apparatuses |
US11280898B2 (en) | 2014-03-07 | 2022-03-22 | Rapiscan Systems, Inc. | Radar-based baggage and parcel inspection systems |
US9891314B2 (en) | 2014-03-07 | 2018-02-13 | Rapiscan Systems, Inc. | Ultra wide band detectors |
US10134254B2 (en) | 2014-11-25 | 2018-11-20 | Rapiscan Systems, Inc. | Intelligent security management system |
US10713914B2 (en) | 2014-11-25 | 2020-07-14 | Rapiscan Systems, Inc. | Intelligent security management system |
US10302807B2 (en) | 2016-02-22 | 2019-05-28 | Rapiscan Systems, Inc. | Systems and methods for detecting threats and contraband in cargo |
EP3764281A1 (en) * | 2016-02-22 | 2021-01-13 | Rapiscan Systems, Inc. | Methods of identifying firearms in radiographic images |
US10768338B2 (en) | 2016-02-22 | 2020-09-08 | Rapiscan Systems, Inc. | Systems and methods for detecting threats and contraband in cargo |
US11287391B2 (en) | 2016-02-22 | 2022-03-29 | Rapiscan Systems, Inc. | Systems and methods for detecting threats and contraband in cargo |
US11163939B2 (en) * | 2017-12-21 | 2021-11-02 | Anritsu Corporation | Article inspection apparatus |
JP2019111473A (en) * | 2017-12-21 | 2019-07-11 | アンリツインフィビス株式会社 | Article inspection device |
JP7062434B2 (en) | 2017-12-21 | 2022-05-06 | アンリツ株式会社 | Goods inspection equipment |
US20210239875A1 (en) * | 2020-01-30 | 2021-08-05 | Hitachi, Ltd. | Alert output timing control apparatus, alert output timing control method, and non-transitory computer readable storage medium |
WO2022062872A1 (en) * | 2020-09-22 | 2022-03-31 | Kyndryl, Inc. | Identifying a distributed threat in a security zone |
US11538319B2 (en) * | 2020-09-22 | 2022-12-27 | Kyndryl, Inc. | Identifying a distributed threat in a security zone |
Also Published As
Publication number | Publication date |
---|---|
WO2006119603A1 (en) | 2006-11-16 |
US20090174554A1 (en) | 2009-07-09 |
CA2608119A1 (en) | 2006-11-16 |
EP1886257A1 (en) | 2008-02-13 |
US20060257005A1 (en) | 2006-11-16 |
US7734102B2 (en) | 2010-06-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20070058037A1 (en) | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same | |
US20080062262A1 (en) | Apparatus, method and system for screening receptacles and persons | |
US7991242B2 (en) | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality | |
US20070041613A1 (en) | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same | |
US8494210B2 (en) | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same | |
US20080152082A1 (en) | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same | |
US20090175411A1 (en) | Methods and systems for use in security screening, with parallel processing capability | |
EP3349050B1 (en) | Inspection devices and methods for detecting a firearm | |
KR102063859B1 (en) | Systems and methods for security search at airport based on AI and deep learning | |
EP3435287A2 (en) | Model-based digital fingerprinting | |
CN109492714B (en) | Image processing apparatus and method thereof | |
EP2140253B1 (en) | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same | |
EP2016532A1 (en) | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality | |
EP1880331A1 (en) | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same | |
WO2006119605A1 (en) | Method and system for screening cargo containers | |
WO2008019473A1 (en) | Method and apparatus for use in security screening providing incremental display of threat detection information and security system incorporating same | |
Kowkabi et al. | Hybrid preprocessing algorithm for endmember extraction using clustering, over-segmentation, and local entropy criterion | |
KR102158967B1 (en) | Image analysis apparatus, image analysis method and recording medium | |
CA2608121A1 (en) | User interface for use in screening luggage, containers, parcels or people and apparatus for implementing same | |
CA2525997A1 (en) | Method and system for screening containers | |
WO2006119629A1 (en) | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same | |
CA2546296C (en) | Apparatus, method and system for screening receptacles and persons, having image distortion correction functionality | |
CA2608124A1 (en) | Database of target objects suitable for use in screening receptacles or people and method and apparatus for generating same | |
ŞAHİNASLAN et al. | A study on remote detection of Turkey digital identity card hologram element | |
CA2979449C (en) | User interface for use in security screening providing image enhancement capabilities and apparatus for implementing same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: OPTOSECURITY INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BERGERON, ERIC;LACHAPELLE, JEAN;PERRON, LUC;AND OTHERS;REEL/FRAME:017825/0795 Effective date: 20060412 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |