US20140055602A1 - Display System - Google Patents
Display System Download PDFInfo
- Publication number
- US20140055602A1 US20140055602A1 US14/070,757 US201314070757A US2014055602A1 US 20140055602 A1 US20140055602 A1 US 20140055602A1 US 201314070757 A US201314070757 A US 201314070757A US 2014055602 A1 US2014055602 A1 US 2014055602A1
- Authority
- US
- United States
- Prior art keywords
- cameras
- vehicle
- video
- navigation system
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H04N5/23203—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/66—Remote control of cameras or camera parts, e.g. by remote control devices
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/246—Calibration of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- H04N5/247—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
Definitions
- the present embodiments relate generally to navigation features and, more particularly, to a vehicle having navigation features that enables an occupant to view external surroundings on a three-dimensional (“3D”) display.
- 3D three-dimensional
- Today's armored vehicles are frequently manufactured with transparent armor—ballistic (bulletproof) glass windows and/or windshields.
- These structures typically consist of a block or pane of glass or, if not glass, a composite layered structure consisting of special tempered glass laminated with different types of plastics mounted in an appropriately dimensioned frame that overlaps or fits into an opening in the body of an armored vehicle.
- these ballistic windows or windshields provide some amount of threat protection to the vehicle's occupants, they are unwieldy and frequently damaged. Whether it is from a high-velocity projectile or a rock even, these ballistic windshields and/or windows frequently crack, or fail in other ways, which invariably leads to visual distortions and obstructions, and to expensive and time-consuming repairs.
- Some conventional vehicles utilize video cameras to assist in navigation.
- these conventional deployments of cameras as navigation tools are in a secondary role (e.g., a back-up camera), and are not intended to serve as the primary or sole navigation system of vehicle.
- these conventional navigation systems are limited to two-dimension displays. This inhibits the depth perception of the operator, leading to less-efficient, dangerous, or ineffective use of the vehicle while using such secondary navigation systems.
- a navigation system for a vehicle comprises one or more video cameras configured to capture views external to the vehicle, a control mechanism for adjusting the sight-line of the one or more video cameras, and a display for displaying the views captured within the sight-line of the one or more video cameras.
- a navigations system for a vehicle having one or more stereoscopic video cameras and a display that is configured to display three-dimensional images to an occupant of the vehicle.
- a method for providing video navigation for a vehicle involves the steps of receiving a user-input to articulate one or more video cameras, adjusting the sight-line of the one or more video cameras based on the received user input, capturing video images from the one or more video cameras and displaying the video images on a display internal to the vehicle.
- FIG. 1 depicts an exemplary navigation system for a vehicle.
- FIG. 2 depicts a block diagram of an exemplary video camera and display system for use in a navigation system.
- FIG. 2 a depicts a block diagram of an exemplary video camera display system for use in a navigation system.
- FIG. 2 b depicts an exemplary configuration of sets of cameras for use in a navigation system.
- FIG. 3 depicts an exemplary navigation system for a vehicle having a control mechanism.
- FIG. 4 depicts an exemplary navigation system for a vehicle having a multi-panel display.
- FIG. 5 depicts an exemplary navigation system for a vehicle having a stereo monitor three-dimensional display.
- FIG. 6 depicts an exemplary navigation system for a vehicle having a holographic display.
- FIG. 7 depicts a module for replacing a conventional ballistic window or windshield with an armored navigation system for a vehicle.
- FIG. 8 depicts an exemplary navigation system for a vehicle having a video monitor within a recessed portion of the vehicle.
- FIG. 9 depicts an exemplary navigation system for a vehicle having variable separation video cameras.
- the exemplary embodiments may be sized or shaped differently, in any suitable manner, and may be adapted to add components not described, or to remove components.
- One possessing ordinary skill in the art will appreciate the exemplary embodiments' use for purposes and benefits in alternative forms and industries, depending upon specific design needs and other considerations.
- any part that fastens, inserts, attaches, or connects any component to or from another feature or vehicle is not limited to any particular type and is instead intended to encompass all known.
- materials for making components of the present invention may be selected from appropriate materials, such as metal, metal alloys, ballistic metals, ballistic metal alloys, natural or manmade fibers, composites, vinyl, plastics, silicone, rubber, and so on. Any and all appropriate manufacturing or production methods, such as casting, pressing, extruding, molding, machining, may be used to construct the exemplary embodiments or their components.
- any reference to relative position (front and back or rear, top and bottom, right and left, upper and lower, and so on) is intended to conveniently describe those embodiments only.
- Positional and spacial references do not limit the exemplary embodiments or its components to any specific position or orientation.
- the present embodiments disclosed herein are generally designed to provide navigation features for a vehicle, such as an armored vehicle. These features may provide a way for a vehicle occupant to navigate a vehicle without the need for a transparent windshield or other windows. These features may, however, be used in conjunction with a transparent windshield or other windows.
- a vehicle 101 may be equipped with one or more display panels 102 and one or more cameras 106 mounted to or on a vehicle.
- the one or more display panels 102 may be designed to replace in whole or part a transparent windshield normally associated with a vehicle.
- the one or more display panels 102 may be substantially flat, though they need not be substantially flat, and may comprise a video monitor, such as a 120 Hz 3D monitor such as the Samsung type 2233RZ or a pair of monitors coupled with a mirror such as the PLANAR SD2620W StereoMirror® video monitor.
- a stereo monitor 3 d video display 502 may be utilized to provide three-dimensional imaging to the occupant 103 .
- glasses may not be required to create three-dimensional imaging and the stereo monitor alone creates a three-dimensional effect.
- a conventional display panel such as display panel 102 may also provide three-dimensional imaging when an occupant, such as occupant 103 , views the display panel through 3D glasses.
- one or more cameras 106 , display panel 102 , and glasses 105 may be configured to utilize stereoscopic imaging to create a three-dimensional viewing effect for occupant 103 .
- other three-dimensional imaging techniques may also be utilized, including autostereoscopic, computer-generated holography, and volumetric displays.
- the one or more cameras 106 may be designed to zoom in and out on images, either through optical or software zoom capabilities, as needed to afford the vehicle's occupants with views ranging from extremely wide angles to telescopic.
- the one or more displays may be configured and designed to project a holographic display. It should be appreciated that any number of camera configurations may be utilized in accordance with the present navigation system for a vehicle.
- the one or more display panels 102 may be presented to the vehicle's occupant 103 directly. That is, the one or more display panels 102 may be positioned where a typical windshield may be located in a vehicle. Alternatively, in another exemplary embodiment, the one or more display panels 102 may be located on or within the dashboard, with the option of being optically projected to the operator 103 through a reflective device, such as a mirror.
- One or more cameras 106 may be mounted to or formed integrally within the vehicle 101 .
- the one or more cameras 106 may be stereoscopic video cameras.
- the one or more cameras 206 e.g., one or more cameras 106
- the camera sync module 201 may further be connected to a video-display adapter module 204 and/or a LCD-sync module 207 .
- the video-display adapter module 204 may receive, further process, and send information to the one or more display panels 202 that are located within a vehicle, such as vehicle 101 .
- the LCD-sync module 207 may receive, further process, and send information to a viewing apparatus 205 , such as shutter or polarized glasses, worn by the vehicle's occupant.
- a viewing apparatus 205 such as shutter or polarized glasses
- shutter glasses may be used in conjunction with a display screen, such as display screen 102 , to create the illusion of a three dimensional image.
- the navigation features of the present embodiments may use either a video-display adapter module 204 or a LCD-sync module 207 , or both.
- FIG. 2 a depicts a block diagram of an exemplary video camera display system for use in a navigation system.
- One or more sets of cameras such as 206 a, 206 b, and 206 c may be mounted or formed integrally to vehicle 101 .
- the one or more sets of cameras 206 a, 206 b, and 206 c may be stereoscopic cameras located at the front of the vehicle 101 in order to provide a 180 degree field of view toward the front of the vehicle.
- the one or more sets of cameras 206 a, 206 b, and 206 c may communicate with an external camera sync 201 a for ensuring that the sequential frames transmitted from the sets of cameras 206 a, 206 b, and 206 c may be captured at the appropriate times and frequencies.
- the sets of cameras 206 a, 206 b, and 206 c may be in communication with a hub 211 for managing the flow of video frames the sets of cameras 206 a, 206 b, and 206 c on their respective video feed cable (or frequency range on the same cable).
- the hub 211 may collect asynchronous data from the sets of cameras 206 a, 206 b, and 206 c before gating it to a synchronized bus.
- the video frames may be gated to communication bus 210 for facilitating the communication of the video frames through the various components of the system.
- Video feeds may be converted from a “raw” communication bus (e.g., IEEE 1394) to a communication bus configured to use a direct memory access (DMA) adaptor.
- the feeds may then be buffered by image buffer 209 .
- the image buffer 209 may comprise a First-In-First-Out (FIFO) buffering circuit to prevent dropped frames in the event that the bus is temporarily unavailable.
- FIFO First-In-First-Out
- the image buffer 209 may operate to ensure an uninterrupted flow of image data packets in the event of occasional bus unavailability.
- a central processing unit 208 may be utilized to provide a wide variety of image processing services to convert the video frames captured by the sets of cameras 206 a, 206 b, and 206 c into a format suitable for creating stereo renderings for display.
- the sets of cameras 206 a, 206 b, and/or 206 c may also include a processor for converting the captured images into signals suitable for a stereoscopic display.
- the central processing unit 208 may convert the frames captured by the sets of cameras 206 a, 206 b, and/or 206 c by constructing frame buffers using a red/blue anaglyph, line interlacing, or page alternating rendering based on left and right video feeds.
- the central processing unit 208 may be responsible for presenting the processed video data to one or more video-display modules 202 a for receiving and further processing before sending information to one or more display panels, such as display panels 202 a, 202 b, and/or 202 c, which may be located within a vehicle, such as vehicle 101 .
- the central processing unit may also provide additional image processing algorithms and may facilitate operator feedback control of the one or more sets of cameras 206 a, 206 b, and/or 206 c.
- an LCD glasses driver 207 a may be configured to receive, further process, and send display signals to an alternative viewing apparatus, such as liquid display shutter glasses 205 a worn by the vehicles occupant.
- the liquid display shutter glasses 205 a may be used as an alternative to display screens 202 a, 202 b, and/or 202 c.
- appropriate signals may be sent to liquid display shutter glasses 205 a and video displays 202 a, 202 b, and/or 202 c to display a wide variety of images to the vehicles occupant.
- any number of configurations may be utilized to provide the vehicle's occupant with any numbers of desired images.
- FIG. 2 b depicts an exemplary configuration for sets of cameras 206 a, 206 b, and/or 206 c, housed either within, or externally to vehicle 101 .
- set of cameras 206 a may be focused towards a left direction
- set of cameras 206 b may be focused towards a forward direction
- set of cameras 206 c may be focused toward a right direction.
- a central processing unit 208 may work to provide a variety of image processing services as discussed above, to facilitate the display of the images captured by the one or more sets of cameras 206 a, 206 b, and/or 206 c on the appropriate display.
- the images captured by the one or more sets of cameras 206 a, 206 b, and/or 206 c may be displayed on an appropriate portion of a display panel, such as display panel 202 b.
- additional display panels may be positioned coincidental to the gaze of the one or more sets of cameras 206 a, 206 b, and/or 206 c.
- video displays 202 a, 202 b, and/or 202 c may each correspond to a respective one or more sets of cameras 206 a, 206 b, and/or 206 c.
- the one or more cameras 106 may be articulating cameras.
- the one or more cameras 106 may contain elements sensitive to the visible spectrum, infrared (IR) wavelengths, or other non-visible wavelengths, and may contain elements like thermal image sensors.
- the one or more cameras 106 may be mounted outside of the vehicle 101 , within an enclosure 107 , such as cupolas or turrets.
- the enclosure 107 may comprise armor in order to protect the one or more cameras 106 from damage.
- the enclosure 107 may provide articulating movement based on various inputs received from the navigation system 100 .
- the enclosure 107 may also be camouflaged to mask the location of the one or more cameras 106 .
- the one ore more cameras 106 may be miniaturized and may be integrated into the structure or the armor of a vehicle 101 itself. In still another embodiment, the one or more cameras 106 may be mounted inside the vehicle 101 , configured and adapted with fiber optic lenses or other miniaturized and protected optical systems to capture and display external views within the vehicle.
- the articulation of one or more cameras 106 may be controlled by the vehicle's occupant 103 .
- the occupant 103 may wear a sensor-fitted helmet 103 synced to the one or more cameras 106 .
- the external view displayed within the vehicle may be coincidental with the of the occupant's gaze.
- the one or more cameras 106 may be operated via a control mechanism 304 inside the vehicle, such as a joystick.
- other sensors (not shown) may be positioned within the inside of the vehicle to sense the position of the eyes of occupant 103 .
- Such sensors may then adjust the line of sight of one or more cameras 106 to be coincidental to the gaze of occupant 103 .
- the occupant 103 may wear sensor-fitted glasses (not shown) synced to the one or more cameras 106 .
- the views displayed within the vehicle may again be coincidental with the occupant's gaze.
- the one or more cameras 106 may be configured to detect movement and/or heat and to automatically articulate to capture images related to the detected movement and/or heat. It should be appreciated that the one or more cameras 106 may be static, controlled by occupant input, or automatically articulating based on pre-programmed instructions.
- FIG. 4 depicts an exemplary embodiment of a vehicle navigation system having a multi-panel display.
- the occupant 103 may control the articulation of the one or more cameras 106 according to the various configurations described above with regard to FIGS. 1 and 3 .
- the external view displayed within the vehicle 401 may be coincidental with the gaze of occupant 401 .
- the multi-panel display 402 may be configured to comprise three panels—left, middle and right.
- the one or more cameras 106 may capture images in front of the vehicle 401 .
- the images captured by the one or more cameras 106 may be displayed on the middle panel of multi-panel display 402 .
- the one or more cameras 106 may also articulate to the right, thereby capturing images from the right side of vehicle 401 . These images may then be displayed on the right panel of multi-panel display 402 .
- the one or more cameras 106 may in turn articulate to the left, and capture images to the left of vehicle 401 . These images may be displayed on the left panel of multi-panel display 402 .
- the multi-panel display 402 may comprise one or more curved panels that wrap around a portion of the interior 408 of the vehicle 401 . It should be understood that multi-panel display 402 may be configured in any number of ways. For example, multi-panel display 402 may wrap around the entirety of an interior portion 408 of vehicle 103 , thereby creating 360 degrees of panel coverage along the “x-axis” from the perspective of occupant 103 .
- the panels may also extend along the “y-axis,” such that display panels are located at the top, bottom, and corners of the interior 408 of vehicle 401 .
- references to “x-axis,” “y-axis,” “top,” “bottom,” and “corners” are for illustrative purposes only and are not intended to limit the position of one or more display panels within the vehicle 401 .
- Those with skill in the art will readily understand that any number of panels may be utilized in any number of configurations, as desired, within the interior 408 of vehicle 401 .
- the one or more cameras 406 may be static.
- the one or more cameras 406 may comprise multiple static cameras that are continuously directed towards the front, rear, and sides of vehicle 401 .
- the images captured by each cameras (or each pair of cameras) may be displayed on four, corresponding panels (front, rear, and sides) within the interior 408 of vehicle 401 . If the interior 408 of vehicle 401 has only one display, the navigation system may provide a selection mechanism for selecting the direction (front, read, and sides) that would like to have displayed on the display. In another exemplary embodiment, the image displayed on the display panel may be automatically selected based upon the direction of the vehicle.
- the one or more cameras 406 may be static.
- the one or more cameras 406 may comprise multiple static cameras that are continuously directed towards the front, rear, and sides of vehicle 401 .
- the images captured by each cameras (or each pair of cameras) may be displayed on four, corresponding panels (front, rear, and sides) within the interior 408 of vehicle 401 .
- the navigation system may provide a
- FIG. 5 depicts an exemplary navigation system for a vehicle having a stereo monitor three-dimensional display.
- the stereo monitor three-dimensional display 502 may be configured to display images captured by the one or more cameras 106 in three dimensions without the use of conventional three-dimensional glasses.
- Those with skill in the art will readily understand that various types of three-dimensional displays may be deployed within the navigation system depending upon the desired configurations.
- FIG. 6 depicts an exemplary navigation system for a vehicle having a holographic display.
- the navigation system may comprise one or more cameras 606 , one or more projectors 605 , holographic computational engine 610 , mirror 609 , and holographic display screen 602 .
- the one or more cameras 606 may be stereo video cameras for capturing video and passing a series of images to the holographic computational engine 610 .
- the holographic computational engine 610 may be configured to process the video received from the one or more cameras 606 before passing display signals to the one or more projectors 605 .
- the one or more projectors 605 may comprise twin LCD projector, but those with skill in the art will readily understand that any type of projector may be used to generate a holographic display.
- the one or more projectors 605 may project a holographic image onto holographic display screen 602 (for example a Physical Optics Corporation—HDS holographic display).
- a curved viewing zone 608 may be positioned proximate to the holographic display screen 602 to create a holographic visual effect for the occupant 103 .
- the occupant 103 does not need to wear three-dimensional glasses to view the image captured by one or more cameras 606 in three-dimensions.
- FIG. 7 depicts an exemplary modular arrangement of a video navigation system for an armored vehicle.
- the exemplary embodiment depicted in FIG. 7 may include a display panel 702 and one or more cameras 706 .
- the display panel 702 and the one or more cameras 706 may be integrally formed within an armor panel 711 .
- This modular arrangement offers several benefits.
- the modular arrangement allows for more efficient mass-production of the navigation system.
- the armor panel 711 may be configured to replace a conventional ballistic window and/or windshield on an existing vehicle. Because the cost of replacing an existing window on an armored vehicle is much less expensive than manufacturing a new vehicle, the modular arrangement provides for a far more cost-effective option to enable the deployment of the present navigation system in existing armored vehicles.
- FIG. 8 depicts an exemplary navigation system for a vehicle having a video monitor within a recessed portion of the vehicle.
- display panel 804 may be recessed within a portion of vehicle 101 .
- display panel 804 may be recessed within a dashboard portion of vehicle 101 .
- the one or more cameras 806 may be configured to capture images external to the vehicle and pass send them to the display panel 804 .
- a mirror 802 may be utilized by occupant 103 to view the images on display panel 804 .
- a series of mirrors (not shown) may be configured to serve the same function.
- the display panel 804 and mirror 802 may be configured to electronically “mirror” the image viewed by the occupant 103 and the image displayed on the display panel 804 . It should be appreciated that recessing display panel 804 within a portion of vehicle 101 provides the benefit of positioning the display panel 804 in a more protected portion of the armored vehicle.
- the mirror 802 may be a rugged silvered or plated metal to provide for a less fragile viewing surface within an interior portion of the vehicle 101 .
- the one or more cameras 806 are positioned at the top of vehicle 101 . It should be understood that this configuration is exemplary only and that the one or more cameras 806 may be positioned in any number of ways (e.g., as depicted in FIG. 1 ).
- FIG. 9 depicts an exemplary navigation system for a vehicle having variable separation video cameras.
- FIG. 9 depicts a pair of cameras 912 for capturing video images.
- the video images captured by each of the pair of cameras 912 may be synthesized into a stereoscopic display to provide a three-dimensional image in any number of ways (e.g., camera sync 201 ).
- the pair of cameras 912 may be positioned on a rail or other device 914 for facilitating variable separation distance between the pair of cameras 912 .
- the distance between the pair of cameras 912 affects the stereoscopic effect of the resulting image, which in turn affects the three-dimensional image that is displayed.
- variable distance between the pair of cameras 912 is dependent upon the distance of a convergence point 913 from the vehicle 101 .
- adjustments to the angles and distance between the pair of cameras 912 may occur relative to the movement of the vehicle. For instance, when travelling at faster speeds, it may be desirable to have the pair of cameras 912 spaced further apart in order to focus on a far off convergence point 913 . On the other hand, when travelling at slower speeds, it may be desirable to have the pair of cameras 912 move closer together to focus on a convergence point 913 that is closer to the vehicle.
- any number of dynamic configurations may be configured for the pair of cameras 912 based on any number of possible operating positions for a vehicle.
- the angles of the pair of cameras 912 may also be configured to dynamically change based on a wide variety of configurations to suit a wide variety of possible operating conditions or user preferences.
- the distance of the convergence point 913 from the vehicle 101 may be determined in any number of ways.
- the distance of convergence point 913 may be variable based on the velocity of the vehicle 101 .
- the distance of the convergence point 913 from the vehicle may be increased as the velocity of the vehicle 101 increases.
- the variable distance between the pair of cameras 912 may be increased.
- the location of the convergence point 913 from the vehicle 101 may decrease.
- the variable distance between the pair of cameras may also be decreased.
- the line of sight, toe angle, and camera distance are significant for creating various stereoscopic images.
- This invention should not be limited to any particular configuration and it should be appreciated that the each of the line of sight, toe angle, and camera distance may be dynamically modified depending upon the operating condition of the vehicle or preference of a user.
- the distance of the convergence point 913 may be determined based upon the gaze of the occupant 103 .
- the inside of the vehicle 101 may comprise one or more sensors (not shown) for analyzing the movement of the eyes of the occupant 103 .
- Such sensors (not shown) may detect movement of the eye, movement of the head, and/or pupil dilation changes to analyze and determine the gaze of the occupant 103 .
- the sensors (not shown) may be configured to send a signal to the pair of variable cameras 912 to adjust their position and desired convergence point 913 .
- the video images captured by the pair of variable cameras 912 is then processed by the navigation system 100 for display on one or more display panels, such as display panel 102 , which is described above with regard to FIG. 1 .
- the sensors (now shown) may be positioned appropriately throughout the interior of the vehicle 101 . In another exemplary embodiment, the sensors may be embedded within glasses or goggles to be worn by the occupant 103 . Those with skill in the art will readily understand that the configuration of the sensors may be modified in any number of ways.
- the navigation features may be integrated in armored, combat vehicles.
- the features may be combined with other management systems or electronic technologies to increase a crew's situational awareness of its surroundings.
- the navigation features may incorporate global positioning system (“GPS”) technology.
- GPS global positioning system
- This GPS technology may, by way of example only, be superimposed on a 3D anaglyph image displayed within the vehicle to its occupants.
- the GPS technology may further assistant with the navigation of a vehicle by indicating where threatening situations may be present, for example.
- Cueing technology may also be incorporated into the navigation features where a vehicle occupant may be able to detect threats using acoustic or infrared detector signals.
- the navigation features of the present embodiments may be linked to other systems external to the vehicle to transmit the same images seen to the vehicle's occupants.
- the displayed imaged may be simultaneously sent to a commander's position so that it too receives the same information and/or displays.
- the foregoing embodiments are not intended to be limited solely to armored, combat, vehicles.
- the navigation system described herein can be configured for use in many other applications.
- the navigation system described herein can be used in conventional vehicles, vehicles used in mining and agriculture, as well as in plains and sea-based vehicles.
Abstract
A system and method for video navigation is disclosed. The system and method may include one or more video cameras configured to capture views external to the vehicle, a control mechanism for adjusting the sight-line of the one or more video cameras, and a display panel positioned within the vehicle for displaying the views captured within the sight-line of the one or more video cameras.
Description
- This application claims the benefit of priority under 35 U.S.C. §119(e) of U.S. Provisional Application Ser. No. 61/323,749, filed Apr. 13, 2010, the context of which is relied upon and incorporated herein by reference in its entirety.
- The present embodiments relate generally to navigation features and, more particularly, to a vehicle having navigation features that enables an occupant to view external surroundings on a three-dimensional (“3D”) display.
- Today's armored vehicles are frequently manufactured with transparent armor—ballistic (bulletproof) glass windows and/or windshields. These structures typically consist of a block or pane of glass or, if not glass, a composite layered structure consisting of special tempered glass laminated with different types of plastics mounted in an appropriately dimensioned frame that overlaps or fits into an opening in the body of an armored vehicle. Though these ballistic windows or windshields provide some amount of threat protection to the vehicle's occupants, they are unwieldy and frequently damaged. Whether it is from a high-velocity projectile or a rock even, these ballistic windshields and/or windows frequently crack, or fail in other ways, which invariably leads to visual distortions and obstructions, and to expensive and time-consuming repairs. These and other problems exist with conventional ballistic glass windows and/or windshields.
- Some conventional vehicles utilize video cameras to assist in navigation. However, these conventional deployments of cameras as navigation tools are in a secondary role (e.g., a back-up camera), and are not intended to serve as the primary or sole navigation system of vehicle. Moreover, these conventional navigation systems are limited to two-dimension displays. This inhibits the depth perception of the operator, leading to less-efficient, dangerous, or ineffective use of the vehicle while using such secondary navigation systems.
- Notwithstanding the number of navigation systems currently in existence, there exists a need for a camera-based navigation system capable of serving as the primary navigation system for a vehicle.
- The present disclosure provides a number of exemplary embodiments that may be used collectively, in various combinations, or alone. The following summary provides examples of such inventions, and does not limit the invention as claimed in any way.
- In one exemplary embodiment, a navigation system for a vehicle is provided. The navigation system comprises one or more video cameras configured to capture views external to the vehicle, a control mechanism for adjusting the sight-line of the one or more video cameras, and a display for displaying the views captured within the sight-line of the one or more video cameras.
- In another exemplary embodiment, a navigations system for a vehicle is provided having one or more stereoscopic video cameras and a display that is configured to display three-dimensional images to an occupant of the vehicle.
- In yet another exemplary embodiment, a method for providing video navigation for a vehicle is provided. The method involves the steps of receiving a user-input to articulate one or more video cameras, adjusting the sight-line of the one or more video cameras based on the received user input, capturing video images from the one or more video cameras and displaying the video images on a display internal to the vehicle.
- Other uses and variations of the foregoing will be apparent to one of ordinary skill in the art.
- The embodiments set forth in the drawings are illustrative and exemplary in nature and not intended to limit the subject matter defined by the claims.
- The invention, both as to its structure and operation together with the additional objects and advantages thereof are best understood through the following description of exemplary embodiments of the present invention when read in conjunction with the accompanying drawings, wherein like structure is indicated with like reference numerals and in which:
-
FIG. 1 depicts an exemplary navigation system for a vehicle. -
FIG. 2 depicts a block diagram of an exemplary video camera and display system for use in a navigation system. -
FIG. 2 a depicts a block diagram of an exemplary video camera display system for use in a navigation system. -
FIG. 2 b depicts an exemplary configuration of sets of cameras for use in a navigation system. -
FIG. 3 depicts an exemplary navigation system for a vehicle having a control mechanism. -
FIG. 4 depicts an exemplary navigation system for a vehicle having a multi-panel display. -
FIG. 5 depicts an exemplary navigation system for a vehicle having a stereo monitor three-dimensional display. -
FIG. 6 depicts an exemplary navigation system for a vehicle having a holographic display. -
FIG. 7 depicts a module for replacing a conventional ballistic window or windshield with an armored navigation system for a vehicle. -
FIG. 8 depicts an exemplary navigation system for a vehicle having a video monitor within a recessed portion of the vehicle. -
FIG. 9 depicts an exemplary navigation system for a vehicle having variable separation video cameras. - The following description conveys an understanding of embodiments that relate generally to vehicles, such as armored vehicles, and more particularly to vehicles having navigation features that may eliminate the need for one or more ballistic window panes. Numerous exemplary embodiments of vehicles having navigation features are described below. Armored vehicles, and other wheeled and non-wheeled vehicles, described by the exemplary embodiments with these navigation features are not limited to only those embodiments, however. For example, exemplary embodiments may be used for any vehicle or other machine or device, including non-military vehicles, machines, or devices.
- The exemplary embodiments may be sized or shaped differently, in any suitable manner, and may be adapted to add components not described, or to remove components. One possessing ordinary skill in the art will appreciate the exemplary embodiments' use for purposes and benefits in alternative forms and industries, depending upon specific design needs and other considerations.
- Terminology used in this description is for describing particular embodiments only and is not intended to limit the scope of an exemplary embodiment. As used throughout this disclosure, the singular forms “a,” “an,” and “the” include the plural, unless the context clearly dictates otherwise. Thus, for example, a reference to “a panel” may encompass a plurality of panels, or other equivalents or variations known to those skilled in the art. Furthermore, permissive language (e.g., “may”) is used for describing some embodiments or features, such use does not suggest that embodiments or features described using other language (e.g., “is,” “are”) are required. Unless defined otherwise, all terms have the same commonly understood meaning that one of ordinary skill in the art to which this invention belongs would expect them to have.
- With regard to the exemplary embodiments of the navigation features, any part that fastens, inserts, attaches, or connects any component to or from another feature or vehicle is not limited to any particular type and is instead intended to encompass all known. Unless otherwise specifically disclosed, materials for making components of the present invention may be selected from appropriate materials, such as metal, metal alloys, ballistic metals, ballistic metal alloys, natural or manmade fibers, composites, vinyl, plastics, silicone, rubber, and so on. Any and all appropriate manufacturing or production methods, such as casting, pressing, extruding, molding, machining, may be used to construct the exemplary embodiments or their components.
- When describing exemplary embodiments, any reference to relative position (front and back or rear, top and bottom, right and left, upper and lower, and so on) is intended to conveniently describe those embodiments only. Positional and spacial references do not limit the exemplary embodiments or its components to any specific position or orientation.
- The present embodiments disclosed herein are generally designed to provide navigation features for a vehicle, such as an armored vehicle. These features may provide a way for a vehicle occupant to navigate a vehicle without the need for a transparent windshield or other windows. These features may, however, be used in conjunction with a transparent windshield or other windows.
- As shown in
FIG. 1 , an exemplary embodiment of navigation features for use in avehicle 101 is shown. These features may collectively and generally be referred to as avideo system 100. In this exemplary embodiment, avehicle 101 may be equipped with one ormore display panels 102 and one ormore cameras 106 mounted to or on a vehicle. The one ormore display panels 102 may be designed to replace in whole or part a transparent windshield normally associated with a vehicle. The one ormore display panels 102 may be substantially flat, though they need not be substantially flat, and may comprise a video monitor, such as a 120 Hz 3D monitor such as the Samsung type 2233RZ or a pair of monitors coupled with a mirror such as the PLANAR SD2620W StereoMirror® video monitor. As depicted inFIG. 5 , a stereo monitor 3d video display 502 may be utilized to provide three-dimensional imaging to theoccupant 103. In the exemplary embodiment depicted inFIG. 5 , glasses may not be required to create three-dimensional imaging and the stereo monitor alone creates a three-dimensional effect. On the other hand, as depicted inFIG. 1 , a conventional display panel, such asdisplay panel 102 may also provide three-dimensional imaging when an occupant, such asoccupant 103, views the display panel through 3D glasses. Those with skill in the art will readily understand that one ormore cameras 106,display panel 102, andglasses 105 may be configured to utilize stereoscopic imaging to create a three-dimensional viewing effect foroccupant 103. Of course, other three-dimensional imaging techniques may also be utilized, including autostereoscopic, computer-generated holography, and volumetric displays. - The one or
more cameras 106 may be designed to zoom in and out on images, either through optical or software zoom capabilities, as needed to afford the vehicle's occupants with views ranging from extremely wide angles to telescopic. Furthermore, the one or more displays may be configured and designed to project a holographic display. It should be appreciated that any number of camera configurations may be utilized in accordance with the present navigation system for a vehicle. - In an exemplary embodiment the one or
more display panels 102 may be presented to the vehicle'soccupant 103 directly. That is, the one ormore display panels 102 may be positioned where a typical windshield may be located in a vehicle. Alternatively, in another exemplary embodiment, the one ormore display panels 102 may be located on or within the dashboard, with the option of being optically projected to theoperator 103 through a reflective device, such as a mirror. - One or
more cameras 106 may be mounted to or formed integrally within thevehicle 101. In an exemplary embodiment, the one ormore cameras 106 may be stereoscopic video cameras. With reference toFIG. 2 , the one or more cameras 206 (e.g., one or more cameras 106) may communicate with a camera-sync module 201 for processing the images and information captured and sent by the one ormore cameras 106. Thecamera sync module 201 may further be connected to a video-display adapter module 204 and/or a LCD-sync module 207. The video-display adapter module 204 may receive, further process, and send information to the one ormore display panels 202 that are located within a vehicle, such asvehicle 101. In an exemplary embodiment, the LCD-sync module 207 may receive, further process, and send information to aviewing apparatus 205, such as shutter or polarized glasses, worn by the vehicle's occupant. In an exemplary embodiment, shutter glasses may be used in conjunction with a display screen, such asdisplay screen 102, to create the illusion of a three dimensional image. It should be understood that the navigation features of the present embodiments may use either a video-display adapter module 204 or a LCD-sync module 207, or both. -
FIG. 2 a depicts a block diagram of an exemplary video camera display system for use in a navigation system. One or more sets of cameras, such as 206 a, 206 b, and 206 c may be mounted or formed integrally tovehicle 101. As described above, the one or more sets ofcameras vehicle 101 in order to provide a 180 degree field of view toward the front of the vehicle. In an exemplary embodiment, the one or more sets ofcameras external camera sync 201 a for ensuring that the sequential frames transmitted from the sets ofcameras cameras hub 211 for managing the flow of video frames the sets ofcameras hub 211 may collect asynchronous data from the sets ofcameras communication bus 210 for facilitating the communication of the video frames through the various components of the system. Video feeds may be converted from a “raw” communication bus (e.g., IEEE 1394) to a communication bus configured to use a direct memory access (DMA) adaptor. The feeds may then be buffered byimage buffer 209. In an exemplary embodiment, theimage buffer 209 may comprise a First-In-First-Out (FIFO) buffering circuit to prevent dropped frames in the event that the bus is temporarily unavailable. Theimage buffer 209 may operate to ensure an uninterrupted flow of image data packets in the event of occasional bus unavailability. Acentral processing unit 208 may be utilized to provide a wide variety of image processing services to convert the video frames captured by the sets ofcameras cameras central processing unit 208 may convert the frames captured by the sets ofcameras central processing unit 208 may be responsible for presenting the processed video data to one or more video-display modules 202 a for receiving and further processing before sending information to one or more display panels, such asdisplay panels vehicle 101. - The central processing unit may also provide additional image processing algorithms and may facilitate operator feedback control of the one or more sets of
cameras LCD glasses driver 207 a may be configured to receive, further process, and send display signals to an alternative viewing apparatus, such as liquiddisplay shutter glasses 205 a worn by the vehicles occupant. In an exemplary embodiment, the liquiddisplay shutter glasses 205 a may be used as an alternative to displayscreens display shutter glasses 205 a andvideo displays -
FIG. 2 b depicts an exemplary configuration for sets ofcameras vehicle 101. In the exemplary embodiment depicted inFIG. 2 b, set ofcameras 206 a may be focused towards a left direction, set ofcameras 206 b may be focused towards a forward direction, and set ofcameras 206 c may be focused toward a right direction. Acentral processing unit 208 may work to provide a variety of image processing services as discussed above, to facilitate the display of the images captured by the one or more sets ofcameras cameras display panel 202 b. In another exemplary embodiment, additional display panels may be positioned coincidental to the gaze of the one or more sets ofcameras cameras - Referring back to
FIG. 1 , the one ormore cameras 106 may be articulating cameras. The one ormore cameras 106 may contain elements sensitive to the visible spectrum, infrared (IR) wavelengths, or other non-visible wavelengths, and may contain elements like thermal image sensors. In an exemplary embodiment, the one ormore cameras 106 may be mounted outside of thevehicle 101, within anenclosure 107, such as cupolas or turrets. Theenclosure 107 may comprise armor in order to protect the one ormore cameras 106 from damage. In another exemplary embodiment theenclosure 107 may provide articulating movement based on various inputs received from thenavigation system 100. Theenclosure 107 may also be camouflaged to mask the location of the one ormore cameras 106. In other embodiments, the one oremore cameras 106 may be miniaturized and may be integrated into the structure or the armor of avehicle 101 itself. In still another embodiment, the one ormore cameras 106 may be mounted inside thevehicle 101, configured and adapted with fiber optic lenses or other miniaturized and protected optical systems to capture and display external views within the vehicle. - In an exemplary embodiment, the articulation of one or
more cameras 106 may be controlled by the vehicle'soccupant 103. For example, theoccupant 103 may wear a sensor-fittedhelmet 103 synced to the one ormore cameras 106. In this exemplary embodiment, the external view displayed within the vehicle may be coincidental with the of the occupant's gaze. As shown inFIG. 3 , the one ormore cameras 106 may be operated via acontrol mechanism 304 inside the vehicle, such as a joystick. In other exemplary embodiments, other sensors (not shown) may be positioned within the inside of the vehicle to sense the position of the eyes ofoccupant 103. Such sensors may then adjust the line of sight of one ormore cameras 106 to be coincidental to the gaze ofoccupant 103. In another exemplary embodiment, theoccupant 103 may wear sensor-fitted glasses (not shown) synced to the one ormore cameras 106. In this exemplary embodiment, the views displayed within the vehicle may again be coincidental with the occupant's gaze. Those with skill in the art will readily understand that there are numerous other configurations that may be employed to provide theoccupant 103 with control over the articulation of one ormore cameras 106. Of course, in other exemplary embodiments, the one ormore cameras 106 may be configured to detect movement and/or heat and to automatically articulate to capture images related to the detected movement and/or heat. It should be appreciated that the one ormore cameras 106 may be static, controlled by occupant input, or automatically articulating based on pre-programmed instructions. -
FIG. 4 depicts an exemplary embodiment of a vehicle navigation system having a multi-panel display. Theoccupant 103 may control the articulation of the one ormore cameras 106 according to the various configurations described above with regard toFIGS. 1 and 3 . In an exemplary embodiment, the external view displayed within thevehicle 401 may be coincidental with the gaze ofoccupant 401. For example, themulti-panel display 402 may be configured to comprise three panels—left, middle and right. When theoccupant 103 gazes forward, the one ormore cameras 106 may capture images in front of thevehicle 401. The images captured by the one ormore cameras 106 may be displayed on the middle panel ofmulti-panel display 402. When the occupant 403 gazes, for example, to the right, the one ormore cameras 106 may also articulate to the right, thereby capturing images from the right side ofvehicle 401. These images may then be displayed on the right panel ofmulti-panel display 402. Of course, when occupant 403 gazes to the left, the one ormore cameras 106 may in turn articulate to the left, and capture images to the left ofvehicle 401. These images may be displayed on the left panel ofmulti-panel display 402. - It should be understood that the foregoing example is exemplary only and not intended to be limited to a multi-panel display having only a left, right, and middle panel. In another exemplary embodiment, the
multi-panel display 402 may comprise one or more curved panels that wrap around a portion of theinterior 408 of thevehicle 401. It should be understood thatmulti-panel display 402 may be configured in any number of ways. For example,multi-panel display 402 may wrap around the entirety of aninterior portion 408 ofvehicle 103, thereby creating 360 degrees of panel coverage along the “x-axis” from the perspective ofoccupant 103. In another exemplary embodiment, the panels may also extend along the “y-axis,” such that display panels are located at the top, bottom, and corners of theinterior 408 ofvehicle 401. It should be appreciated that references to “x-axis,” “y-axis,” “top,” “bottom,” and “corners” are for illustrative purposes only and are not intended to limit the position of one or more display panels within thevehicle 401. Those with skill in the art will readily understand that any number of panels may be utilized in any number of configurations, as desired, within theinterior 408 ofvehicle 401. - In another exemplary embodiment, the one or more cameras 406 may be static. For example, the one or more cameras 406 may comprise multiple static cameras that are continuously directed towards the front, rear, and sides of
vehicle 401. In this exemplary embodiment, the images captured by each cameras (or each pair of cameras) may be displayed on four, corresponding panels (front, rear, and sides) within theinterior 408 ofvehicle 401. If theinterior 408 ofvehicle 401 has only one display, the navigation system may provide a selection mechanism for selecting the direction (front, read, and sides) that would like to have displayed on the display. In another exemplary embodiment, the image displayed on the display panel may be automatically selected based upon the direction of the vehicle. Those with skill in the art will readily understand that any number of cameras may be utilized in conjunction with any number of display panels. -
FIG. 5 depicts an exemplary navigation system for a vehicle having a stereo monitor three-dimensional display. The stereo monitor three-dimensional display 502 may be configured to display images captured by the one ormore cameras 106 in three dimensions without the use of conventional three-dimensional glasses. Those with skill in the art will readily understand that various types of three-dimensional displays may be deployed within the navigation system depending upon the desired configurations. -
FIG. 6 depicts an exemplary navigation system for a vehicle having a holographic display. In accordance with an exemplary embodiment, the navigation system may comprise one ormore cameras 606, one ormore projectors 605, holographiccomputational engine 610,mirror 609, andholographic display screen 602. The one ormore cameras 606 may be stereo video cameras for capturing video and passing a series of images to the holographiccomputational engine 610. The holographiccomputational engine 610 may be configured to process the video received from the one ormore cameras 606 before passing display signals to the one ormore projectors 605. In one exemplary embodiment, the one ormore projectors 605 may comprise twin LCD projector, but those with skill in the art will readily understand that any type of projector may be used to generate a holographic display. - Still referring to
FIG. 6 , the one ormore projectors 605 may project a holographic image onto holographic display screen 602 (for example a Physical Optics Corporation—HDS holographic display). Acurved viewing zone 608 may be positioned proximate to theholographic display screen 602 to create a holographic visual effect for theoccupant 103. According to this exemplary embodiment, theoccupant 103 does not need to wear three-dimensional glasses to view the image captured by one ormore cameras 606 in three-dimensions. -
FIG. 7 depicts an exemplary modular arrangement of a video navigation system for an armored vehicle. The exemplary embodiment depicted inFIG. 7 may include adisplay panel 702 and one ormore cameras 706. Thedisplay panel 702 and the one ormore cameras 706 may be integrally formed within an armor panel 711. This modular arrangement offers several benefits. For example, the modular arrangement allows for more efficient mass-production of the navigation system. Furthermore, the armor panel 711 may be configured to replace a conventional ballistic window and/or windshield on an existing vehicle. Because the cost of replacing an existing window on an armored vehicle is much less expensive than manufacturing a new vehicle, the modular arrangement provides for a far more cost-effective option to enable the deployment of the present navigation system in existing armored vehicles. -
FIG. 8 depicts an exemplary navigation system for a vehicle having a video monitor within a recessed portion of the vehicle. According to the exemplary embodiment depicted inFIG. 8 ,display panel 804 may be recessed within a portion ofvehicle 101. In one exemplary embodiment,display panel 804 may be recessed within a dashboard portion ofvehicle 101. The one ormore cameras 806 may be configured to capture images external to the vehicle and pass send them to thedisplay panel 804. Because, in this exemplary embodiment,display panel 804 is recessed within a portion ofvehicle 101, amirror 802 may be utilized byoccupant 103 to view the images ondisplay panel 804. In the alternative, a series of mirrors (not shown) may be configured to serve the same function. In any event, thedisplay panel 804 and mirror 802 (or series of mirrors) may be configured to electronically “mirror” the image viewed by theoccupant 103 and the image displayed on thedisplay panel 804. It should be appreciated that recessingdisplay panel 804 within a portion ofvehicle 101 provides the benefit of positioning thedisplay panel 804 in a more protected portion of the armored vehicle. Themirror 802 may be a rugged silvered or plated metal to provide for a less fragile viewing surface within an interior portion of thevehicle 101. Still referring toFIG. 8 , the one ormore cameras 806 are positioned at the top ofvehicle 101. It should be understood that this configuration is exemplary only and that the one ormore cameras 806 may be positioned in any number of ways (e.g., as depicted inFIG. 1 ). -
FIG. 9 depicts an exemplary navigation system for a vehicle having variable separation video cameras. In accordance with an exemplary embodiment,FIG. 9 depicts a pair ofcameras 912 for capturing video images. The video images captured by each of the pair ofcameras 912 may be synthesized into a stereoscopic display to provide a three-dimensional image in any number of ways (e.g., camera sync 201). Importantly, the pair ofcameras 912 may be positioned on a rail orother device 914 for facilitating variable separation distance between the pair ofcameras 912. In accordance with this exemplary embodiment, the distance between the pair ofcameras 912 affects the stereoscopic effect of the resulting image, which in turn affects the three-dimensional image that is displayed. Therefore, the variable distance between the pair ofcameras 912 is dependent upon the distance of aconvergence point 913 from thevehicle 101. In an exemplary embodiment, adjustments to the angles and distance between the pair ofcameras 912 may occur relative to the movement of the vehicle. For instance, when travelling at faster speeds, it may be desirable to have the pair ofcameras 912 spaced further apart in order to focus on a far offconvergence point 913. On the other hand, when travelling at slower speeds, it may be desirable to have the pair ofcameras 912 move closer together to focus on aconvergence point 913 that is closer to the vehicle. Those with skill in the art will appreciate that any number of dynamic configurations may be configured for the pair ofcameras 912 based on any number of possible operating positions for a vehicle. Similarly, the angles of the pair ofcameras 912 may also be configured to dynamically change based on a wide variety of configurations to suit a wide variety of possible operating conditions or user preferences. - The distance of the
convergence point 913 from thevehicle 101 may be determined in any number of ways. In accordance with one exemplary embodiment, the distance ofconvergence point 913 may be variable based on the velocity of thevehicle 101. As discussed above, the distance of theconvergence point 913 from the vehicle may be increased as the velocity of thevehicle 101 increases. In accordance with this exemplary embodiment, the variable distance between the pair ofcameras 912 may be increased. On the other hand, as the velocity of thevehicle 101 decreases, the location of theconvergence point 913 from thevehicle 101 may decrease. In turn, the variable distance between the pair of cameras may also be decreased. In accordance with this exemplary embodiment, it may be desirable to increase the distance of theconvergence point 913 from thevehicle 101 at higher velocities because theoccupant 103 may desire to view objects further away fromvehicle 101 than when travelling at slower speeds. It should be appreciated that the combination and various configurations of the line of sight, toe angle, and camera distance are significant for creating various stereoscopic images. This invention should not be limited to any particular configuration and it should be appreciated that the each of the line of sight, toe angle, and camera distance may be dynamically modified depending upon the operating condition of the vehicle or preference of a user. - Still referring to
FIG. 9 , in accordance with another exemplary embodiment, the distance of the convergence point 913 (and in turn the variable distance between the pair of cameras 912) may be determined based upon the gaze of theoccupant 103. The inside of thevehicle 101 may comprise one or more sensors (not shown) for analyzing the movement of the eyes of theoccupant 103. Such sensors (not shown) may detect movement of the eye, movement of the head, and/or pupil dilation changes to analyze and determine the gaze of theoccupant 103. The sensors (not shown) may be configured to send a signal to the pair ofvariable cameras 912 to adjust their position and desiredconvergence point 913. The video images captured by the pair ofvariable cameras 912 is then processed by thenavigation system 100 for display on one or more display panels, such asdisplay panel 102, which is described above with regard toFIG. 1 . The sensors (now shown) may be positioned appropriately throughout the interior of thevehicle 101. In another exemplary embodiment, the sensors may be embedded within glasses or goggles to be worn by theoccupant 103. Those with skill in the art will readily understand that the configuration of the sensors may be modified in any number of ways. - In other embodiments of the present disclosure, the navigation features may be integrated in armored, combat vehicles. The features may be combined with other management systems or electronic technologies to increase a crew's situational awareness of its surroundings. For example, the navigation features may incorporate global positioning system (“GPS”) technology. This GPS technology may, by way of example only, be superimposed on a 3D anaglyph image displayed within the vehicle to its occupants. The GPS technology may further assistant with the navigation of a vehicle by indicating where threatening situations may be present, for example. Cueing technology may also be incorporated into the navigation features where a vehicle occupant may be able to detect threats using acoustic or infrared detector signals. These and other technologies may be incorporated with the navigation features to enhance the threat-level protection features of a vehicle. In addition, the navigation features of the present embodiments may be linked to other systems external to the vehicle to transmit the same images seen to the vehicle's occupants. For example, in the case of an armored combat vehicle, the displayed imaged may be simultaneously sent to a commander's position so that it too receives the same information and/or displays.
- It should be appreciated that the foregoing embodiments are not intended to be limited solely to armored, combat, vehicles. The navigation system described herein can be configured for use in many other applications. By way of non-limiting example, the navigation system described herein can be used in conventional vehicles, vehicles used in mining and agriculture, as well as in plains and sea-based vehicles.
- It will be readily understood by those persons skilled in the art that embodiments of the present inventions are susceptible to broad utility and application. Many embodiments and adaptations of the present inventions, other than those herein described, as well as many variations, modifications and equivalent arrangements, will be apparent from or reasonably suggested by the present invention and foregoing descriptions thereof, without departing from the substance or scope of the invention.
- Accordingly, it is to be understood that this disclosure is only illustrative and exemplary and is made to provide an enabling disclosure. Accordingly, the foregoing disclosure is not intended to be construed to limit the present invention or otherwise to exclude any other such embodiments, adaptations, variations, modifications or equivalent arrangements.
- While particular embodiments have been illustrated and described herein, it should be understood that various other changes and modifications may be made without departing from the spirit and scope of the claimed subject matter. Moreover, although various aspects of the claimed subject matter have been described herein, such aspects need not be utilized in combinations. It is therefore intended that the appended claims cover all such changes and modifications that are within the scope of the claimed subject matter.
Claims (18)
1-30. (canceled)
31. A navigation system for a vehicle, comprising:
one or more pairs of video cameras, each pair having a first camera with a first sight-line and a second camera with a second sight-line, the second camera being spaced apart from the first camera by a selectively variable separation distance;
a control mechanism for adjusting the first and second sight-lines, the control mechanism being configured to determine a desired convergence point and to adjust the first and second sight lines and the separation distance so that the first and second sight-lines intersect at the desired convergence point; and
a display panel positioned within the vehicle for displaying the views captured by the one or more pairs of video cameras.
32. A navigation system according to claim 31 further comprising:
one or more sensors for determining a direction of a vehicle occupant's gaze.
33. A navigation system according to claim 32 , wherein the control mechanism adjusts the first and second sight-lines of at least one of the one or more pairs of cameras using the determined direction of the occupant's gaze.
34. A navigation system according to claim 32 , wherein the one or more sensors includes at least one of the set consisting of a sensor integrated into a helmet wearable by a vehicle occupant and a sensor integrated into eyeglasses wearable by a vehicle occupant.
35. A navigation system according to claim 31 , wherein the control mechanism is configured to determine the desired convergence point using at least one of the set consisting of a direction of a vehicle occupant's gaze and a velocity of the vehicle.
36. A navigation system according to claim 31 , wherein the control mechanism is configured to determine the desired convergence point using a direction of a vehicle occupant's gaze and a velocity of the vehicle.
37. A navigation system according to claim 31 , further comprising:
a power supply;
a camera sync module; and
a video display adapter for sending video captured by the one or more cameras to the display panel.
38. A navigation system according to claim 37 , wherein the video display adapter is an LCD-sync module or LCD shutter glasses system.
39. A navigation system according to claim 31 , wherein he one or more video camera pairs are mounted to an exterior surface of the vehicle.
40. A navigation system according to claim 31 , wherein the one or more video cameras and the display panel are integrated into a single armor panel.
41. A navigation system according to claim 38 , wherein the armor panel is configured to replace an existing window or windshield.
42. A method of navigating a vehicle having first and second externally mounted video cameras and an internally mounted display panel, the method comprising:
determining a desired convergence point for the first and second video cameras;
adjusting at least one of the set consisting of a sightline of the first camera, a sight line of the second camera, and a separation distance between the first and second cameras so that the sightlines of the first and second cameras intersect at the desired convergence point;
capturing video images from the first and second cameras; and
displaying the video images from the first and second cameras on the display panel.
43. A method according to claim 42 , further comprising:
determining a gaze direction of a vehicle occupant.
44. The method of claim 43 , wherein the action of determining a desired convergence point is accomplished using the determined gaze direction.
45. A method according to claim 42 , further comprising:
determining a vehicle velocity.
46. The method of claim 45 , wherein the action of determining a desired convergence point is accomplished using the determined vehicle velocity.
47. A method according to claim 42 , wherein the vehicle has third and fourth externally mounted video cameras, the method further comprising:
determining a desired convergence point for the third and fourth video cameras;
adjusting at least one of the set consisting of a sightline of the third camera, a sight line of the fourth camera, and a separation distance between the third and fourth cameras so that the sightlines of the third and fourth cameras intersect at the desired convergence point;
capturing video images from the third and fourth cameras; and
displaying the video images from the third and fourth cameras on the display panel.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/070,757 US20140055602A1 (en) | 2010-04-13 | 2013-11-04 | Display System |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US32374910P | 2010-04-13 | 2010-04-13 | |
US13/085,233 US8576286B1 (en) | 2010-04-13 | 2011-04-12 | Display system |
US14/070,757 US20140055602A1 (en) | 2010-04-13 | 2013-11-04 | Display System |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/085,233 Continuation US8576286B1 (en) | 2010-04-13 | 2011-04-12 | Display system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140055602A1 true US20140055602A1 (en) | 2014-02-27 |
Family
ID=49487867
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/085,233 Active 2032-01-26 US8576286B1 (en) | 2010-04-13 | 2011-04-12 | Display system |
US14/070,757 Abandoned US20140055602A1 (en) | 2010-04-13 | 2013-11-04 | Display System |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/085,233 Active 2032-01-26 US8576286B1 (en) | 2010-04-13 | 2011-04-12 | Display system |
Country Status (1)
Country | Link |
---|---|
US (2) | US8576286B1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160113345A1 (en) * | 2013-06-18 | 2016-04-28 | Alexandr Alexandrovich KOLOTOV | Helmet for motorcyclists and for people who engage in extreme activities |
US20160131633A1 (en) * | 2014-11-06 | 2016-05-12 | Moba Mobile Automation Ag | Device and method for determining the temperature of a road building material applied by a construction machine, and construction machine comprising such a device |
CN106030614A (en) * | 2014-04-22 | 2016-10-12 | 史內普艾德有限公司 | System and method for controlling a camera based on processing an image captured by other camera |
US20170174227A1 (en) * | 2015-12-21 | 2017-06-22 | Igor Tatourian | Dynamic sensor range in advanced driver assistance systems |
CN109409172A (en) * | 2017-08-18 | 2019-03-01 | 安徽三联交通应用技术股份有限公司 | Pilot's line of vision detection method, system, medium and equipment |
JP2020059388A (en) * | 2018-10-10 | 2020-04-16 | 株式会社デンソー | Display device and display control device |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9666091B2 (en) | 2008-01-10 | 2017-05-30 | Lifelong Driver Llc | Driver training system |
JP5499985B2 (en) | 2010-08-09 | 2014-05-21 | ソニー株式会社 | Display assembly |
US9349300B2 (en) * | 2011-10-31 | 2016-05-24 | Lifelong Driver Llc | Senior driver training |
US10013893B2 (en) | 2011-10-31 | 2018-07-03 | Lifelong Driver Llc | Driver training |
JP6070703B2 (en) * | 2012-06-07 | 2017-02-01 | コニカミノルタ株式会社 | Room lighting method |
US9975483B1 (en) * | 2013-02-08 | 2018-05-22 | Amazon Technologies, Inc. | Driver assist using smart mobile devices |
US20140267687A1 (en) * | 2013-03-15 | 2014-09-18 | Ziaollah Loghmani | Vehicular image projection systems and methods of using the same |
SE537279C2 (en) * | 2013-07-12 | 2015-03-24 | BAE Systems Hägglunds AB | System and procedure for handling tactical information in combat vehicles |
US20150062340A1 (en) * | 2013-09-03 | 2015-03-05 | International Business Machines Corporation | High Occupancy Toll Lane Compliance |
US9354073B2 (en) * | 2013-12-09 | 2016-05-31 | Harman International Industries, Inc. | Eye gaze enabled navigation system |
US9756319B2 (en) * | 2014-02-27 | 2017-09-05 | Harman International Industries, Incorporated | Virtual see-through instrument cluster with live video |
KR101622622B1 (en) * | 2014-10-13 | 2016-05-31 | 엘지전자 주식회사 | Apparatus for providing under vehicle image and vehicle including the same |
US10614726B2 (en) | 2014-12-08 | 2020-04-07 | Life Long Driver, Llc | Behaviorally-based crash avoidance system |
US9898078B2 (en) * | 2015-01-12 | 2018-02-20 | Dell Products, L.P. | Immersive environment correction display and method |
US10394237B2 (en) * | 2016-09-08 | 2019-08-27 | Ford Global Technologies, Llc | Perceiving roadway conditions from fused sensor data |
US11615566B2 (en) * | 2017-05-10 | 2023-03-28 | Fotonation Limited | Multi-camera vehicle vision system and method |
US10810751B2 (en) * | 2017-06-23 | 2020-10-20 | Panasonic Intellectual Property Corporation Of America | Distance measuring apparatus and distance measuring method |
US10899358B2 (en) * | 2018-05-31 | 2021-01-26 | Accenture Global Solutions Limited | Vehicle driver monitoring system and method for capturing driver performance parameters |
GB2583483B (en) * | 2019-04-29 | 2021-05-19 | Envisics Ltd | Image capture and display system |
US11388352B2 (en) * | 2020-07-10 | 2022-07-12 | Drivingplates.com LLC | Camera array |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010043266A1 (en) * | 2000-02-02 | 2001-11-22 | Kerry Robinson | Method and apparatus for viewing stereoscopic three- dimensional images |
US20050086000A1 (en) * | 2003-10-17 | 2005-04-21 | Fuji Jukogyo Kabushiki Kaisha | Information display apparatus and information display method |
US20060098093A1 (en) * | 2002-09-12 | 2006-05-11 | Stefan Hahn | Stereo night vision system for vehicles |
US20070247522A1 (en) * | 2003-12-18 | 2007-10-25 | University Of Durham | Method and Apparatus for Generating a Stereoscopic Image |
US20080212215A1 (en) * | 1998-01-07 | 2008-09-04 | Donnelly Corporation | Information display system for a vehicle |
US20080258063A1 (en) * | 2007-04-23 | 2008-10-23 | John Rapanotti | Vehicle threat detection system |
US20080278821A1 (en) * | 2007-05-09 | 2008-11-13 | Harman Becker Automotive Systems Gmbh | Head-mounted display system |
US8139109B2 (en) * | 2006-06-19 | 2012-03-20 | Oshkosh Corporation | Vision system for an autonomous vehicle |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5037166A (en) * | 1985-10-07 | 1991-08-06 | Astronautics Corporation Of America | Holographic optical element for instrument panel displays |
US7436429B2 (en) * | 2003-11-24 | 2008-10-14 | The Boeing Company | Virtual pan/tilt camera system and method for vehicles |
DE102005055350A1 (en) * | 2005-11-21 | 2007-05-24 | Robert Bosch Gmbh | Method for controlling the visual field size of a video system and video system for a motor vehicle |
-
2011
- 2011-04-12 US US13/085,233 patent/US8576286B1/en active Active
-
2013
- 2013-11-04 US US14/070,757 patent/US20140055602A1/en not_active Abandoned
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080212215A1 (en) * | 1998-01-07 | 2008-09-04 | Donnelly Corporation | Information display system for a vehicle |
US20010043266A1 (en) * | 2000-02-02 | 2001-11-22 | Kerry Robinson | Method and apparatus for viewing stereoscopic three- dimensional images |
US20060098093A1 (en) * | 2002-09-12 | 2006-05-11 | Stefan Hahn | Stereo night vision system for vehicles |
US20050086000A1 (en) * | 2003-10-17 | 2005-04-21 | Fuji Jukogyo Kabushiki Kaisha | Information display apparatus and information display method |
US20070247522A1 (en) * | 2003-12-18 | 2007-10-25 | University Of Durham | Method and Apparatus for Generating a Stereoscopic Image |
US8139109B2 (en) * | 2006-06-19 | 2012-03-20 | Oshkosh Corporation | Vision system for an autonomous vehicle |
US20080258063A1 (en) * | 2007-04-23 | 2008-10-23 | John Rapanotti | Vehicle threat detection system |
US20080278821A1 (en) * | 2007-05-09 | 2008-11-13 | Harman Becker Automotive Systems Gmbh | Head-mounted display system |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160113345A1 (en) * | 2013-06-18 | 2016-04-28 | Alexandr Alexandrovich KOLOTOV | Helmet for motorcyclists and for people who engage in extreme activities |
EP3134850A4 (en) * | 2014-04-22 | 2018-02-21 | Snap-Aid Patents Ltd. | System and method for controlling a camera based on processing an image captured by other camera |
EP4250738A3 (en) * | 2014-04-22 | 2023-10-11 | Snap-Aid Patents Ltd. | Method for controlling a camera based on processing an image captured by other camera |
CN106030614A (en) * | 2014-04-22 | 2016-10-12 | 史內普艾德有限公司 | System and method for controlling a camera based on processing an image captured by other camera |
US10473637B2 (en) * | 2014-11-06 | 2019-11-12 | Moba-Mobile Automation Ag | Device and method for determining the temperature of a road building material applied by a construction machine, and construction machine comprising such a device |
US20160131633A1 (en) * | 2014-11-06 | 2016-05-12 | Moba Mobile Automation Ag | Device and method for determining the temperature of a road building material applied by a construction machine, and construction machine comprising such a device |
US9889859B2 (en) * | 2015-12-21 | 2018-02-13 | Intel Corporation | Dynamic sensor range in advanced driver assistance systems |
WO2017112205A1 (en) * | 2015-12-21 | 2017-06-29 | Tatourian Igor | Dynamic sensor range in advanced driver assistance systems |
CN109476309A (en) * | 2015-12-21 | 2019-03-15 | 英特尔公司 | Dynamic pickup range in advanced driving assistance system |
US20170174227A1 (en) * | 2015-12-21 | 2017-06-22 | Igor Tatourian | Dynamic sensor range in advanced driver assistance systems |
CN109409172A (en) * | 2017-08-18 | 2019-03-01 | 安徽三联交通应用技术股份有限公司 | Pilot's line of vision detection method, system, medium and equipment |
JP2020059388A (en) * | 2018-10-10 | 2020-04-16 | 株式会社デンソー | Display device and display control device |
JP7099239B2 (en) | 2018-10-10 | 2022-07-12 | 株式会社デンソー | Display device and display control device |
Also Published As
Publication number | Publication date |
---|---|
US8576286B1 (en) | 2013-11-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8576286B1 (en) | Display system | |
CN106797423B (en) | Sight line detection device | |
US10061992B2 (en) | Image display system | |
US9720231B2 (en) | Display, imaging system and controller for eyewear display device | |
EP1883850B1 (en) | Method of navigating in a surrounding world captured by one or more image sensors and a device for carrying out the method | |
US20060072206A1 (en) | Image display apparatus and image display system | |
US9544578B2 (en) | Portable electronic equipment and method of controlling an autostereoscopic display | |
US20090195652A1 (en) | Interactive Virtual Window Vision System For Mobile Platforms | |
CN106707508A (en) | Cap type virtual reality display image system | |
CN111295702A (en) | Virtual image display device and head-mounted display using the same | |
US20080297590A1 (en) | 3-d robotic vision and vision control system | |
US20190155031A1 (en) | Display apparatus for superimposing a virtual image into the field of vision of a user | |
WO2011102136A1 (en) | Three-dimensional display system and three-dimensional viewing glasses | |
KR20180037887A (en) | Smart glasses | |
JP2009278234A (en) | Display system | |
EP3291170B1 (en) | Aircraft cockpit view enhancing system and method | |
KR20150026201A (en) | A digital device and method of controlling the same | |
KR20180037909A (en) | Smart glasses | |
EP0065837B1 (en) | Display of information | |
US8780179B2 (en) | Robot vision with three dimensional thermal imaging | |
CN115032788B (en) | Head-wearing device | |
CN110024381A (en) | Display device and electronics rearview mirror | |
CN112673620B (en) | Optoelectronic device for piloting an aircraft | |
JPH085955A (en) | Device and method for displaying virtual image type stereoscopic picture | |
JP4399789B2 (en) | 3D image viewing glasses |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GENERAL DYNAMICS ARMAMENT AND TECHNICAL PRODUCTS, Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHILDS, JOHN;REEL/FRAME:031536/0033 Effective date: 20130927 |
|
AS | Assignment |
Owner name: GENERAL DYNAMICS-OTS, INC., FLORIDA Free format text: CHANGE OF NAME;ASSIGNOR:GENERAL DYNAMICS ARMAMENT AND TECHNICAL PRODUCTS, INC.;REEL/FRAME:031690/0920 Effective date: 20130820 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |