CA2376772C - Method and system for a computer-rendered three-dimensional mannequin - Google Patents
Method and system for a computer-rendered three-dimensional mannequin Download PDFInfo
- Publication number
- CA2376772C CA2376772C CA2376772A CA2376772A CA2376772C CA 2376772 C CA2376772 C CA 2376772C CA 2376772 A CA2376772 A CA 2376772A CA 2376772 A CA2376772 A CA 2376772A CA 2376772 C CA2376772 C CA 2376772C
- Authority
- CA
- Canada
- Prior art keywords
- model
- selection control
- garments
- environment
- action
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Classifications
-
- A—HUMAN NECESSITIES
- A41—WEARING APPAREL
- A41H—APPLIANCES OR METHODS FOR MAKING CLOTHES, e.g. FOR DRESS-MAKING OR FOR TAILORING, NOT OTHERWISE PROVIDED FOR
- A41H3/00—Patterns for cutting-out; Methods of drafting or marking-out such patterns, e.g. on the cloth
- A41H3/007—Methods of drafting or marking-out patterns using computers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0641—Shopping interfaces
- G06Q30/0643—Graphical representation of items or shoppers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/16—Cloth
Abstract
A human body (1) is three-dimensionally reproduced to create a computer-rendered model. The model is dynami-cally manipulable to adjust or customize the dimensions of the model to specified measurements. Computer renderings of garments are stored in the computer. The model may be "clothed" with one or more garments (3), the representation of a garment being accu-rately adjusted to reflect its fit on a model of the model's particular dimensions. The computer will accurately represent the specific dimensions of the items of clothing in relation to the customized dimensions of the virtual mannequin. The virtual mannequin may be rotated and may be animated to thereby model the clothing and observe the interaction of the clothing with the environment. Once satisfied to the accuracy/appropriateness of the clothing, the computer can output appropriate dimensional, color and related coordinates.
Description
Title: Method And System For A Computer-Rendered Three-Dimensional Mannequin The present invention relates generally to the field of virtual reality, and more particularly, to the accurate rendering of a three-dimensional model of a person wearing clothing and illustrating the fit and movement of the clothing.
BACKGROUND OF THE INVENTION
TECHNICAL FIELD
Purchasers and potential purchasers of clothing items are able to conveniently preview clothing items displayed in a catalog, in retail stores, and in on-line catalogs. One of the frustrations of purchasers, however, is that clothing items when purchased and actually worn do not meet the expectations developed by the catalog or in the dressing room.
Models in catalogs tend to be taller and thinner than most purchasers, thus developing high expectations for the garment appearance. A garment worn by a person not having model-type proportions might look quite different and might not be flattering.
Moreover, a garment that is photographed or tried on under studio or dressing room lights may have quite a different coloring and reflectivity in other settings, such as day light, candle light, office lighting, and so forth. A person's coloring also affects whether a particular garment is appropriate.
Furthermore, the manner of photographing a garment, typically in a front pose, does not demonstrate back and side fit, and the flow of the garment in various activities. Fitting rooms attempt to solve the problem of front, back and side views by using multiple mirrors. Even so, observing the back view of one's self in a fitting room can be awkward.
Fitting rooms obviously do not permit much testing of a garment in an active activity situation, or observance of a garment from a distance. It is desirable, nevertheless, for a potential purchaser to observe the reaction of the garment to activities such as walking, running, sitting, and so on.
Clothing purchased for a different person, such as a gift, cannot be tried on before purchase. There is no practical way to preliminarily ascertain whether a particular garment will be flattering when worn and when in action.
BACKGROUND ART
Various methods and systems for illustrating the look of a garment on a particular person have been attempted. U.S. Pat. No. 5,850,222, Cone, has attempted a "virtual dressing room", in which a person's measurements are used to create a body data structure that represents the person's figure, by adjusting a standard body data structure. Unfortunately, a garment is represented by a two-dimensional image of the garment worn by a physical mannequin; the garment is inaccurately "stretched" to approximate the adjusted body structure, rather than representing the actual garment.
An on-line clothing catalog by Land's End, available at www.landsend.com, provides a two-dimensional, static representation of a personalized model wearing clothing specified by the user. The static, two-dimensional nature of the model neither permits various viewpoints of the model during activity nor observation of the garment's reaction to the environment.
U.S. Pat. No. 5,557,527, Kotaki et al., concerns the mesh mapping of a simulated knit garment (loop simulation image) on a virtual mannequin, for use in conjunction with designing knit garments. The loop simulation image is dragged and distorted to fit over a scanned-in model of a person. Thus, although Kotaki starts with an accurate representation of a garment, the drawbacks of Cone are magnified in Kotaki. Additionally, Kotaki does not address the accurate representation of a person.
BACKGROUND OF THE INVENTION
TECHNICAL FIELD
Purchasers and potential purchasers of clothing items are able to conveniently preview clothing items displayed in a catalog, in retail stores, and in on-line catalogs. One of the frustrations of purchasers, however, is that clothing items when purchased and actually worn do not meet the expectations developed by the catalog or in the dressing room.
Models in catalogs tend to be taller and thinner than most purchasers, thus developing high expectations for the garment appearance. A garment worn by a person not having model-type proportions might look quite different and might not be flattering.
Moreover, a garment that is photographed or tried on under studio or dressing room lights may have quite a different coloring and reflectivity in other settings, such as day light, candle light, office lighting, and so forth. A person's coloring also affects whether a particular garment is appropriate.
Furthermore, the manner of photographing a garment, typically in a front pose, does not demonstrate back and side fit, and the flow of the garment in various activities. Fitting rooms attempt to solve the problem of front, back and side views by using multiple mirrors. Even so, observing the back view of one's self in a fitting room can be awkward.
Fitting rooms obviously do not permit much testing of a garment in an active activity situation, or observance of a garment from a distance. It is desirable, nevertheless, for a potential purchaser to observe the reaction of the garment to activities such as walking, running, sitting, and so on.
Clothing purchased for a different person, such as a gift, cannot be tried on before purchase. There is no practical way to preliminarily ascertain whether a particular garment will be flattering when worn and when in action.
BACKGROUND ART
Various methods and systems for illustrating the look of a garment on a particular person have been attempted. U.S. Pat. No. 5,850,222, Cone, has attempted a "virtual dressing room", in which a person's measurements are used to create a body data structure that represents the person's figure, by adjusting a standard body data structure. Unfortunately, a garment is represented by a two-dimensional image of the garment worn by a physical mannequin; the garment is inaccurately "stretched" to approximate the adjusted body structure, rather than representing the actual garment.
An on-line clothing catalog by Land's End, available at www.landsend.com, provides a two-dimensional, static representation of a personalized model wearing clothing specified by the user. The static, two-dimensional nature of the model neither permits various viewpoints of the model during activity nor observation of the garment's reaction to the environment.
U.S. Pat. No. 5,557,527, Kotaki et al., concerns the mesh mapping of a simulated knit garment (loop simulation image) on a virtual mannequin, for use in conjunction with designing knit garments. The loop simulation image is dragged and distorted to fit over a scanned-in model of a person. Thus, although Kotaki starts with an accurate representation of a garment, the drawbacks of Cone are magnified in Kotaki. Additionally, Kotaki does not address the accurate representation of a person.
Other methods have attempted to compare garment measurements to standardized or personalized sizes. For example, in U.S. Pat. No. 5,530,652, Croyle et al., a person and clothing are separately measured by a machine vision system. The data can be used to determine whether the garments are within tolerances, or to determine the best size of a particular piece of clothing for a person.
In a different field of application, computers have been used to study fabric characteristics, such as friction and wear. For example, in U.S. Pat. No.
5,495,568, Beavin, a three dimensional model moves, such as raising the arms, bending, walking or running, and the response of a fabric model to motion, stretching and friction is evaluated.
Computers have also been used to create and alter garment patterns to fit standard or individualized body measurements, as in U.S. Pat. No. 4,926,344, Collins et al.
It is generally known in the field of computer graphics to manipulate three-dimensional objects, including shape and texture (U.S. Pat. No. 5,818,420, Mitsumine et al.); to render three-dimensional objects in two-dimensional space (U.S. Pat. No.
5,745,666, Gilley et al.) and vice-versa (U.S. Pat. No. 5,363,476, Kurashige et al.); to manipulate two-dimensional objects in three-dimensional space (U.S. Pat. No. 5,729,673, Cooper et al.);
and to move or animate three-dimensional objects (e.g., U.S. Pat. No. 5,581,665, Sugiura et al.; U.S. Pat. No.
5,577,175, Naka et al.; and U.S. Pat. No. 5,483,630, Unuma et al.).
Thus, there remains a need for a straightforward way for someone to view on a computer screen how a particular garment will look and flow on a particular person and/or in a particular setting. Further, there remains a need for accuracy in rendering such modeled garments.
In a different field of application, computers have been used to study fabric characteristics, such as friction and wear. For example, in U.S. Pat. No.
5,495,568, Beavin, a three dimensional model moves, such as raising the arms, bending, walking or running, and the response of a fabric model to motion, stretching and friction is evaluated.
Computers have also been used to create and alter garment patterns to fit standard or individualized body measurements, as in U.S. Pat. No. 4,926,344, Collins et al.
It is generally known in the field of computer graphics to manipulate three-dimensional objects, including shape and texture (U.S. Pat. No. 5,818,420, Mitsumine et al.); to render three-dimensional objects in two-dimensional space (U.S. Pat. No.
5,745,666, Gilley et al.) and vice-versa (U.S. Pat. No. 5,363,476, Kurashige et al.); to manipulate two-dimensional objects in three-dimensional space (U.S. Pat. No. 5,729,673, Cooper et al.);
and to move or animate three-dimensional objects (e.g., U.S. Pat. No. 5,581,665, Sugiura et al.; U.S. Pat. No.
5,577,175, Naka et al.; and U.S. Pat. No. 5,483,630, Unuma et al.).
Thus, there remains a need for a straightforward way for someone to view on a computer screen how a particular garment will look and flow on a particular person and/or in a particular setting. Further, there remains a need for accuracy in rendering such modeled garments.
DISCLOSURE OF THE INVENTION
It is an object of the present invention to provide a three-dimensional computerized mannequin, having adjustable characteristics such as size, coloring and proportions, for use in modeling garments.
It is a further object of the present invention to provide for a computerized rendering of a model wearing a garment in a particular environment, in order to observe the reaction and flow of the garment to the environment.
It is another object of the present invention to provide for a computerized rending of a model wearing a garment, wherein the model engages in an activity, in order to observe the reaction and flow of the garment during the activity.
It is yet another object of the present invention to provide an adjustable three-dimensional computer-rendered model in connection with an on-line catalog of garments.
Another object of the invention is to observe a garment on a personalized mannequin from a variety of camera positions, distances, and angles.
According to the invention, there is provided a system and method for displaying garments on a computer-rendered three-dimensional mannequin. There is provided a computer-rendered model of a person, the model being three-dimensional and rotatable. A
size selection control is provided, for selecting one of a plurality of sizes, the sizes to be applied to the model.
A proportion selection control is provided, for selecting one of a plurality of proportions, a selected one of the proportions to be applied to the model. A garment selection control is provided for selecting one of a plurality of garments, a selected one of the garments to be displayed on the model.
Optionally, the method and system further include a motion selection control for selecting a motion to be applied to the model.
Further , the method and system may include an action selection control to apply an action to the model.
Moreover, the method and system may include an appearance feature selection control for selecting one of a plurality of appearance features, a selected one of the appearance features to be applied to the model.
Also included might be a appearance feature selection control corresponding to hair length, hair color and/or skin tone.
In the provided system, he action selection control might automatically select the action corresponding to the selected garment from a plurality of actions, each of the actions corresponding to ones of the garments.
A further option is a rotation selection control for rotating the model.
Further provided may be stored data corresponding to each of a plurality of models, the plurality of proportions, and/or the plurality of garments. In the system and method, the stored data corresponding to the garments or models may be stored on a magnetic storage media.
Optionally, the system may include a centralized computer on which the model is rendered, and a remote computer on which the model is displayed.
These and other objects, features and advantages of the present invention are readily apparent from the following drawings and detailed description.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is an illustration of a screen showing a personalized computer-rendered model of a person;
It is an object of the present invention to provide a three-dimensional computerized mannequin, having adjustable characteristics such as size, coloring and proportions, for use in modeling garments.
It is a further object of the present invention to provide for a computerized rendering of a model wearing a garment in a particular environment, in order to observe the reaction and flow of the garment to the environment.
It is another object of the present invention to provide for a computerized rending of a model wearing a garment, wherein the model engages in an activity, in order to observe the reaction and flow of the garment during the activity.
It is yet another object of the present invention to provide an adjustable three-dimensional computer-rendered model in connection with an on-line catalog of garments.
Another object of the invention is to observe a garment on a personalized mannequin from a variety of camera positions, distances, and angles.
According to the invention, there is provided a system and method for displaying garments on a computer-rendered three-dimensional mannequin. There is provided a computer-rendered model of a person, the model being three-dimensional and rotatable. A
size selection control is provided, for selecting one of a plurality of sizes, the sizes to be applied to the model.
A proportion selection control is provided, for selecting one of a plurality of proportions, a selected one of the proportions to be applied to the model. A garment selection control is provided for selecting one of a plurality of garments, a selected one of the garments to be displayed on the model.
Optionally, the method and system further include a motion selection control for selecting a motion to be applied to the model.
Further , the method and system may include an action selection control to apply an action to the model.
Moreover, the method and system may include an appearance feature selection control for selecting one of a plurality of appearance features, a selected one of the appearance features to be applied to the model.
Also included might be a appearance feature selection control corresponding to hair length, hair color and/or skin tone.
In the provided system, he action selection control might automatically select the action corresponding to the selected garment from a plurality of actions, each of the actions corresponding to ones of the garments.
A further option is a rotation selection control for rotating the model.
Further provided may be stored data corresponding to each of a plurality of models, the plurality of proportions, and/or the plurality of garments. In the system and method, the stored data corresponding to the garments or models may be stored on a magnetic storage media.
Optionally, the system may include a centralized computer on which the model is rendered, and a remote computer on which the model is displayed.
These and other objects, features and advantages of the present invention are readily apparent from the following drawings and detailed description.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is an illustration of a screen showing a personalized computer-rendered model of a person;
FIG. 2 is an illustration of a portion of a screen showing a computer-rendered model of a person prior to personalization;
FIG. 3. is a block diagram showing the model database and virtual garment database;
FIG. 4 is a block diagram showing the model database, virtual garment database, and environment database;
FIG. 5 is an illustration of a portion of a screen showing a personalized computer-rendered model of a person, showing rotation of the model;
FIG. 6 is a block diagram illustrating the computer-rendered three-dimensional mannequin utilized in an on-line catalog application accessible over the internet;
FIG. 7 is an illustration of the personalized computer-rendered model of a person, with action;
FIG. 8 is an illustration of the personalized computer-rendered model of a person, with a specified action/environment.; and FIGS. 9A-C are an illustration of the personalized computer-rendered model of a person in a specified action/environment, with varying camera angles.
BEST MODE FOR CARRYING OUT THE INVENTION
Reference is made to Fig. 1, illustrating a personalized, computer-rendered three-dimensional model of a person 1 wearing a garment 3. The model is provided with a rotational control 5. The user interface provides for a size selection control 7, a proportion selection control 9 and a clothing selection control 21. Further the user interface provides for a motion selection control 11, an action/environment selection control 13, and controls for selection of skin tone, hair color and hair length 15, 17, 19.
FIG. 3. is a block diagram showing the model database and virtual garment database;
FIG. 4 is a block diagram showing the model database, virtual garment database, and environment database;
FIG. 5 is an illustration of a portion of a screen showing a personalized computer-rendered model of a person, showing rotation of the model;
FIG. 6 is a block diagram illustrating the computer-rendered three-dimensional mannequin utilized in an on-line catalog application accessible over the internet;
FIG. 7 is an illustration of the personalized computer-rendered model of a person, with action;
FIG. 8 is an illustration of the personalized computer-rendered model of a person, with a specified action/environment.; and FIGS. 9A-C are an illustration of the personalized computer-rendered model of a person in a specified action/environment, with varying camera angles.
BEST MODE FOR CARRYING OUT THE INVENTION
Reference is made to Fig. 1, illustrating a personalized, computer-rendered three-dimensional model of a person 1 wearing a garment 3. The model is provided with a rotational control 5. The user interface provides for a size selection control 7, a proportion selection control 9 and a clothing selection control 21. Further the user interface provides for a motion selection control 11, an action/environment selection control 13, and controls for selection of skin tone, hair color and hair length 15, 17, 19.
In the preferred embodiment, the computer model 1 of a person is rendered by any appropriate computer software for creating and displaying three-dimensional models. Preferably such software initially creates the model based on measurements and accurately depicts the sizing and proportion of the subject person. Alternatively, the model 1 may be based on standard sizes. It is advantageous for the computer model to be stored as a mesh database, further described below. One appropriate software package is sold by Auto Desk under the trademark 3D Studio Max.
Similarly, the garment 3 is rendered by appropriate computer software for creating three-dimensional models, preferably based on accurate measurements of the garment.
Moreover, such software preferably takes into consideration garment qualities such as fabric drape, flow and reflectivity. Advantageously, garment information may be stored in mesh format. Where the garments are sold in various sizes, a measurement of a garment of each size should be taken, and separate information on each garment size should be stored. An appropriate package for creating and displaying such mesh databases is sold under the aforementioned trademark 3D Studio Max.
The rotational control 5 is provided, so that the user may rotate the computer model 1 to observe different aspects of the garment 3 as worn by the model 1. Here, the rotational control is provided in a vertical axis, to simulate a person turning left or right.
Preferably, the rotational control provides full rotation of the model about the vertical axis in accordance with traditional animation principals. The rotational control is illustrated as a double arrow;
it will be appreciated that there are many other methods for allowing a user to control rotation.
The size selection control 7 is provided so that the user may specify the overall size of the model 1 and garment 3. In the illustrated embodiment, the size options are small, medium and large. A finer granularity may be applied, if desired, for example to select a full range of standard women's dress sizes.
Since most individuals are not perfect fits for standardized sizes, the proportion selection control 9 is provided so that the user may fine-tune proportions.
Advantageously, it is the most significant proportions, in terms of garment fit and appearance, that may be modified by the user.
In the illustrated embodiment, the proportion selection control 9 allows selection of cup size A, B, C or D. The cup size proportion selection is particularly suitable in connection with high-end designer clothing and lingerie applications. It may be desirable in other applications to include additional or alternative proportion selections, such as hips, derriere, waist, etc. Notably, the suggested and preferred proportion selection and its effect on the garment 3 may be observed in the rotated appearance of the model 1 in a three-dimensional aspect, rather than a rotated 2-dimensional aspect.
The static appearance of the garment does not tell the complete story of the garment's ornamental appearance. Contributing to the appearance is the flow of the garment in reaction to motion. For example, some garments provide for limited stride; other garments provide for dramatic contrast between back and front necklines; still other garments provide for fluttering hemlines, etc. Therefore, the motion selection control 11 is provided to place the model 1 in motion. In the simplest version, the motion selection control 11 places the model 1 in an abbreviated runway demonstration, that is, the model takes several paces forward, turns, and returns. More elaborate motion control could be provided if desired. For example, other patterns of motion could be provided. Alternatively, motion could be provided specific to a garment, in order to highlight certain features of a garment that might otherwise remain unobserved.
Similarly, the garment 3 is rendered by appropriate computer software for creating three-dimensional models, preferably based on accurate measurements of the garment.
Moreover, such software preferably takes into consideration garment qualities such as fabric drape, flow and reflectivity. Advantageously, garment information may be stored in mesh format. Where the garments are sold in various sizes, a measurement of a garment of each size should be taken, and separate information on each garment size should be stored. An appropriate package for creating and displaying such mesh databases is sold under the aforementioned trademark 3D Studio Max.
The rotational control 5 is provided, so that the user may rotate the computer model 1 to observe different aspects of the garment 3 as worn by the model 1. Here, the rotational control is provided in a vertical axis, to simulate a person turning left or right.
Preferably, the rotational control provides full rotation of the model about the vertical axis in accordance with traditional animation principals. The rotational control is illustrated as a double arrow;
it will be appreciated that there are many other methods for allowing a user to control rotation.
The size selection control 7 is provided so that the user may specify the overall size of the model 1 and garment 3. In the illustrated embodiment, the size options are small, medium and large. A finer granularity may be applied, if desired, for example to select a full range of standard women's dress sizes.
Since most individuals are not perfect fits for standardized sizes, the proportion selection control 9 is provided so that the user may fine-tune proportions.
Advantageously, it is the most significant proportions, in terms of garment fit and appearance, that may be modified by the user.
In the illustrated embodiment, the proportion selection control 9 allows selection of cup size A, B, C or D. The cup size proportion selection is particularly suitable in connection with high-end designer clothing and lingerie applications. It may be desirable in other applications to include additional or alternative proportion selections, such as hips, derriere, waist, etc. Notably, the suggested and preferred proportion selection and its effect on the garment 3 may be observed in the rotated appearance of the model 1 in a three-dimensional aspect, rather than a rotated 2-dimensional aspect.
The static appearance of the garment does not tell the complete story of the garment's ornamental appearance. Contributing to the appearance is the flow of the garment in reaction to motion. For example, some garments provide for limited stride; other garments provide for dramatic contrast between back and front necklines; still other garments provide for fluttering hemlines, etc. Therefore, the motion selection control 11 is provided to place the model 1 in motion. In the simplest version, the motion selection control 11 places the model 1 in an abbreviated runway demonstration, that is, the model takes several paces forward, turns, and returns. More elaborate motion control could be provided if desired. For example, other patterns of motion could be provided. Alternatively, motion could be provided specific to a garment, in order to highlight certain features of a garment that might otherwise remain unobserved.
WO 00/77744 PCT[US00/40022 To further illustrate features of the garments, there is provided an action/environment selection control 13. This control places the model 1 into action in a specific environment. In the preferred embodiment, there is provided for each garment 3 a specific pattern of action for the model and environment that is displayed when the user selects the action/environment selection control 13. For example, when clothing control 21 specifies "suit", the computer determines that the model 1 is displayed with the suit 3, and the corresponding action/environment "discotech" is displayed. This particular action/environment provides numerous bright lights, such as would occur in a discotech; the model 1 strides in a pattern about the virtual discotech, while modeling the garment, in three dimensions.] The action/environment illustrates the response of the garment to particular activity, such as stretching, fluttering, flow, highlights, sheerness, etc. Moreover, the action/environment permits the garment 3 to be illustrated in a setting designed to evoke a mood appropriate to the particular garment 3.
Preferably, additional controls are provided to permit adjustments to tailor the model's appearance to closely resemble a particular person. Such appearance controls advantageously include the skin tone selection control 15; the hair color selection control 17; and the hair length control 19. The skin tone selection control 15 allows the user to select one of a variety of skin tones and to apply the skin tone to the model 1. In the embodiment, the skin tones include none (a neutral gray), light, olive and dark. Of course, additional or fewer skin tones could be provided. The hair color control 17 similarly allows the user to select one of several hair colors and to apply the hair color to the model. In the preferred embodiment, the hair colors include blond, brunette and red; alternative or additional colors could be provided.
The preferred embodiment also provides for selection of hair length via the hair length control 19. Here, two hair lengths are provided - short or long. Additional lengths or hair styles could be provided changing a texture. It will be appreciated that other controls could be provided to tailor the model's appearance, such as eye color and hair style; or that the controls could provide many more choices of hair color, for example.
The clothing selection control 21 provides the ability for the user to select one of several garments 3 which the model 1 will "wear". The preferred embodiment provides for a selection of no clothing, a suit or a dress as a selected garment 3. When the user selects a garment listed in the clothing selection control 21, the garment 3 is displayed on the model 1 as follows. The vertice information is stored in a mesh database. The set of garments from which a user may select should correspond to the garments available in a garment database, discussed in greater detail below. Where there are provided numerous garments in the garment database, the user interface may become more elaborate to permit selection of a particular garment, in accordance with known techniques. Further, it is possible for the user to select a combination of garments for the model 1 to "wear", for example, a skirt and blouse combination.
FIG. 2 is an illustration of a portion of a screen showing a computer-rendered model 1 of a person prior to personalization. In the preferred embodiment, the model 1 may be provided with facial features, e.g., neutral or highly stylized. Further, the model 1 is illustrated in a neutral stance, preferably blocked at an angle to the viewer with arms and legs comfortably spaced to permit viewing of the garment 3 (not illustrated). When rotated by the rotational control 5, the model 1 in the preferred neutral stance will rotate about the vertical axis x.
Alternative axes are, of course, possible. The model 1 may be provided with neutral or realistic anatomic features as preferred, in accordance with standard practice in this field. It is important that the model be three-dimensional and rotatable, so that the fit and appearance of the clothing may be observed from all angles.
FIG. 3 is a block diagram showing the model database 31 and virtual garment database 33 resident on a host computer 37. The model database 31 and garment database 33 should be capable of supporting an accurate three-dimensional rendering, preferably one that may be readily rotated and support altered environments. The data for the model database 31 is preferably selected as the designer desires, preferably to resemble a reasonably neutral figure which proportions would be acceptable as a base figure from where the proportions can be adjusted. by In the preferred embodiment, the model 1 is stored in a model database 31 as mesh data. In order to provide adjustable size and proportions, the model database 31 stores mesh data corresponding to each adjustable size. For example, in the illustrated embodiment, the model database 31 stores the entire mesh data for a small model 1, a medium model 1, and a large model 1. The various proportions selected by the proportion control are stored as mesh data for the affected body part only; the mesh data for the proportion is blended to the mesh data for the model when selected by the user. For example, in the illustrated embodiment, the model database 31 also stores the proportion mesh data for cup sizes A, B, C and D.
Similarly, the garment database 33 is preferably provided as mesh data. For each garment in each size, the garment database 33 stores garment data 35. However, it is possible for the garment database 33 to store basic mesh data for a garment, and for the garment to be adjusted according to the user-selected size and proportion(s). The data for each garment is created by the artist in order to accurately reflect the appearance, color, material (including reflection), material flow, texture, for example.
FIG. 4 illustrates the model database 31 and virtual garment database 33 in connection with an optional environment/action database 39. The environment/action database 39 contains data sufficient to render scenery, lighting, special effects, and to place the model 1 in motion on the screen. In the preferred embodiment, the environment/action database 39 contains one or more of the following elements:
= Scenery = Lighting = Special effects = Model motion (scenery, light, textures, sound, music, streaming video. A
script determines the model movement and other action in the scenery such as waves on a beach birds, etc.) FIG. 5 illustrates the model 1 with selections different from that illustrated in Figure 1, illustrated as rotated 180 degrees. Here, the model 1 has a size selection control 7 at "small", proportion selection control 9 at "A", skin tone selection control 15 at "dark", hair color selection control 17 at "brunette", and hair length selection control 19 at "short." The clothing selection control 21 has selected "suit" as the garment 3.
FIG. 6 is a block diagram illustrating the computer-rendered three-dimensional mannequin utilized over the internet. The computer-rendered three-dimensional mannequin can be accessed by a local user 61 or other user with direct communications connection to the host 37. The mannequin can also be accessed by a remote user 63 via internet 65, an intranet, a cable modem, or other network access.
Advantageously, the mannequin is incorporated into an interactive on-line clothing sales catalog. In that instance, the garment database 33 is created as described above from clothing offered in the catalog.
FIG. 7 is an illustration of the model 1, with action. In this illustration, the model 1 walks on a catwalk, in accordance with motion listed in a script.
FIG. 8 is an illustration of the model 1, within a specified action/environment. Here, the environment is a beach. Note that the environment includes motion, such as waves lapping on a shore, and a breeze. A garment 3 illustrated in such an environment would react to the breeze, and to the motion of the model. The model would move in the environment in accordance with a script.
FIGS. 9A-B are an illustration of the personalized computer-rendered model of a person in a specified action/environment, with varying camera angles. Here, the model 1 is illustrated in a discotheque. The first camera angle as shown in Fig. 9A is a close up front view. The second camera angle as shown in Fig. 9B is a rear flyover camera angle. Other camera angles could be provided. The model 1 is clearly 3-dimensional.
Reference is made again to Fig. 1. The computer-rendered three-dimensional mannequin is utilized by a user as follows. Initially, the system displays a default model 1, and selection and feature controls 7, 9, 11, 13, 15, 17, 19 and 21. In the preferred embodiment, the default model 1 has intermediate size and proportions, with no motion, action or garment selected, as illustrated in Fig. 2. The default model alternatively could be set, for example, to any other combination of sizes, proportions, clothing, etc; or could be set using a cookie to a prior model utilized by the user.
Utilizing the selection and feature controls 7, 9, 11, 1, 15, 17,19 and 21, in no particular order, the user personalizes the model 1. The user selects one of several sizes using the size selection control 7; and selects one of several proportions using the proportion selection control 9. Although the illustrated embodiment provides proportion selection control only for cup size, alternative or additional proportions may be provided for, as discussed above.
If desired, the user selects one of several features via the feature controls 15, 17, 19 for skin tone, hair color and hair length. Preferably, as each selection is made by the user, the model 1 is automatically modified to reflect the selection.
The user utilizes the clothing selection control 21 to select one of several garments that may be "worn" by the model 1. The clothing selection control 21 also provides for a garment selection of "none", in which the model 1 is unclothed. When utilized in connection with an on-line or interactive catalog, the clothing selection control 21 provides a selection of garments corresponding to those garments available from the catalog or manufacturer.
Where there are numerous garments, a more detailed user interface should be provided to select an appropriate garment, garment color, etc. from the catalog.
The user rotates the model 1 utilizing the rotational selection control 5.
In order to observe the model 1 (and garment) in motion, the user selects the motion selection control 11. The model 1 is then displayed in motion on the computer screen, following a pre-set pattern of motion. In order to observe the model 1 (and garment if selected) in action and in an environment, the user selects the action control 13. There are provided several actions and environments. In the preferred embodiment, the action and environment corresponds to the garment selected by the user; thus, when the user selects the action selection control 13, the action is automatically selected to correspond to the garment selected by the user. While an environment/action is selected, the user may select one of several camera angles for observing the model in action.
The user may change the garment on the model by selecting the garment selection control 21. Similarly, the user may adjust the size, proportions or feature selections by selecting one of the controls 7, 9, 15, 17, 19.
Although the model as illustrated and discussed is a female model, it will be appreciated that the principles herein readily could be applied to a male model or a child model.
The system could be provided to a local user, on a central system.
Alternatively, the garment data and/or model data could be stored on a disk, and provided to a user on a periodic basis (for example as a catalogue)l; with the controlling software provided on a central system accessible to the user via communications. Alternatively, the system could be made available via a multiple-user communications system, such as the internet.
While the preferred mode and best mode for carrying out the invention have been described, those familiar with the art to which this invention relates will appreciate that various alternative designs and embodiments for practicing the invention are possible, and will fall within the scope of the following claims.
Preferably, additional controls are provided to permit adjustments to tailor the model's appearance to closely resemble a particular person. Such appearance controls advantageously include the skin tone selection control 15; the hair color selection control 17; and the hair length control 19. The skin tone selection control 15 allows the user to select one of a variety of skin tones and to apply the skin tone to the model 1. In the embodiment, the skin tones include none (a neutral gray), light, olive and dark. Of course, additional or fewer skin tones could be provided. The hair color control 17 similarly allows the user to select one of several hair colors and to apply the hair color to the model. In the preferred embodiment, the hair colors include blond, brunette and red; alternative or additional colors could be provided.
The preferred embodiment also provides for selection of hair length via the hair length control 19. Here, two hair lengths are provided - short or long. Additional lengths or hair styles could be provided changing a texture. It will be appreciated that other controls could be provided to tailor the model's appearance, such as eye color and hair style; or that the controls could provide many more choices of hair color, for example.
The clothing selection control 21 provides the ability for the user to select one of several garments 3 which the model 1 will "wear". The preferred embodiment provides for a selection of no clothing, a suit or a dress as a selected garment 3. When the user selects a garment listed in the clothing selection control 21, the garment 3 is displayed on the model 1 as follows. The vertice information is stored in a mesh database. The set of garments from which a user may select should correspond to the garments available in a garment database, discussed in greater detail below. Where there are provided numerous garments in the garment database, the user interface may become more elaborate to permit selection of a particular garment, in accordance with known techniques. Further, it is possible for the user to select a combination of garments for the model 1 to "wear", for example, a skirt and blouse combination.
FIG. 2 is an illustration of a portion of a screen showing a computer-rendered model 1 of a person prior to personalization. In the preferred embodiment, the model 1 may be provided with facial features, e.g., neutral or highly stylized. Further, the model 1 is illustrated in a neutral stance, preferably blocked at an angle to the viewer with arms and legs comfortably spaced to permit viewing of the garment 3 (not illustrated). When rotated by the rotational control 5, the model 1 in the preferred neutral stance will rotate about the vertical axis x.
Alternative axes are, of course, possible. The model 1 may be provided with neutral or realistic anatomic features as preferred, in accordance with standard practice in this field. It is important that the model be three-dimensional and rotatable, so that the fit and appearance of the clothing may be observed from all angles.
FIG. 3 is a block diagram showing the model database 31 and virtual garment database 33 resident on a host computer 37. The model database 31 and garment database 33 should be capable of supporting an accurate three-dimensional rendering, preferably one that may be readily rotated and support altered environments. The data for the model database 31 is preferably selected as the designer desires, preferably to resemble a reasonably neutral figure which proportions would be acceptable as a base figure from where the proportions can be adjusted. by In the preferred embodiment, the model 1 is stored in a model database 31 as mesh data. In order to provide adjustable size and proportions, the model database 31 stores mesh data corresponding to each adjustable size. For example, in the illustrated embodiment, the model database 31 stores the entire mesh data for a small model 1, a medium model 1, and a large model 1. The various proportions selected by the proportion control are stored as mesh data for the affected body part only; the mesh data for the proportion is blended to the mesh data for the model when selected by the user. For example, in the illustrated embodiment, the model database 31 also stores the proportion mesh data for cup sizes A, B, C and D.
Similarly, the garment database 33 is preferably provided as mesh data. For each garment in each size, the garment database 33 stores garment data 35. However, it is possible for the garment database 33 to store basic mesh data for a garment, and for the garment to be adjusted according to the user-selected size and proportion(s). The data for each garment is created by the artist in order to accurately reflect the appearance, color, material (including reflection), material flow, texture, for example.
FIG. 4 illustrates the model database 31 and virtual garment database 33 in connection with an optional environment/action database 39. The environment/action database 39 contains data sufficient to render scenery, lighting, special effects, and to place the model 1 in motion on the screen. In the preferred embodiment, the environment/action database 39 contains one or more of the following elements:
= Scenery = Lighting = Special effects = Model motion (scenery, light, textures, sound, music, streaming video. A
script determines the model movement and other action in the scenery such as waves on a beach birds, etc.) FIG. 5 illustrates the model 1 with selections different from that illustrated in Figure 1, illustrated as rotated 180 degrees. Here, the model 1 has a size selection control 7 at "small", proportion selection control 9 at "A", skin tone selection control 15 at "dark", hair color selection control 17 at "brunette", and hair length selection control 19 at "short." The clothing selection control 21 has selected "suit" as the garment 3.
FIG. 6 is a block diagram illustrating the computer-rendered three-dimensional mannequin utilized over the internet. The computer-rendered three-dimensional mannequin can be accessed by a local user 61 or other user with direct communications connection to the host 37. The mannequin can also be accessed by a remote user 63 via internet 65, an intranet, a cable modem, or other network access.
Advantageously, the mannequin is incorporated into an interactive on-line clothing sales catalog. In that instance, the garment database 33 is created as described above from clothing offered in the catalog.
FIG. 7 is an illustration of the model 1, with action. In this illustration, the model 1 walks on a catwalk, in accordance with motion listed in a script.
FIG. 8 is an illustration of the model 1, within a specified action/environment. Here, the environment is a beach. Note that the environment includes motion, such as waves lapping on a shore, and a breeze. A garment 3 illustrated in such an environment would react to the breeze, and to the motion of the model. The model would move in the environment in accordance with a script.
FIGS. 9A-B are an illustration of the personalized computer-rendered model of a person in a specified action/environment, with varying camera angles. Here, the model 1 is illustrated in a discotheque. The first camera angle as shown in Fig. 9A is a close up front view. The second camera angle as shown in Fig. 9B is a rear flyover camera angle. Other camera angles could be provided. The model 1 is clearly 3-dimensional.
Reference is made again to Fig. 1. The computer-rendered three-dimensional mannequin is utilized by a user as follows. Initially, the system displays a default model 1, and selection and feature controls 7, 9, 11, 13, 15, 17, 19 and 21. In the preferred embodiment, the default model 1 has intermediate size and proportions, with no motion, action or garment selected, as illustrated in Fig. 2. The default model alternatively could be set, for example, to any other combination of sizes, proportions, clothing, etc; or could be set using a cookie to a prior model utilized by the user.
Utilizing the selection and feature controls 7, 9, 11, 1, 15, 17,19 and 21, in no particular order, the user personalizes the model 1. The user selects one of several sizes using the size selection control 7; and selects one of several proportions using the proportion selection control 9. Although the illustrated embodiment provides proportion selection control only for cup size, alternative or additional proportions may be provided for, as discussed above.
If desired, the user selects one of several features via the feature controls 15, 17, 19 for skin tone, hair color and hair length. Preferably, as each selection is made by the user, the model 1 is automatically modified to reflect the selection.
The user utilizes the clothing selection control 21 to select one of several garments that may be "worn" by the model 1. The clothing selection control 21 also provides for a garment selection of "none", in which the model 1 is unclothed. When utilized in connection with an on-line or interactive catalog, the clothing selection control 21 provides a selection of garments corresponding to those garments available from the catalog or manufacturer.
Where there are numerous garments, a more detailed user interface should be provided to select an appropriate garment, garment color, etc. from the catalog.
The user rotates the model 1 utilizing the rotational selection control 5.
In order to observe the model 1 (and garment) in motion, the user selects the motion selection control 11. The model 1 is then displayed in motion on the computer screen, following a pre-set pattern of motion. In order to observe the model 1 (and garment if selected) in action and in an environment, the user selects the action control 13. There are provided several actions and environments. In the preferred embodiment, the action and environment corresponds to the garment selected by the user; thus, when the user selects the action selection control 13, the action is automatically selected to correspond to the garment selected by the user. While an environment/action is selected, the user may select one of several camera angles for observing the model in action.
The user may change the garment on the model by selecting the garment selection control 21. Similarly, the user may adjust the size, proportions or feature selections by selecting one of the controls 7, 9, 15, 17, 19.
Although the model as illustrated and discussed is a female model, it will be appreciated that the principles herein readily could be applied to a male model or a child model.
The system could be provided to a local user, on a central system.
Alternatively, the garment data and/or model data could be stored on a disk, and provided to a user on a periodic basis (for example as a catalogue)l; with the controlling software provided on a central system accessible to the user via communications. Alternatively, the system could be made available via a multiple-user communications system, such as the internet.
While the preferred mode and best mode for carrying out the invention have been described, those familiar with the art to which this invention relates will appreciate that various alternative designs and embodiments for practicing the invention are possible, and will fall within the scope of the following claims.
Claims (37)
1. A system for displaying garments, comprising :
at least one computer-rendered three-dimensional rotatable model in at least one interactive virtual reality software product;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments selected from the at least one interactive virtual reality software product and to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
at least one computer-rendered three-dimensional rotatable model in at least one interactive virtual reality software product;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments selected from the at least one interactive virtual reality software product and to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
2. The system of Claim 1, further comprising at least one appearance feature selection control for selecting at least one of a plurality of appearance features, a selected at least one of the appearance features to be applied to the at least one model, the at least one appearance feature selection control corresponding to hair length, hair color, or skin tone, or any combination thereof.
3. The system of Claim 1, further comprising at least one rotation selection control for rotating the at least one model.
4. The system of Claim 1, further comprising at least one centralized computer on which the at least one model is rendered, and at least one remote computer on which the at least one model is displayed.
5. The system of Claim 1, wherein the at least one of the plurality of proportions includes:
at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof.
at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof.
6. The system of Claim 1, wherein the at least one of the plurality of garments is rendered in three dimensions via input of a plurality of measurements of at least one actual garment.
7. The system of Claim 1, wherein the at least one of the plurality of garments illustrates qualities comprising: appearance, color, material, reflection, material flow, texture, stretching, fluttering, flow, highlights, fabric drape, or sheerness, or any combination thereof.
8. The system of Claim 1, wherein the at least one environment includes scenery, lighting, special effects, sound, music, or streaming video, or any combination thereof.
9. A method for displaying garments on at least one computer-rendered three-dimensional rotatable model in at least one interactive virtual reality software product, comprising:
selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
selecting at least one of a plurality of garments, the selected at least one of the plurality of garments selected from the at least one interactive virtual reality software product and to be displayed on the at least one model;
applying at least one action and environment to the at least one model to apply at least one of at least one action and at least one environment to the at least one model;
the method being characterized in that the selected at least one of the plurality of garments on the at least one model reacts to the at least one of the at least one action and the at least one environment.
selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
selecting at least one of a plurality of garments, the selected at least one of the plurality of garments selected from the at least one interactive virtual reality software product and to be displayed on the at least one model;
applying at least one action and environment to the at least one model to apply at least one of at least one action and at least one environment to the at least one model;
the method being characterized in that the selected at least one of the plurality of garments on the at least one model reacts to the at least one of the at least one action and the at least one environment.
10. The method of Claim 9, further comprising selecting at least one of a plurality of appearance features, a selected at least one of the appearance features to be applied to the at least one model, the appearance features comprising hair length, hair color, or skin tone, or any combination thereof.
11. The method of Claim 9, further comprising rotating the at least one model.
12. The method of Claim 9, wherein the at least one of a plurality of proportions includes: at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof.
13. The method of Claim 9, wherein the at least one of the plurality of garments is rendered in three dimensions via input of a plurality of measurements of at least one actual garment.
14. The method of Claim 9, wherein the at least one of the plurality of garments illustrates qualities comprising: appearance, color, material, reflection, material flow, texture, stretching, fluttering, fabric drape, flow, highlights, or sheerness, or any combination thereof.
15. The method of Claim 9, wherein the at least one environment includes scenery, lighting, special effects, sound, music, or streaming video, or any combination thereof.
16. A system for displaying garments (3), comprising:
at least one computer-rendered three-dimensional rotatable model of a person;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model, wherein the proportions include: at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
at least one computer-rendered three-dimensional rotatable model of a person;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model, wherein the proportions include: at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
17. The system of Claim 16, further comprising at least one appearance feature selection control for selecting at least one of a plurality of appearance features, a selected at least one of the appearance features to be applied to the at least one model, the at least one appearance feature selection control corresponding to hair length, hair color, or skin tone, or any combination thereof.
18. The system of Claim 16, further comprising at least one rotation selection control for rotating the at least one model.
19. The system of Claim 16, further comprising at least one centralized computer on which the at least one model is rendered, and at least one remote computer on which the at least one model is displayed.
20. The system of Claim 16, wherein the at least one model is in at least one of at least one interactive on-line clothing sales catalog and at least one interactive virtual reality software product.
21. The system of Claim 16, wherein the at least one of the plurality of garments is rendered in three dimensions via input of a plurality of measurements of at least one actual garment.
22. The system of Claim 16, wherein the at least one of the plurality of garments illustrates qualities comprising: appearance, color, material, reflection, material flow, texture, stretching, fluttering, flow, highlights, fabric drape, or sheerness, or any combination thereof.
23. The system of Claim 16, wherein the at least one environment includes scenery, lighting, special effects, sound, music, or streaming video, or any combination thereof.
24. A system for displaying garments, comprising:
at least one computer-rendered three-dimensional rotatable model;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the lighting of the at least one of the at least one action and the at least one environment.
at least one computer-rendered three-dimensional rotatable model;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the lighting of the at least one of the at least one action and the at least one environment.
25. The system of Claim 24, further comprising at least one appearance feature selection control for selecting at least one of a plurality of appearance features, a selected at least one of the appearance features to be applied to the at least one model, the at least one appearance feature selection control corresponding to hair length, hair color, or skin tone, or any combination thereof.
26. The system of Claim 24, further comprising at least one rotation selection control for rotating the at least one model.
27. The system of Claim 24, wherein the at least one of a plurality of proportions includes: at least one cup size proportion, at least one hip proportion, at least one derriere proportion, or at least one waist proportion, or any combination thereof.
28. The system of Claim 24, wherein the at least one of the plurality of garments illustrates qualities comprising: appearance, color, material, reflection, material flow, texture, stretching, fluttering, flow, highlights, fabric drape, or sheerness, or any combination thereof.
29. The system of Claim 24, wherein the at least one environment also includes scenery, lighting, special effects, sound, music, or streaming video, or any combination thereof.
30. The system of Claim 24, wherein the at least one lighting further comprises: coloring, reflectivity, highlights, sheerness, day light, dressing room light, candle light, or office light, or any combination thereof.
31. The system of Claim 24, wherein the selected at least one of the garments on the at least one model also reacts to the skin tone of the at least one model.
32. A system for displaying at least one garment, comprising:
at least one computer-rendered three-dimensional rotatable model;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control automatically selects the at least one of the at least one action and the at least one environment corresponding to the selected at least one of the garments from a plurality of at least actions and environments.
at least one computer-rendered three-dimensional rotatable model;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments to be displayed on the at least one model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the at least one model;
the system being characterized in that the at least one action and environment selection control automatically selects the at least one of the at least one action and the at least one environment corresponding to the selected at least one of the garments from a plurality of at least actions and environments.
33. The system of Claim 32, further comprising at least one appearance feature selection control for selecting at least one of a plurality of appearance features, a selected at least one of the appearance features to be applied to the at least one model, the at least one appearance feature selection control corresponding to hair length, hair color, or skin tone, or any combination thereof.
34. The system of Claim 32, further comprising at least one centralized computer on which the at least one model is rendered, and at least one remote computer on which the at least one model is displayed.
35. The system of Claim 32, wherein the at least one of the plurality of garments, illustrates qualities comprising: appearance, color, material, reflection, material flow, texture, stretching, fluttering, flow, highlights, fabric drape, or sheerness, or any combination thereof.
36. The system of Claim 32, wherein the at least one environment includes scenery, lighting, special effects, sound, music, or streaming video, or any combination thereof.
37. A system for displaying garments, comprising:
at least one computer-rendered three-dimensional rotatable model in at least one interactive on-line clothing sales catalog;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments selected from the at least one interactive on-line clothing sales catalog with stored garment data for at least one garment, the selected at least one of the garments to be displayed on the model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
at least one computer-rendered three-dimensional rotatable model in at least one interactive on-line clothing sales catalog;
at least one size selection control for selecting at least one of a plurality of sizes, the selected at least one of the sizes to be applied to the at least one model;
at least one proportion selection control for selecting at least one of a plurality of proportions, the selected at least one of the proportions to be applied to the at least one model;
at least one garment selection control for selecting at least one of a plurality of garments, the selected at least one of the garments selected from the at least one interactive on-line clothing sales catalog with stored garment data for at least one garment, the selected at least one of the garments to be displayed on the model;
at least one action and environment selection control to apply at least one of at least one action and at least one environment to the model;
the system being characterized in that the at least one action and environment selection control causes the selected at least one of the garments on the at least one model to react to the at least one of the at least one action and the at least one environment.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/329,995 | 1999-06-11 | ||
US09/329,995 US6404426B1 (en) | 1999-06-11 | 1999-06-11 | Method and system for a computer-rendered three-dimensional mannequin |
PCT/US2000/040022 WO2000077744A1 (en) | 1999-06-11 | 2000-06-01 | Method and system for a computer-rendered three-dimensional mannequin |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2376772A1 CA2376772A1 (en) | 2000-12-21 |
CA2376772C true CA2376772C (en) | 2010-12-07 |
Family
ID=23287884
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA2376772A Expired - Lifetime CA2376772C (en) | 1999-06-11 | 2000-06-01 | Method and system for a computer-rendered three-dimensional mannequin |
Country Status (7)
Country | Link |
---|---|
US (7) | US6404426B1 (en) |
EP (1) | EP1407423B1 (en) |
AT (1) | ATE419599T1 (en) |
AU (1) | AU5049200A (en) |
CA (1) | CA2376772C (en) |
DE (1) | DE60041286D1 (en) |
WO (1) | WO2000077744A1 (en) |
Families Citing this family (155)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6404426B1 (en) | 1999-06-11 | 2002-06-11 | Zenimax Media, Inc. | Method and system for a computer-rendered three-dimensional mannequin |
US7328177B1 (en) * | 1999-07-20 | 2008-02-05 | Catherine Lin-Hendel | System and method for interactive, computer assisted personalization of on-line merchandise purchases |
US8843850B2 (en) * | 1999-07-22 | 2014-09-23 | Tavusi Data Solutions Llc | Graphic-information flow for visually analyzing patterns and relationships |
US7418407B2 (en) * | 1999-10-14 | 2008-08-26 | Jarbridge, Inc. | Method for electronic gifting using merging images |
US7917397B1 (en) | 1999-10-14 | 2011-03-29 | Jarbridge, Inc. | Merging private images for gifting |
US6903756B1 (en) | 1999-10-14 | 2005-06-07 | Jarbridge, Inc. | Merged images viewed via a virtual storage closet |
AU7809600A (en) * | 1999-10-18 | 2001-04-30 | T-Bra Limited | Tape measure |
US7663648B1 (en) * | 1999-11-12 | 2010-02-16 | My Virtual Model Inc. | System and method for displaying selected garments on a computer-simulated mannequin |
JP2001283079A (en) | 2000-03-28 | 2001-10-12 | Sony Corp | Communication service method, its device, communication terminal unit, communication system and advertisement publicizing method |
US7149665B2 (en) * | 2000-04-03 | 2006-12-12 | Browzwear International Ltd | System and method for simulation of virtual wear articles on virtual models |
US7079134B2 (en) * | 2000-05-12 | 2006-07-18 | Societe Civile T.P.C. International | Three-dimensional digital method of designing clothes |
AU2001278318A1 (en) * | 2000-07-24 | 2002-02-05 | Jean Nicholson Prudent | Modeling human beings by symbol manipulation |
AUPQ921600A0 (en) * | 2000-08-04 | 2000-08-31 | Canon Kabushiki Kaisha | Automatic person meta-data labeller |
US6791584B1 (en) * | 2000-09-05 | 2004-09-14 | Yiling Xie | Method of scaling face image with spectacle frame image through computer |
US20020069110A1 (en) * | 2000-11-27 | 2002-06-06 | David Sonnenberg | Internet web site |
GB0101371D0 (en) * | 2001-01-19 | 2001-03-07 | Virtual Mirrors Ltd | Production and visualisation of garments |
JP2002297971A (en) * | 2001-01-24 | 2002-10-11 | Sony Computer Entertainment Inc | Electronic commerce system, commodity fitness determining device and method |
US7016824B2 (en) | 2001-02-06 | 2006-03-21 | Geometrix, Inc. | Interactive try-on platform for eyeglasses |
US7496526B2 (en) * | 2001-06-27 | 2009-02-24 | Razumov Sergey N | Method and system for selling clothes |
JP2003058747A (en) * | 2001-08-10 | 2003-02-28 | Fujitsu Ltd | Method, server and program for supporting commodity selection, and recording medium with the program recorded thereon |
US20030050864A1 (en) * | 2001-09-13 | 2003-03-13 | Koninklijke Philips Electronics N.V. | On-line method for aiding a customer in the purchase of clothes |
US7953648B2 (en) * | 2001-11-26 | 2011-05-31 | Vock Curtis A | System and methods for generating virtual clothing experiences |
FR2837593B1 (en) * | 2002-03-22 | 2004-05-28 | Kenneth Kuk Kei Wang | METHOD AND DEVICE FOR VIEWING, ARCHIVING AND TRANSMISSION ON A NETWORK OF COMPUTERS OF A CLOTHING MODEL |
JP2005530578A (en) * | 2002-06-27 | 2005-10-13 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Medical experience method |
FR2843684B1 (en) * | 2002-08-23 | 2004-12-10 | Jean Bernard Condat | DIGITAL HUMAN TEMPLATE AND COMPUTERIZED MANNEQUIN FOR THE USE OF STYLISTS, MODELISTS, PATTERNERS, PLACERS, CORSELIERS AND ANY MAN OF ART IN THE TEXTILE WORLD |
GB0220514D0 (en) * | 2002-09-04 | 2002-10-09 | Depuy Int Ltd | Acetabular cup spacer arrangement |
US20040212674A1 (en) * | 2002-10-16 | 2004-10-28 | Massey Ryan W. | Fashion photography system for clothing stores |
ES2211357B1 (en) * | 2002-12-31 | 2005-10-16 | Reyes Infografica, S.L. | METHOD ASSISTED BY COMPUTER TO DESIGN CLOTHING. |
US20040236552A1 (en) * | 2003-05-22 | 2004-11-25 | Kimberly-Clark Worldwide, Inc. | Method of evaluating products using a virtual environment |
US7099734B2 (en) * | 2003-05-22 | 2006-08-29 | Kimberly-Clark Worldwide, Inc. | Method of evaluating the performance of a product using a virtual environment |
US20040236455A1 (en) * | 2003-05-22 | 2004-11-25 | Kimberly-Clark Worldwide, Inc. | Method of designing a product in a virtual environment |
US20040236457A1 (en) * | 2003-05-22 | 2004-11-25 | Kimberly-Clark Worldwide, Inc. | Method of evaluating articles used on a body in a virtual environment |
US20050131776A1 (en) * | 2003-12-15 | 2005-06-16 | Eastman Kodak Company | Virtual shopper device |
US6882897B1 (en) * | 2004-01-05 | 2005-04-19 | Dennis S. Fernandez | Reconfigurable garment definition and production method |
US7937253B2 (en) * | 2004-03-05 | 2011-05-03 | The Procter & Gamble Company | Virtual prototyping system and method |
US20050267614A1 (en) * | 2004-03-05 | 2005-12-01 | Looney Michael T | System and method of virtual modeling of thin materials |
US20050234782A1 (en) * | 2004-04-14 | 2005-10-20 | Schackne Raney J | Clothing and model image generation, combination, display, and selection |
US7724258B2 (en) * | 2004-06-30 | 2010-05-25 | Purdue Research Foundation | Computer modeling and animation of natural phenomena |
US8660902B2 (en) * | 2004-07-23 | 2014-02-25 | Lori Coulter, Llc | Methods and systems for selling apparel |
US7296372B2 (en) * | 2004-10-04 | 2007-11-20 | Target Brands, Inc. | Retail display article and system |
GB2420643B (en) * | 2004-11-30 | 2011-01-12 | Innoval Technology Ltd | Fabric modelling |
US7950925B2 (en) * | 2004-12-30 | 2011-05-31 | Kimberly-Clark Worldwide, Inc. | Interacting with consumers to inform, educate, consult, and assist with the purchase and use of personal care products |
US20060149638A1 (en) * | 2005-01-06 | 2006-07-06 | Allen Anita L | Electronic personalized clothing makeover assistant |
JP4473754B2 (en) * | 2005-03-11 | 2010-06-02 | 株式会社東芝 | Virtual fitting device |
US20060286512A1 (en) * | 2005-06-17 | 2006-12-21 | Cogliano Mary A | Computerized system for designing a pattern to be applied to a doll or toy |
US20070005174A1 (en) * | 2005-06-29 | 2007-01-04 | Sony Ericsson Mobile Communications Ab | Virtual apparel fitting |
US20080222262A1 (en) * | 2005-09-30 | 2008-09-11 | Sk C&C Co. Ltd. | Digital Album Service System for Showing Digital Fashion Created by Users and Method for Operating the Same |
KR20070050165A (en) * | 2005-11-10 | 2007-05-15 | 종 해 김 | Business method & system related to a fashionable items utilizing internet. |
ES2279708B1 (en) * | 2005-11-15 | 2008-09-16 | Reyes Infografica, S.L. | METHOD OF GENERATION AND USE OF A VIRTUAL CLOTHING CLOTHING TEST AND SYSTEM. |
US7728853B2 (en) * | 2005-11-15 | 2010-06-01 | Honeywell International Inc. | Systems and methods for changing view perspective in 3-D graphical displays of buildings with stepped rotation |
US7675520B2 (en) * | 2005-12-09 | 2010-03-09 | Digital Steamworks, Llc | System, method and computer program for creating two dimensional (2D) or three dimensional (3D) computer animation from video |
US7657341B2 (en) | 2006-01-31 | 2010-02-02 | Dragon & Phoenix Software, Inc. | System, apparatus and method for facilitating pattern-based clothing design activities |
US7657340B2 (en) * | 2006-01-31 | 2010-02-02 | Dragon & Phoenix Software, Inc. | System, apparatus and method for facilitating pattern-based clothing design activities |
EP1826723B1 (en) * | 2006-02-28 | 2015-03-25 | Microsoft Corporation | Object-level image editing |
WO2008015565A2 (en) * | 2006-08-04 | 2008-02-07 | Auckland Uniservices Limited | Biophysical virtual model database and applications |
US8212805B1 (en) | 2007-01-05 | 2012-07-03 | Kenneth Banschick | System and method for parametric display of modular aesthetic designs |
US20080177641A1 (en) * | 2007-01-19 | 2008-07-24 | Edward Herniak | Method and system for online cooperative shopping |
US7979256B2 (en) * | 2007-01-30 | 2011-07-12 | The Procter & Gamble Company | Determining absorbent article effectiveness |
US9530142B2 (en) * | 2007-02-13 | 2016-12-27 | Claudia Juliana Minsky | Method and system for creating a multifunctional collage useable for client/server communication |
US20090033667A1 (en) * | 2007-08-02 | 2009-02-05 | Motorola, Inc. | Method and Apparatus to Facilitate Depicting an Object in Combination with an Accessory |
US20090216659A1 (en) * | 2008-02-27 | 2009-08-27 | Parker Tracy L | Method and System for Assisting Cutomers in Making Purchase Decisions |
US20090222861A1 (en) * | 2008-02-29 | 2009-09-03 | Edmark Tomima L | System and Method for Displaying a High Impact Video Test |
US20090228543A1 (en) * | 2008-03-05 | 2009-09-10 | Collum Dana S | Agent representative for the internet |
JP5159375B2 (en) | 2008-03-07 | 2013-03-06 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Object authenticity determination system and method in metaverse, and computer program thereof |
US7900385B2 (en) | 2008-03-12 | 2011-03-08 | Target Brands, Inc. | Retail display systems and methods |
US8185450B2 (en) * | 2008-06-12 | 2012-05-22 | International Business Machines Corporation | Method and system for self-service manufacture and sale of customized virtual goods |
US9824495B2 (en) * | 2008-09-11 | 2017-11-21 | Apple Inc. | Method and system for compositing an augmented reality scene |
US8704832B2 (en) * | 2008-09-20 | 2014-04-22 | Mixamo, Inc. | Interactive design, synthesis and delivery of 3D character motion data through the web |
US8749556B2 (en) | 2008-10-14 | 2014-06-10 | Mixamo, Inc. | Data compression for real-time streaming of deformable 3D models for 3D animation |
US8941642B2 (en) * | 2008-10-17 | 2015-01-27 | Kabushiki Kaisha Square Enix | System for the creation and editing of three dimensional models |
US8659596B2 (en) | 2008-11-24 | 2014-02-25 | Mixamo, Inc. | Real time generation of animation-ready 3D character models |
WO2010060113A1 (en) * | 2008-11-24 | 2010-05-27 | Mixamo, Inc. | Real time generation of animation-ready 3d character models |
US8982122B2 (en) * | 2008-11-24 | 2015-03-17 | Mixamo, Inc. | Real time concurrent design of shape, texture, and motion for 3D character animation |
US20100259547A1 (en) | 2009-02-12 | 2010-10-14 | Mixamo, Inc. | Web platform for interactive design, synthesis and delivery of 3d character motion data |
US20120284148A1 (en) * | 2009-02-27 | 2012-11-08 | Clonnequin Pty Ltd | System and method of facilitating on line purchase of clothing |
WO2010129721A2 (en) * | 2009-05-05 | 2010-11-11 | Mixamo, Inc. | Distributed markerless motion capture |
US8364561B2 (en) * | 2009-05-26 | 2013-01-29 | Embodee Corp. | Garment digitization system and method |
US8700477B2 (en) * | 2009-05-26 | 2014-04-15 | Embodee Corp. | Garment fit portrayal system and method |
US8818883B2 (en) * | 2009-07-23 | 2014-08-26 | Apple Inc. | Personalized shopping avatar |
GB0915016D0 (en) * | 2009-08-28 | 2009-09-30 | Digimania Ltd | Animation of characters |
WO2011038275A1 (en) | 2009-09-25 | 2011-03-31 | Avazap Inc. | Frameless video system |
US8175741B2 (en) * | 2009-12-22 | 2012-05-08 | Ko Young-A | Method for creating panels and pattern-making |
US8736606B2 (en) * | 2010-02-01 | 2014-05-27 | SathyaKumar Andre Ramalingam | Method and apparatus to create 3-dimensional computer models of persons from specially created 2-dimensional images |
US8090465B2 (en) * | 2010-03-04 | 2012-01-03 | Belinda Luna Zeng | Fashion design method, system and apparatus |
US20110234591A1 (en) * | 2010-03-26 | 2011-09-29 | Microsoft Corporation | Personalized Apparel and Accessories Inventory and Display |
US8928672B2 (en) | 2010-04-28 | 2015-01-06 | Mixamo, Inc. | Real-time automatic concatenation of 3D animation sequences |
US8908928B1 (en) | 2010-05-31 | 2014-12-09 | Andrew S. Hansen | Body modeling and garment fitting using an electronic device |
US8655053B1 (en) | 2010-05-31 | 2014-02-18 | Andrew S Hansen | Body modeling and garment fitting using an electronic device |
US10628729B2 (en) | 2010-06-08 | 2020-04-21 | Styku, LLC | System and method for body scanning and avatar creation |
US10628666B2 (en) | 2010-06-08 | 2020-04-21 | Styku, LLC | Cloud server body scan data system |
US11640672B2 (en) | 2010-06-08 | 2023-05-02 | Styku Llc | Method and system for wireless ultra-low footprint body scanning |
US11244223B2 (en) | 2010-06-08 | 2022-02-08 | Iva Sareen | Online garment design and collaboration system and method |
US8797328B2 (en) | 2010-07-23 | 2014-08-05 | Mixamo, Inc. | Automatic generation of 3D character animation from 3D meshes |
AU2011293106B2 (en) * | 2010-08-27 | 2016-07-28 | Clonnequin Pty Ltd | Mannequin, method and system for purchase, making and alteration of clothing |
US8711175B2 (en) * | 2010-11-24 | 2014-04-29 | Modiface Inc. | Method and system for simulating superimposition of a non-linearly stretchable object upon a base object using representative images |
GB201102794D0 (en) | 2011-02-17 | 2011-03-30 | Metail Ltd | Online retail system |
US10049482B2 (en) | 2011-07-22 | 2018-08-14 | Adobe Systems Incorporated | Systems and methods for animation recommendations |
US10748325B2 (en) | 2011-11-17 | 2020-08-18 | Adobe Inc. | System and method for automatic rigging of three dimensional characters for facial animation |
CN102521878B (en) * | 2011-12-20 | 2013-08-14 | 恒源祥(集团)有限公司 | Programmable three-dimensional human body modeling and human body measurement system and method thereof |
US8782565B2 (en) | 2012-01-12 | 2014-07-15 | Cisco Technology, Inc. | System for selecting objects on display |
CN102622699B (en) * | 2012-02-27 | 2016-12-14 | 浙江工商大学 | The method showing real product in 3D virtual world |
US9747495B2 (en) | 2012-03-06 | 2017-08-29 | Adobe Systems Incorporated | Systems and methods for creating and distributing modifiable animated video messages |
US9147207B2 (en) * | 2012-07-09 | 2015-09-29 | Stylewhile Oy | System and method for generating image data for on-line shopping |
US20140201023A1 (en) * | 2013-01-11 | 2014-07-17 | Xiaofan Tang | System and Method for Virtual Fitting and Consumer Interaction |
US9378586B2 (en) * | 2013-02-21 | 2016-06-28 | Seoul National University Industry Foundation | Method of smart grading based on parameterized draft |
US10366175B2 (en) * | 2013-03-15 | 2019-07-30 | 3D Tech Llc | System and method for automated manufacturing of custom apparel |
US20140368499A1 (en) * | 2013-06-15 | 2014-12-18 | Rajdeep Kaur | Virtual Fitting Room |
US9119487B2 (en) | 2013-09-13 | 2015-09-01 | Target Brands, Inc. | Display system |
US9635895B1 (en) | 2013-10-29 | 2017-05-02 | Vf Imagewear, Inc. | System and method for mapping wearer mobility for clothing design |
US20150134302A1 (en) | 2013-11-14 | 2015-05-14 | Jatin Chhugani | 3-dimensional digital garment creation from planar garment photographs |
US9773274B2 (en) | 2013-12-02 | 2017-09-26 | Scott William Curry | System and method for online virtual fitting room |
US10366439B2 (en) | 2013-12-27 | 2019-07-30 | Ebay Inc. | Regional item reccomendations |
US9699123B2 (en) | 2014-04-01 | 2017-07-04 | Ditto Technologies, Inc. | Methods, systems, and non-transitory machine-readable medium for incorporating a series of images resident on a user device into an existing web browser session |
US20150324103A1 (en) * | 2014-05-09 | 2015-11-12 | Warangkana Tepmongkol | Social network system for sharing fashions |
TWI560627B (en) * | 2014-05-09 | 2016-12-01 | Happy Ray Co Ltd | Method and system for matching graphics |
US10509865B2 (en) * | 2014-09-18 | 2019-12-17 | Google Llc | Dress form for three-dimensional drawing inside virtual reality environment |
US20160092956A1 (en) | 2014-09-30 | 2016-03-31 | Jonathan Su | Garment size mapping |
US10204375B2 (en) | 2014-12-01 | 2019-02-12 | Ebay Inc. | Digital wardrobe using simulated forces on garment models |
US20170352091A1 (en) * | 2014-12-16 | 2017-12-07 | Metail Limited | Methods for generating a 3d virtual body model of a person combined with a 3d garment image, and related devices, systems and computer program products |
US9984409B2 (en) | 2014-12-22 | 2018-05-29 | Ebay Inc. | Systems and methods for generating virtual contexts |
EP3038053B1 (en) | 2014-12-22 | 2019-11-06 | Reactive Reality GmbH | Method and system for generating garment model data |
US10475113B2 (en) | 2014-12-23 | 2019-11-12 | Ebay Inc. | Method system and medium for generating virtual contexts from three dimensional models |
US9754417B2 (en) * | 2014-12-31 | 2017-09-05 | Canon Information And Imaging Solutions, Inc. | Methods and systems for displaying virtual objects |
US10310616B2 (en) | 2015-03-31 | 2019-06-04 | Ebay Inc. | Modification of three-dimensional garments using gestures |
US9905019B2 (en) | 2015-08-07 | 2018-02-27 | Selfiestyler Inc. | Virtual apparel fitting systems and methods |
US10430867B2 (en) | 2015-08-07 | 2019-10-01 | SelfieStyler, Inc. | Virtual garment carousel |
CN106690617A (en) * | 2015-11-17 | 2017-05-24 | 北京三件客科技有限公司 | 3D model figure measuring system and shirt personalized customization 3D model figure measuring system |
US9901192B2 (en) | 2015-12-28 | 2018-02-27 | James Tiggett, JR. | Robotic mannequin system |
US10127717B2 (en) | 2016-02-16 | 2018-11-13 | Ohzone, Inc. | System for 3D Clothing Model Creation |
US11615462B2 (en) | 2016-02-16 | 2023-03-28 | Ohzone, Inc. | System for virtually sharing customized clothing |
US10373386B2 (en) | 2016-02-16 | 2019-08-06 | Ohzone, Inc. | System and method for virtually trying-on clothing |
US10262440B2 (en) | 2016-03-25 | 2019-04-16 | Ebay Inc. | Publication modification using body coordinates |
US10062198B2 (en) | 2016-06-23 | 2018-08-28 | LoomAi, Inc. | Systems and methods for generating computer ready animation models of a human head from captured data images |
US10559111B2 (en) | 2016-06-23 | 2020-02-11 | LoomAi, Inc. | Systems and methods for generating computer ready animation models of a human head from captured data images |
US10395425B2 (en) | 2016-08-04 | 2019-08-27 | International Business Machines Corporation | System, method and recording medium for mannequin pose control |
US20180096505A1 (en) * | 2016-10-04 | 2018-04-05 | Facebook, Inc. | Controls and Interfaces for User Interactions in Virtual Spaces |
CN106780700B (en) * | 2017-01-13 | 2020-05-01 | 炫彩互动网络科技有限公司 | Game personalized 3D coloring method based on VR technology |
CN106984039A (en) * | 2017-04-14 | 2017-07-28 | 苏州蜗牛数字科技股份有限公司 | A kind of method for projecting to real human body in VR game |
US11145138B2 (en) * | 2017-04-28 | 2021-10-12 | Linden Research, Inc. | Virtual reality presentation of layers of clothing on avatars |
US11094136B2 (en) | 2017-04-28 | 2021-08-17 | Linden Research, Inc. | Virtual reality presentation of clothing fitted on avatars |
US10467818B2 (en) | 2017-05-31 | 2019-11-05 | Marie Manvel | System and method for virtual wardrobe selection |
US11948057B2 (en) * | 2017-06-22 | 2024-04-02 | Iva Sareen | Online garment design and collaboration system and method |
US10613710B2 (en) | 2017-10-22 | 2020-04-07 | SWATCHBOOK, Inc. | Product simulation and control system for user navigation and interaction |
US10765948B2 (en) | 2017-12-22 | 2020-09-08 | Activision Publishing, Inc. | Video game content aggregation, normalization, and publication systems and methods |
US10262432B1 (en) | 2017-12-30 | 2019-04-16 | Gabriel Keilholz | System and method for measuring and comparing items using computer vision |
US10872475B2 (en) | 2018-02-27 | 2020-12-22 | Soul Vision Creations Private Limited | 3D mobile renderer for user-generated avatar, apparel, and accessories |
CN108694739B (en) * | 2018-04-26 | 2022-04-05 | 中山大学 | Fabric realistic appearance rendering system and method based on micro-appearance model |
US10198845B1 (en) | 2018-05-29 | 2019-02-05 | LoomAi, Inc. | Methods and systems for animating facial expressions |
CN109324687B (en) * | 2018-08-14 | 2021-10-01 | 华为技术有限公司 | Display method and virtual reality equipment |
US11803664B2 (en) | 2018-10-09 | 2023-10-31 | Ebay Inc. | Distributed application architectures using blockchain and distributed file systems |
US11551393B2 (en) | 2019-07-23 | 2023-01-10 | LoomAi, Inc. | Systems and methods for animation generation |
US11712627B2 (en) | 2019-11-08 | 2023-08-01 | Activision Publishing, Inc. | System and method for providing conditional access to virtual gaming items |
WO2022006683A1 (en) * | 2020-07-10 | 2022-01-13 | Wimalasuriya Daya Karunita | Tension-map based virtual fitting room systems and methods |
CN113129450B (en) * | 2021-04-21 | 2024-04-05 | 北京百度网讯科技有限公司 | Virtual fitting method, device, electronic equipment and medium |
US11651564B2 (en) | 2021-06-15 | 2023-05-16 | Tailr LLC | System and method for virtual fitting of garments over a communications network |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4734690A (en) * | 1984-07-20 | 1988-03-29 | Tektronix, Inc. | Method and apparatus for spherical panning |
EP0175039B1 (en) * | 1984-09-18 | 1990-01-31 | Wolf Merz | Device for putting down and picking up patterns of a part of a garment |
US4926344A (en) | 1988-03-16 | 1990-05-15 | Minnesota Mining And Manufacturing Company | Data storage structure of garment patterns to enable subsequent computerized prealteration |
US5495568A (en) | 1990-07-09 | 1996-02-27 | Beavin; William C. | Computerized clothing designer |
US5483630A (en) | 1990-07-12 | 1996-01-09 | Hitachi, Ltd. | Method and apparatus for representing motion of multiple-jointed object, computer graphic apparatus, and robot controller |
JP3179474B2 (en) * | 1990-11-28 | 2001-06-25 | 株式会社日立製作所 | Computer graphic display method and information processing apparatus |
JP3117097B2 (en) | 1992-01-28 | 2000-12-11 | ソニー株式会社 | Image conversion device |
JPH0696100A (en) * | 1992-09-09 | 1994-04-08 | Mitsubishi Electric Corp | Remote transaction system |
US5581665A (en) | 1992-10-27 | 1996-12-03 | Matsushita Electric Industrial Co., Ltd. | Three-dimensional object movement and transformation processing apparatus for performing movement and transformation of an object in a three-diamensional space |
JP3052681B2 (en) | 1993-08-06 | 2000-06-19 | 松下電器産業株式会社 | 3D video generation device |
US5530652A (en) | 1993-08-11 | 1996-06-25 | Levi Strauss & Co. | Automatic garment inspection and measurement system |
US5557527A (en) | 1993-08-31 | 1996-09-17 | Shima Seiki Manufacturing Ltd. | Knit design system and a method for designing knit fabrics |
GB9322260D0 (en) * | 1993-10-28 | 1993-12-15 | Pandora Int Ltd | Digital video processor |
US5544291A (en) | 1993-11-10 | 1996-08-06 | Adobe Systems, Inc. | Resolution-independent method for displaying a three dimensional model in two-dimensional display space |
JP3543395B2 (en) * | 1994-11-17 | 2004-07-14 | 株式会社日立製作所 | Service provision and usage |
US5729673A (en) | 1995-04-07 | 1998-03-17 | Avid Technology, Inc. | Direct manipulation of two-dimensional moving picture streams in three-dimensional space |
US5850222A (en) | 1995-09-13 | 1998-12-15 | Pixel Dust, Inc. | Method and system for displaying a graphic image of a person modeling a garment |
US5818420A (en) | 1996-07-31 | 1998-10-06 | Nippon Hoso Kyokai | 3D object graphics display device, 3D object graphics display method, and manipulator for 3D object graphics display |
US5930769A (en) * | 1996-10-07 | 1999-07-27 | Rose; Andrea | System and method for fashion shopping |
US6307568B1 (en) * | 1998-10-28 | 2001-10-23 | Imaginarix Ltd. | Virtual dressing over the internet |
US6634949B1 (en) * | 1999-02-26 | 2003-10-21 | Creative Kingdoms, Llc | Multi-media interactive play system |
US6196429B1 (en) * | 1999-04-28 | 2001-03-06 | Cyberform Corp. | Dress or clothing form |
US7062454B1 (en) * | 1999-05-06 | 2006-06-13 | Jarbridge, Inc. | Previewing system and method |
US6404426B1 (en) | 1999-06-11 | 2002-06-11 | Zenimax Media, Inc. | Method and system for a computer-rendered three-dimensional mannequin |
US6968075B1 (en) * | 2000-05-09 | 2005-11-22 | Chang Kurt C | System and method for three-dimensional shape and size measurement |
-
1999
- 1999-06-11 US US09/329,995 patent/US6404426B1/en not_active Expired - Lifetime
-
2000
- 2000-06-01 AT AT00932820T patent/ATE419599T1/en not_active IP Right Cessation
- 2000-06-01 AU AU50492/00A patent/AU5049200A/en not_active Abandoned
- 2000-06-01 CA CA2376772A patent/CA2376772C/en not_active Expired - Lifetime
- 2000-06-01 WO PCT/US2000/040022 patent/WO2000077744A1/en active Application Filing
- 2000-06-01 DE DE60041286T patent/DE60041286D1/en not_active Expired - Lifetime
- 2000-06-01 EP EP00932820A patent/EP1407423B1/en not_active Expired - Lifetime
-
2001
- 2001-08-08 US US09/923,410 patent/US7212202B2/en not_active Expired - Lifetime
-
2007
- 2007-01-26 US US11/627,549 patent/US7522165B2/en not_active Expired - Fee Related
-
2008
- 2008-12-23 US US12/343,144 patent/US8189000B2/en not_active Expired - Fee Related
-
2012
- 2012-04-23 US US13/453,827 patent/US8970585B2/en not_active Expired - Fee Related
-
2015
- 2015-01-22 US US14/603,096 patent/US9877531B2/en not_active Expired - Fee Related
-
2017
- 2017-12-05 US US15/831,911 patent/US20180092416A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
EP1407423A1 (en) | 2004-04-14 |
US20180092416A1 (en) | 2018-04-05 |
US8189000B2 (en) | 2012-05-29 |
US8970585B2 (en) | 2015-03-03 |
US20020021297A1 (en) | 2002-02-21 |
US9877531B2 (en) | 2018-01-30 |
US20090213117A1 (en) | 2009-08-27 |
US20120293497A1 (en) | 2012-11-22 |
US20070182736A1 (en) | 2007-08-09 |
AU5049200A (en) | 2001-01-02 |
EP1407423B1 (en) | 2008-12-31 |
CA2376772A1 (en) | 2000-12-21 |
DE60041286D1 (en) | 2009-02-12 |
EP1407423A4 (en) | 2005-03-16 |
US7212202B2 (en) | 2007-05-01 |
US6404426B1 (en) | 2002-06-11 |
US7522165B2 (en) | 2009-04-21 |
WO2000077744A1 (en) | 2000-12-21 |
US20150199753A1 (en) | 2015-07-16 |
ATE419599T1 (en) | 2009-01-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2376772C (en) | Method and system for a computer-rendered three-dimensional mannequin | |
US20200380333A1 (en) | System and method for body scanning and avatar creation | |
US10573091B2 (en) | Systems and methods to create a virtual object or avatar | |
US10872475B2 (en) | 3D mobile renderer for user-generated avatar, apparel, and accessories | |
CN107636585B (en) | Generation of three-dimensional fashion objects by drawing inside a virtual reality environment | |
US11640672B2 (en) | Method and system for wireless ultra-low footprint body scanning | |
AU739559B2 (en) | A method and a device for displaying at least part of the human body with a modified appearance thereof | |
US7149665B2 (en) | System and method for simulation of virtual wear articles on virtual models | |
US7663648B1 (en) | System and method for displaying selected garments on a computer-simulated mannequin | |
US20100097375A1 (en) | Three-dimensional design support apparatus and three-dimensional model display system | |
CN109598798A (en) | Virtual object approximating method and virtual object are fitted service system | |
Loker et al. | Dress in the third dimension: Online interactivity and its new horizons | |
JP2000512039A (en) | Programmable computer graphic objects | |
Gray | In virtual fashion | |
CN113610612A (en) | 3D virtual fitting method, system and storage medium | |
WO2020104990A1 (en) | Virtually trying cloths & accessories on body model | |
WO2013120453A1 (en) | System and method for natural person digitized image design | |
US11948057B2 (en) | Online garment design and collaboration system and method | |
WO2001075750A1 (en) | System and method for virtual shopping of wear articles | |
CA2461038C (en) | System and method for displaying selected garments on a computer-simulated mannequin | |
Dvořák et al. | Presentation of historical clothing digital replicas in motion | |
JP7365457B2 (en) | Data processing device, program, and data processing method | |
JP2022117813A (en) | Virtual try-on system and program used for the same | |
Kaur et al. | The Future Possibilities of Artificial Intelligence in Modern Drapes |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request |