WO1999042918A2 - Priority-based virtual environment - Google Patents
Priority-based virtual environment Download PDFInfo
- Publication number
- WO1999042918A2 WO1999042918A2 PCT/IB1999/000299 IB9900299W WO9942918A2 WO 1999042918 A2 WO1999042918 A2 WO 1999042918A2 IB 9900299 W IB9900299 W IB 9900299W WO 9942918 A2 WO9942918 A2 WO 9942918A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- priority
- participant
- objects
- rendering
- avatar
- Prior art date
Links
- 238000009877 rendering Methods 0.000 claims abstract description 58
- 239000013598 vector Substances 0.000 claims description 66
- 238000000034 method Methods 0.000 claims description 9
- 230000000007 visual effect Effects 0.000 claims description 3
- 230000001419 dependent effect Effects 0.000 claims 5
- 230000001965 increasing effect Effects 0.000 abstract description 19
- 230000004438 eyesight Effects 0.000 description 38
- 230000004048 modification Effects 0.000 description 32
- 238000012986 modification Methods 0.000 description 32
- 238000011524 similarity measure Methods 0.000 description 26
- 230000003993 interaction Effects 0.000 description 23
- 230000000903 blocking effect Effects 0.000 description 19
- 238000007654 immersion Methods 0.000 description 18
- 239000011435 rock Substances 0.000 description 18
- 241000607479 Yersinia pestis Species 0.000 description 16
- 238000013459 approach Methods 0.000 description 14
- 238000004891 communication Methods 0.000 description 13
- 230000003997 social interaction Effects 0.000 description 11
- 230000008447 perception Effects 0.000 description 10
- 230000003247 decreasing effect Effects 0.000 description 9
- 230000007246 mechanism Effects 0.000 description 9
- 239000002131 composite material Substances 0.000 description 6
- 230000033001 locomotion Effects 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 238000007667 floating Methods 0.000 description 5
- 230000005236 sound signal Effects 0.000 description 5
- 229910001385 heavy metal Inorganic materials 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 230000004913 activation Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 3
- 230000000295 complement effect Effects 0.000 description 3
- 230000001186 cumulative effect Effects 0.000 description 3
- 230000008921 facial expression Effects 0.000 description 3
- 230000008676 import Effects 0.000 description 3
- 206010020772 Hypertension Diseases 0.000 description 2
- 230000002238 attenuated effect Effects 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000001815 facial effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 230000010354 integration Effects 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 208000010125 myocardial infarction Diseases 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 230000000391 smoking effect Effects 0.000 description 2
- 238000010408 sweeping Methods 0.000 description 2
- 238000013518 transcription Methods 0.000 description 2
- 230000035897 transcription Effects 0.000 description 2
- 230000001960 triggered effect Effects 0.000 description 2
- 206010063659 Aversion Diseases 0.000 description 1
- 241000282472 Canis lupus familiaris Species 0.000 description 1
- 241000320126 Pseudomugilidae Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 229910052763 palladium Inorganic materials 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000002250 progressing effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 230000035939 shock Effects 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G06Q50/50—
Definitions
- This invention relates to a system, apparatus and method for enabling rendering in a virtual environment and, particularly, a system, apparatus and method for enabling priority-based rendering in a virtual environment.
- the high- volume, consumer electronics market segment is progressing toward the integration of computer-based advances and traditional audio/video products.
- networking systems e.g., the Internet
- Traditional audio and video offers ease of use and consumer comfort and acceptance in the home, particularly the family room.
- Virtual environments comprise computer-generated, generally three-dimensional representations of a simulated or real, physical setting or scene.
- the setting/scene can be a city, a mall, an individual store, a building, a suite of offices, an individual office or some other space.
- the representations can be more or less realistic both in terms of their rendering of the real world and in terms of the human senses that are supported.
- a virtual environment generally comprises virtual objects, the objects typically including entities that are animate or inanimate.
- Inanimate entities may include features of the environment, such as, for example, walls of a virtual office that are always inanimate walls in the office.
- Animate entities may include so-called avatars and bots.
- Bots broadly, are images that, generally, operate autonomously from the participants, performing predetermined tasks or providing features within the environment.
- a bot can include, for example, a wall that transforms to deliver incoming messages.
- An avatar by comparison, generally is an image that represents, and is controlled by, a participant.
- An avatar typically supports one or more of body gestures, facial expressions, speech and motion.
- a virtual environment generally beckons its participants to become immersed in the sensory experience it provides.
- the quality of the participant's experience depends, broadly, on the quality of interaction supported by the environment.
- studies of existing virtual environments have indicated that the quality of social interaction between participants is a dominant factor in determining whether participants are stimulated to repeatedly visit a virtual environment.
- the studies have also shown that the quality of social interactions depends on support, in avatars, of facial expressions, as well as other body language, generally associated with face-to-face encounters in the real world.
- An object of this invention is to provide a virtual environment that overcomes the shortcomings associated with rendering and data communication.
- a system that comprises a priority component.
- the priority component determines priorities for avatars and objects, such that highest priority avatars/objects have (a) enhanced rendering (e.g., enhancement of resolution, frame rate, colors and/or other qualities) as to the senses and dimensions supported by the system and (b) enhanced quality of service from networks and operating systems.
- the priority component preferably is responsive to selected parameters, including, for example, one or more of (i) objects' relative positions, (ii) the direction, orientation and span of sense and focus spaces, (iii) participants' profiles, (iv) parameters set by developers and/or participants, (v) social parameters, (vi) economic features, e.g., cost structures and (vii) combinations of these and other selected parameters.
- FIG. 1 a showing a block diagram of a virtual environment system, including a priority component, according to the principles of the invention
- Fig. lb showing a block diagram of a virtual environment system, showing attention sense and focus spaces of a attention component, according to the principles of the invention
- FIG. 2 showing a sense space
- Fig. 3 showing a focus space
- Fig. 4 showing one embodiment of a priority component's priority regions as projected in a plane containing the center of the focus and sense spheres and the focus and sense vectors, according to the principles of the invention
- Fig. 5 showing one embodiment of a priority component's priority regions as projected in a plane containing the center of the focus and sense spheres and the focus and sense vectors, according to the principles of the invention
- Fig. 6 showing one embodiment of a priority component's priority regions as projected in a plane containing the center of the focus and sense spheres and the focus and sense vectors, according to the principles of the invention
- Fig. 7 showing one embodiment of a priority component's priority regions as projected in a plane containing the center of the focus and sense spheres and the focus and sense vectors, according to the principles of the invention.
- Fig. 8 showing an interest specification tool, responsive to selection of interest parameters associated with the priority component, and including an interest specifier component and a vocabulary specifier component.
- Virtual environment system refers to any individual or plural, central or distributed, computing-based system capable of supporting a virtual world or other virtual environment, or settings/scenes thereof, particularly any such environment characterized by avatars, bots and other virtual objects.
- Virtual object (or, sometimes, object alone) is any thing or part of the virtual environment, whether animate, inanimate or combinations thereof.
- Inanimate virtual objects may include features of the environment, e.g., inanimate walls in a virtual office.
- Animate virtual objects may include avatars and bots.
- Avatar refers to a virtual object that represents, and is controlled by, a participant.
- An avatar typically supports, e.g., one or more of body gestures, facial expressions, speech and motion.
- An avatar, while so representing and supporting, can be other than anthropomo ⁇ hic and can comprise a combination of both animate and inanimate virtual objects.
- Non-avatar object refers to a virtual object that does not represent a participant.
- non- avatar objects can operate autonomously from the participants, supporting predetermined senses (e.g., analogous to avatar sight), having predetermined functions (e.g., provide information), performing predetermined tasks (e.g., collect information) or providing features (e.g., image of a tree) of the environment.
- a non-avatar object can be a component of an avatar.
- a non- avatar object can include a wall that, in interaction with an avatar, either remains a wall or animates to deliver message directed to the interacting avatar.
- a non- avatar object can include an appendage or other part of a first avatar that, when triggered by a second avatar, provides to the second avatar selected information, such as personal information of the participant associated with the first avatar.
- a virtual environment system 8 comprises a network system 10 to and/or through which are coupled, via communication channels 12, one or more hosts 20.
- Each of the hosts 20 has associated therewith a participant 14 who interacts with the virtual environment via the respective station 20.
- the network system 10 comprises a selected topology.
- An example topology of the system 10 includes a star network having a centralized computing system to which all the hosts 20 are connected.
- This example topology provides: (i) the centralized computing system runs server software that administers the virtual environment (e.g., receiving from and transmitting to participants 14 data concerning interaction, motion and other selected features and behaviors of virtual objects); (ii) the hosts 20 run client software that controls the local virtual experience, including obtaining and transmitting data to the server respecting its participant's interaction with and in the environment ("interaction data") and multimedia data (such as images, sound, and or text) responsive to received data; and (iii) the server software controls the distribution of data, including interaction data received from each host 20, among all hosts 20.
- the centralized computing system can comprise a host 20.
- Another example topology employs direct connections among hosts 20, wherein (i) typical connections rely on protocols such as transmission control protocol (TCP) and/or user datagram protocol (UDP), (ii) server/client software is essentially distributed among each of the hosts 20 and (iii) each host's software sends its participant's interaction data to relevant other hosts 20.
- Yet another example topology employs direct connections wherein, rather than communicating interaction data among all participants one by one, a participant communicates their data to a selected group address (i.e., multicasting) such that each host 20 of a particular group address can choose whether to accept/reject the data. It is to be recognized that other topologies can be employed without departing from the principles of the invention, including, for example topologies that combine one or more features of the above topologies.
- a host 20 comprises a computing device 22 coupled, via connections 21, to each of one or more display devices 24 and one or more input/output (I/O) devices 26.
- the connections 21 typically comprise wire/cabling, but, in the case of one or more of the I/O devices 26 and display devices 24, the connections can be implemented using a wireless technology (e.g., infrared technology).
- the interaction connections 36 provide a human-machine-type link between the participant and the I/O devices 26 and/or the display device 24.
- system 8 provides a third person view of the environment
- system can provide either/both third and first person views, without departing from the principles of the invention. That is, system preferably is implemented so that the virtual environment is presented to its participant either in first person or third person view: (i) first person view provides for experiencing and interacting in the virtual environment effectively as if the participant is the avatar (e.g., as if the participant sees the environment effectively through the avatar's eyes) and (ii) third person view provides for experiencing and interacting in the virtual environment separate from the avatar (e.g., as if seeing the environment through a camera mounted at a view point effectively above the avatar and, in that regard, enabled to observe the avatar in the environment).
- first person view provides for experiencing and interacting in the virtual environment effectively as if the participant is the avatar (e.g., as if the participant sees the environment effectively through the avatar's eyes)
- third person view provides for experiencing and interacting in the virtual environment separate from the avatar (e.g., as if seeing the environment through a
- Each display device 24 has a screen 28, the screen imaging the environment 16, particularly setting/scenes thereof.
- the screen 28 images the participant's avatar 18, other avatars 17 and a non-avatar object 19 (e.g., a lightning bolt seeking to shock the avatar 18).
- a non-avatar object 19 e.g., a lightning bolt seeking to shock the avatar 18.
- a preferred display device 24 comprises a head-mounted display (HMD) equipped with spatial disposition technology (i.e., to detect head and/or eye motions relative to, e.g., the displayed virtual environment 16).
- HMD head-mounted display
- spatial disposition technology i.e., to detect head and/or eye motions relative to, e.g., the displayed virtual environment 16.
- other types of display devices 24 can be used without departing from the principles of the invention (e.g., a single or an array of monitors).
- the I/O devices 26 preferably include one or more data acquisition devices that provide data representative of the participant's interaction with and in the virtual environment.
- the I/O devices 26 preferably are responsive to the movements of the participant's body parts or sense-based socialization.
- the devices can include one or more of a keyboard, a microphone, audio speakers, olfactory generators/detectors, pointing devices (e.g., a mouse, trackball, touch pad, and/or a joystick), steering devices (e.g., steering wheel or bar, bicycle frame, virtual skis and/or skates), movement sensors, facial gesture detecting devices, virtual reality gloves, haptic suits and/or other haptic attire, as well as other data acquisition devices, alone or in combination, without departing from the principles of the invention.
- pointing devices e.g., a mouse, trackball, touch pad, and/or a joystick
- steering devices e.g., steering wheel or bar, bicycle frame, virtual skis and/or skates
- movement sensors
- the computing device 22 typically includes a general pu ⁇ ose computer 29 having a processor 30, a memory system 32 (e.g., volatile memory and/or nonvolatile mass memory) and supporting software 34.
- the software 34 preferably includes an operating system ("OS") 33, a virtual environment package 35 (e.g., either client, client/host or other software associated with the topology of the environment), and some number of other application programs 37.
- the OS 33 preferably supports multimedia, e.g., through a graphical user interface (GUI).
- GUI graphical user interface
- the OS 33 typically comprises a collection of component programs.
- the OS component programs generally include (i) device drivers associated with the respective I/O devices 26 and (ii) display device drivers associated with the display devices 24.
- the invention admits a broad range of OS architectures. In that regard, no particular commercial OS, or architectural characteristics thereof, are preferred for implementing this invention. Moreover, it is to be recognized that any implemented OS can be other than as described above, without departing from the principles of the invention. As an example, the OS 33 can omit, combine or re-arrange various of the device drivers, with or without adding new components.
- system 8 is depicted to have plural hosts 20, each comprising a computing device, and each implemented in association with but separate from the network system 10, it is to be recognized that other implementations can be used without departing from the principles of the invention.
- the system 8 comprises a single host 20 used by one or more participants via one or more display devices 24, wherein the host 20 provides the network system 10.
- a virtual environment system 8, as illustrated in Figure lb, comprises an avatar attention component 52 and a priority component 54.
- the avatar attention component 52 comprises an attention module 38 and an interaction control engine 39, both of the virtual environment package 35, together with associated hardware (whether disposed in one or more hosts 20 and/or the network system 10) and software (e.g., the implicated portions of the OS 33, other applications 37, and any other software of one or more hosts 20 and/or the network system 10)
- the priority component 54 comprises an priority module 40 of the virtual environment package 35, together with associated hardware (whether disposed in one or more hosts 20 and/or the network system 10) and software (e.g., the implicated portions of the OS 33, other applications 37, and any other software of one or more hosts 20 and/or the network system 10).
- either or both of the attention component 52 and the priority component 54 can be implemented in the absence of one or more of respective modules 38, 39, 40 and/or other software.
- the avatar attention component 52 employs one or more attention spaces 55, each such space 55 comprising a sense space 56 and a focus space 58.
- the sense space 56 preferably comprises a sense volume (e.g., a sense sphere 62) and a sense volume section (e.g., a sense cone 60).
- the sense cone 60 is contained by the sense sphere 62 and has a sense vector 64.
- the focus space 58 preferably comprises a focus volume (e.g., a focus sphere 68) and a focus volume section (e.g., a focus cone 66).
- the focus cone 66 is contained by the focus sphere 68 and has a focus vector 70.
- the focus cone 66 is also contained by the sense cone 60.
- Sense and focus spheres are sometimes referred to herein individually or collectively as “attention spheres”; sense and focus cones are sometimes referred to herein individually or collectively as “attention cones”; and sense and focus vectors are sometimes referred to herein individually or collectively as “attention vectors”.
- the attention component 52 is associated with one or more senses and, as described in greater detail below, its elements and parameters preferably are employed in the operation of the priority component 54.
- the attention component provides an attention space.
- That space's sense space 56 preferably is used to control the objects that the participant/avatar is enabled to perceive (e.g., perception being via rendering and control via priorities).
- that space's focus space 58 preferably not only is used to determine the prospective partners (e.g., avatars and/or other virtual objects) with which the participant's avatar is enabled to interact spontaneously, but also preferably is used in computing priorities.
- avatars 17a-c are all within the attention space 55 of avatar 18, but (i) only avatars 17a and 17b are in the attention sense space 56 of avatar 18 and (ii) only avatar 17a is also in the attention focus space 58 of avatar 18. (As described further hereinafter, the two spaces preferably contribute to priority-based rendering of avatars and other virtual objects.)
- Attention Component The discussion that follows describes the attention component 52 in additional detail, as respects the sense of sight. This detail, among other things, provides background to the later description of the priority component 54. It is to be recognized that senses other than sight generally are supported by the system, such support being similar to the details that follow for sight. (Throughout the remainder of this document, vectors generally are represented by underscored labels; and participants, avatars and objects generally are represented by italicized capital letters.)
- the attention component's attention space 55 has (i) a sense space 56 comprising a scanning space 100 and (ii) a focus space 58 comprising a gazing space 150.
- the scanning space 100 comprises a vision cone 102 having a vision vector 106.
- the vertex "A" 108 of the cone 102 is associated with a selected immersion point 103.
- the immersion point 103 typically is separate from the avatar 101, e.g., a view point.
- the immersion point 103 preferably is a part of the avatar 101.
- the immersion point 103 preferably is disposed at the center of the avatar's face 110.
- the immersion point 103 can be disposed at the top third of the avatar's face 110 (e.g., the "eyes") or in association with the avatar's body 112. It is to be recognized that these examples rely on the avatar having a face, eyes and/or a body, whereas in any particular implementation, the avatar may be absent one or more of these features.
- the attention vectors preferably are associated with the selected immersion point 103. Moreover, the attention vectors preferably are oriented at the immersion point 103. In first person view, because the immersion point 103 is associated with some portion of the avatar 101, the attention vectors are also associated with that avatar portion and preferably are oriented substantially pe ⁇ endicular to a surface of that portion.
- the vision vector (AA") 106 is co-linear with the longitudinal axis 114 of the cone 102 and has a length described by the value
- the vector 106 resides in the three dimensional space of a vision sphere 104.
- the avatar 101 preferably is disposed at the center of the vision sphere 104.
- the participant preferably is effectively so disposed.
- the vector's direction describes the center of effective attention and its length limits the distance within which attention can be directed.
- the gazing space 150 comprises a gaze cone 152 having a gaze vector 154.
- the gaze cone 152 preferably shares the vertex (A) 108 and immersion point 103 of the associated vision cone 102.
- the gaze vector (AA') 154 comprises the longitudinal axis 114 and has a length described by the value [AA'
- the vector 154 preferably is co-linear with the vision vector 106, directed along axis 114. However, it is understood that other orientations of vector 154 can be selected without departing from the principles of the invention.
- the length of the gaze vector 154 is less than or equal to the length of the vision vector 106 (i.e.,
- ⁇
- the gaze vector's direction describes the center of focus and its length limits the distance within which attention can be focused.
- the scanning and gazing spaces 100, 150 preferably are bounded.
- the gazing space 150 preferably is bounded by two surfaces: (i) the surface 156 of the gaze cone 152 and (ii) the surface 160 of a gaze sphere 158.
- the scanning space 100 preferably is bounded by two surfaces: (i) the surface 116 of the vision cone 102 and (ii) the surface 118 of the vision sphere 104.
- the center of the gaze and vision spheres 158, 104 is the immersion point 103 associated with the vertex 108 (e.g., A) of the respective cones 152, 102.
- the radius of the gaze sphere 158 is the length of the gaze vector 154 (e.g.,
- Bounding in first person view is illustrated in Figure 3 for the gazing space 150 and in Figure 2 for the scanning space 100. Similar bounding applies to the spaces in third person view.
- the gazing space 150 typically is contained by the scanning space 100.
- the spaces 100 and 150 preferably are bounded in part by respective spheres 104, 158, the spaces 100 and 150 preferably comprise conical sections of the respective spheres 104, 158.
- these spheres preferably comprise the bounding volumes in that the associated attention cones are thereby able to be enlarged within the respective spheres (i.e., by increasing a cone's internal angle of rotation) so as to, at an extreme and in the absence of any orientation change within the volume, encompass substantially all or entirely all of the sphere.
- the bounding of the scanning space 100 typically establishes effective limitations to such perception (e.g., as described below, rendering typically is limited to a predetermined space).
- the bounding of the gazing space 150 establishes effective limitations as to the focus in gazing.
- the vision sphere's surface 118 typically defines the avatar-based maximum perception limits
- the vision cone's surface 116 defines the avatar-based effective perception limits.
- the gaze cone's boundaries effectively limit the avatar-based visual focus as to the other avatars and other virtual objects.
- the system controls rendering of objects disposed in the vision sphere 104 based on computed priorities, the priority computations preferably being based on disposition of the object relative to the above described elements of a selected attention space.
- the use and bounding of attention spaces, spheres, cones, vectors and the like preferably is selectably applicable in first person view, third person view, or both. If both such views are implemented, it is preferred that they be independently configurable.
- first person view can be implemented to have associated therewith each of attention spaces, spheres, cones, vectors, vertex, and immersion point while, for third person view, having associated a combination of some or all of these elements.
- the participant in third person view, preferably is not constrained by the configuration of first person view: the third person immersion point typically is separate from the participant's avatar and the relevant scanning space preferably is a different, selectable portion of the virtual environment.
- elements of the third person spaces e.g., spheres, cones, etc
- the participant preferably has selected dimensional and spatial relationships with correlative elements of first person spaces.
- the third person vision sphere can be larger, smaller or equal in size to the first person vision sphere
- a supported third person vision cone can include some, all or none of the first person vision cone
- supported attention cones of third person view may contain some, all or none of the participant's avatar.
- supported third person spaces are implemented so that rendering enables the participant to perceive not only what they would perceive if in first person view, but also a selected portion of the virtual environment supplemental thereto. If both first and third person views are implemented, it is preferred that, when one view is selected, the attention spaces thereof control attention and priority computation (the "control view"). In such case, the attention spaces of the other view are off or defeated.
- the views can be otherwise implemented, without departing from the principles of the invention.
- Dispositions in the virtual environment of avatars, objects, immersion points and the like can be determined in various ways.
- the disposition of an avatar can generally be determined by the disposition of the vertex of an avatar's first person focus cone, as each such vertex typically is associated with a immersion point at the avatar (e.g., face or eyes).
- an avatar is within a participant's first or third person gaze cone when a selected focus cone vertex of the avatar is within the participant's first or third person gaze cone.
- the disposition of an avatar, and particularly non-avatar objects can be determined independently of any associated cone and/or vertex.
- the dispositions of these objects can be determined relative to the first person immersion point of the avatar.
- the dispositions of these and other objects, as well as the avatar can be determined based on one or more markers placed in the virtual environment or based on some other mechanism, all without departing from the principles of the invention.
- Such disposition as is described below, preferably is employed in determining priorities under this invention.
- explicit controls can be relevant to either/both interaction and priorities. As to priorities, certain explicit controls preferably are supported, these controls implicating participants, avatars (e.g., typically based on a participant's affirmative action) and non-avatar objects (e.g., based on programming). It is understood that explicit controls can be used in first and third person views.
- the system 8 preferably supports the following explicit controls respecting priorities:
- Show-Profile a control to display the profile of selected objects.
- the content of a profile preferably includes, among other information, the respective participant's likes, dislikes, hobbies, and personalities.
- This control can be variously implemented, including via a key stroke ⁇ 'p' ⁇ with a pointing device (e.g. a mouse) positioned on the object.
- Set-Profile a control to enable change in the content of an object's profile.
- a profile can be configured so that different versions or portions thereof can be provided and/or rendered accessible.
- the provided profile can be selected by the object's type.
- the provided profile can be selected by the avatar's associated participant, or by groups and/or categories of participants.
- participants appearing in a friend list associated with the profile can be enabled to receive a more-detailed version than participants that are not so listed.
- participants appearing in a pest list may be restricted entirely from receiving or otherwise accessing a profile.
- This control can be variously implemented, including by providing forms for displaying and changing profiles.
- Set-Priority a control to set parameters for computing priorities.
- a participant Whether in first or third person view, a participant preferably can selectively change priorities for objects encountered in the virtual environment. Examples of such objects are avatars (including the participant's avatar, specifically the visible parts thereof) and non-avatar objects (including parts of an avatar).
- this control enables the participant to raise the priority of the audio signal of a friend's avatar so that the participant can hear the friend's audio signal clearly whenever the friend enters the virtual environment.
- this control enables the participant to lower the priority of a non-partner object that is broadcasting unwelcome messages.
- the participant is also enabled, through this control, to configure the system for resolving conflicts associated with multiple objects contending for the participant's attention.
- This control can be variously implemented, including by providing forms and sliders.
- the system 8 comprises a priority component 54.
- the priority component 54 generally stated, provides, separately to each participant, priorities for objects within the participant's attention space or spaces.
- the priorities preferably are used to determine not only which objects to render, but also the rendering detail.
- the priorities preferably are also used to determine which objects should be sent over the network and at which level of quality.
- system resources can generally be used with enhanced efficiently.
- advantages of the priority-based rendering and data communication arrangement include, as examples, one or more of enhanced coordination of the priority and attention components, reduced complexity of and overall system loading (e.g., relating to rendering computations and communications), and enhanced allocation of system resources in supporting a virtual environment.
- the priorities preferably are used to provide rendering/communication in enhanced detail and/or quality for those objects that the participant (or developer) deems important.
- an object with which a participant is engaged in an interaction session preferably has a high priority and, as such, may be rendered so that, as to one or more selected senses, the participant experiences enhanced manifestations of the object's features.
- the details typically include the object's features, expressions, gestures and other manifestations; as such, the enhancements are directed to one or more of resolution (e.g., numbers of polygons), enhanced animation, and increased frame rates.
- the details typically include the object's voice(s), each such voice typically being speech, music or other articulation subject to being heard, alone or in combination, by the participant; as such, the enhancement is directed to one or more of volume, resolution, increased bandwidth, and reduced masking from other sounds (e.g., so that the voices are louder, clearer or otherwise distinguished from the voice's of other non-interacting objects).
- Priority preferably is determined in connection with selected parameters.
- exemplary parameters include one or more of (i) the position of an object relative to one or more of a participant's selected attention spaces, (ii) the direction, orientation and span of the selected sense and focus spaces, (iii) the participants' profiles, (iv) parameters explicitly set by a developer or participant for priority computation, (v) models of social interaction/cultural norms, and (vi) economic features, e.g., cost structures.
- one or more parameters preferably serve as base parameters, while the other parameters serve as modifying parameters.
- the priorities are first computed using the base parameters.
- the priorities are subject to modification based on computations using one or more of the modifying parameters.
- the context is first person view and two base parameters are described: (a) one base parameter respects the relative distance between (i) the participant's avatar A and (ii) another object B resident in the attention space associated with avatar A and (b) another base parameter respects the angle between (i) the vector AB associated with the relative positions of objects A and B and (ii) an attention vector associated with A.
- these base parameters are preferred in this first person context, it is to be recognized that other base parameters can be selected, without departing from the principles of the invention.
- one base parameter can be selected to be the relative distance between (i) the immersion point of participant A and (ii) another object B resident in the attention space associated with A and (b) another base parameter can be selected to be the angle between (i) the vector AB associated with the relative positions of object B and the immersion point of A and (ii) an attention vector associated with A.
- Priority-based rendering and data communication contemplates the employ of plural attention spaces, each such space being associated with a particular sense (e.g., sight, hearing or smell) as described above respecting the attention component.
- Rendering and communication in such case, is enabled to have separate priority structures for each sense.
- sight preferably has associated priority structures respecting the base parameters of a vision space
- hearing preferably has associated priority structures respecting base parameters of a hearing space, wherein the base parameters of the hearing space can be the same or different than those of the vision space.
- Priority-based rendering and data communication also contemplates concurrent support of plural senses, each such sense having associated therewith one or more attention spaces.
- objects in a virtual environment comprise source objects and media objects.
- Source objects include any object with which a participant can interact, e.g. avatars, dogs, and TV sets.
- Media objects include any object directed to a participants' senses when rendered, e.g. animation, video, sound, text, and graphics.
- more than one media object typically is associated with each source object.
- participants perceive each source object through its associated media objects.
- priorities preferably are initially assigned to source objects (e.g., base priorities) in a relevant attention space.
- each source object's media objects preferably take priorities according to the respective source object's priority.
- the media objects of a particular type and associated with a single source object preferably have a priority equivalent to the priority of the source object.
- priorities of certain graphics, video and image media objects are associated with a source object and the source object is a single avatar, such that the media objects' priorities are equal to the priority of the avatar.
- the base priorities of the media objects preferably are equal to the base priority of the avatar as to the selected hearing space.
- the priorities may be given to media objects in various implementations.
- the implementations include express assignment (e.g., the system stores each media object together with its priority) and implied assignment (e.g., the system relates the media objects to a priority by relating the media objects to a source object having the priority).
- the computation of priorities for each resident object B comprises (i) dividing A's relevant attention space 55 (see Figure lb) into regions and (ii) assigning a number to resident objects B based on the region in which each such object resides.
- the assignment preferably follows selected conditions, including: (a) a region disposed relatively closer to A (e.g., closer to the vertex of A's attention cones) is assigned a higher priority than a region farther away and (b) a region disposed relatively closer to A's attention vectors (e.g., vision vector AA ”) is assigned a higher priority than a region farther away.
- the disposition of a resident object B relative to A is measured by the distance ⁇ AB ⁇ , while its disposition relative to AA " is measured by angle ? (angle BAA ").
- angle BAA angle BAA
- the applicable attention space (or a selected portion thereof) is divided along two dimensions: (i) a radial dimension (R-dimension) related to the one or more radii of concentric spheres within the attention space and (ii) an angular dimension ( ⁇ -dimension) related to an angle ⁇ .
- R-dimension a radial dimension
- ⁇ -dimension an angular dimension
- the attention space is selectably divided in either equal, substantially equal or unequal intervals.
- each supported attention space preferably is a sphere 57.
- the R-dimension preferably divides such sphere 57 using concentric spheres, with a priority number being assigned to each region between successive spheres.
- the regions closer to the center of sphere 57 preferably receive higher priorities (e.g., smaller-valued priority numbers).
- the ⁇ dimension preferably divides such sphere 57 using coaxial cones whose vertices are at the center of the sphere 57.
- Priority numbers are assigned to the regions between successive cones.
- the innermost cone preferably receives the highest priority (e.g., the smallest-valued priority number), with the priorities decreasing for regions formed as the internal angle of rotation becomes larger.
- the size of the regions preferably is controlled by selecting the value of the dividing interval(s) associated with each dimension.
- smaller regions generally are provided by selecting relatively smaller values for the dividing interval(s).
- larger regions generally are provided by selecting relatively larger values for the dividing interval(s).
- n The values of "n” in the above expressions can be different for different division forms as well as for different senses. Developers can select “n” so as to provide finer or coarser regions. For example, developers may select "n” for coarse division if the rendering engine can only handle a small number of quality levels (e.g., four levels of animation).
- division forms for each dimension leads to various combinations. Of the possible combinations (including those based on division forms other than the examples), four are described below: (i) the combination of R-division and ⁇ -division; (ii) the combination of R 2 -division and cos 2-division; (iii) the combination of R-division and cos/?- division; and (iv) the combination of R -division and / ⁇ -division. It is understood that, for the senses of sight and hearing, the combination of R 2 -division and cos ?-division generally approaches the behavior of the real world. It is also understood that division along only a single dimension can be used in the priority computation, without departing from the principles of the invention.
- Each combination provides priority numbers to the objects inA's attention space.
- each combination preferably generates three base priority structures.
- a first structure contains priority numbers assigned in R-dimension, e.g., using R- or R 2 -division.
- a second structure contains priority numbers assigned in the ⁇ dimension, e.g., using ⁇ or cos 2-division.
- a third structure contains composite priority numbers from both dimensions.
- a structure is a base priority structure if the priority numbers thereof are computed from only the base parameter.
- a structure is a modified priority l o structure if the priority numbers thereof are computed from the base parameter and one or more of the modifying parameters.
- A is the ith object's identifier
- P is the priority number of A] using the division form indicated by type-of-division
- the '*' means zero or more pairs of (A thread P .
- Valid types-of- division include, for example, R, R2, B, C, RB, RC, R2C, and R2B, where R represents R- division, R2 represents R 2 -division, B represents ⁇ -division, C represents cos 2-division.
- the first four of the valid types-of-division indicate that priority numbers are generated by dividing the sense sphere along one dimension using only the indicated division form.
- the last four of the valid types-of-division indicate that composite priority numbers are generated by dividing the sense sphere along two dimensions using the indicated, corresponding division forms.
- higher priority can be associated with larger priority numbers.
- the priority numbers can be other than numbers; rather, any selected designation ranking priority in the attention space (e.g., by mapping the numbers P, in the priority structure to the selected ranking system) can be used without departing from the principles of the invention.
- developers can configure their respective systems as to priority computation. For example, developers can select (i) one of the above division forms, or other division forms, for computing priorities, (ii) all or a subset of the three priority structures and (iii) the parameters employed in implementing the combinations and priority structures, as well as the associated computations.
- the selected configuration generally is responsive to various factors, including, available system resources and specifics of the application domain (e.g., the assessed importance of each priority structure to controlling the rendering of objects of different media or of different aspects of the same media).
- the object identifiers in the base priority structure represent source objects.
- priority modification may replace affected source objects by associated media objects.
- source objects generally are represented by italicized capital letters; media objects are generally denoted using the associated source object letter, together with an extension, e.g. ".m”.
- the attention space's sphere 57 is divided using R-division and ⁇ -division.
- D be the distance between avatar A and an object resident in A's attention space
- d be the radial interval used to divide the sphere 57 in the R-dimension (i.e., using concentric spheres)
- P d be the priority numbers assigned to resident objects.
- P d is:
- P d D /d
- Table 1 :
- the attention space's sphere 57 is also divided using coaxial cones, the cones having common axis AA ".
- P b angular interval used division
- P b the priority number assigned to an object resident in regions of this space.
- Table 2 shows exemplary values of P b for ranges of ?when b-22.5° . In the first column, for instance, Table 2 shows that, when ?is between 0 ° and 22.5°, P b has a value between 0 and 1.
- the priority numbers of objects within A 's focus cone are decreased in value, thereby increasing the associated priority. If the decrease is by an amount designated by P g , the composite priority numbers Pi for each object in A's attention space become:
- Figure 4 shows the projection of the priority regions in a plane containing the center of the attention space's sphere 57 and containing the sense and focus vectors.
- the attention space's sphere 57 is divided into 27 regions (24 due to rotational divisions and 3 due to the focus cone).
- the area enclosed by the heavy lines is the projection of A 's focus cone.
- the attention space's sphere 57 is divided using R 2 -division and cos 2-division.
- Advantages of using these division forms include that the divisions approximately comport with human visual and auditory perceptions, these perceptions being generally proportional to ⁇ IR 2 and 1/cos ⁇ .
- D 2 be the square of the distance between avatar A and an object resident in A's attention space
- d2 be the radial interval used to divide the sphere 57 in the R -dimension (i.e., using concentric spheres)
- P d ⁇ be the priority numbers assigned to the resident object.
- the attention space's sphere 57 is also divided using coaxial cones, the cones having common axis AA ".
- Priority numbers are again assigned to the regions between successive coaxial cones.
- the inner-most cone preferably receives the smallest- valued priority number, with the values increasing as the internal angle of rotation becomes larger. More specifically, however, in the ⁇ -dimension, it is preferred to assign smaller (i.e., higher priority) values to the larger cosine-valued regions (i.e., regions closer to AA ").
- the priority numbers of objects within A's focus cone e.g., the gaze cone
- the priority numbers of objects within A's focus cone preferably can be decreased in value, thereby increasing the associated priority. If the decrease is by an amount designated by P g , the composite priority numbers P 2 for each object in A's attention space become:
- this approach generates one or more of the following base priority structures:
- R2 ⁇ R2 (A, P d21 ) (A 2 P d22 ) ... (A Repeat P d2n ) ⁇
- C ⁇ C (A 1 P cl ) (A 2 P c2 ) ... (A n P cn ) ⁇
- Figure 5 shows the projection of the priority regions in a plane containing the center of the attention space's sphere 57 and containing the sense and focus vectors.
- the attention space's sphere 57 is again divided into 27 regions.
- the area enclosed by the heavy lines is again the projection of A s focus cone.
- the priority numbers assigned to the regions are the same except the regions unevenly divide the attention space's sphere 57 in both the R- and ⁇ dimensions.
- C Computing priority using R -division and cosffdivision.
- P d comprises the priority numbers assigned to resident objects using R-division
- P c comprises the priority numbers assigned to resident objects using cos ?-division.
- the computations for P d and P c are:
- the computation of P d can include integer operations, without departing from the principles of the invention.
- the weight coefficients have the same import as previously described.
- this approach generates one or more of the following base priority structures:
- Figure 6 shows the projection of the priority regions in a plane containing the center of the attention space's sphere 57 and containing the sense and focus vectors.
- the attention space's sphere 57 is again divided into 27 regions and the area enclosed by the heavy lines again is the projection of A 's focus cone.
- the priority numbers assigned to the regions are the same except the regions unevenly divide the attention space's sphere 57 in the / ⁇ dimension.
- the priority numbers assigned to the regions are the same except the regions evenly divide the attention space's sphere 57 in the R-dimension.
- the attention space's sphere 57 is divided using R 2 -division and / ⁇ -division.
- Pd2 comprises the priority numbers assigned to resident objects using R -division
- P b comprises the priority numbers assigned to resident objects using ⁇ division.
- the computations for E 2 and P b are:
- the priority numbers of objects within A's focus cone e.g., the gaze cone
- the composite priority numbers P for each object in A's attention space become:
- this approach generates one or more of the following base priority structures:
- R2 ⁇ R2 (A, P d21 ) (A 2 P d22 ) ... (A n P d2n ) ⁇
- R2B ⁇ R2B (A ! P 41 ) (A 2 P 42 ) ... (A n P 4n ) ⁇
- Figure 7 shows the projection of the priority regions in a plane containing the center of the attention space's sphere 57 and containing the sense and focus vectors.
- the attention space's sphere 57 is again divided into 27 regions and the area enclosed by the heavy lines again is the projection of A 's focus cone.
- the priority numbers assigned to the regions are the same except the regions unevenly divide the attention space's sphere 57 in the R-dimension.
- the priority numbers assigned to the regions are the same except the regions evenly divide the attention space's sphere 57 in the ⁇ -dimension.
- the priority numbers assigned to the regions are the same except the regions evenly divide the attention space's sphere 57 in the ⁇ -dimension and unevenly divide the attention space's sphere 57 in the R-dimension.
- priority numbers for objects within A's focus cone (e.g., the gaze cone), so as to increase priority.
- the priority numbers can also be decreased based on whether or not objects are in A' s sense cone. Such decrease can be in place of or together with the decrease based on the focus cone.
- priority numbers Pi use of both cones effects (i) a decrease for the focus cone by an amount designated by Pf and (ii) a decrease for the sense cone by an amount designated by P s .
- W d , w b , w and w s comprise weight coefficients, the coefficients being individually and/or collectively selectable. Similar reductions can be applied in the other priority structures described above.
- base parameters preferably are employed to establish one or more base priority structures, these priority structures relating a priority number to each respective object in the applicable attention space.
- the priority numbers so computed preferably are subject to modification.
- the modification computations are based on one or more modifying parameters.
- One such modifying parameter is a participant's profile which profile comprises, in a general sense, one or more lists of data directed to the participant's categorical views and opinions. Examples of the lists that may be provided include: friends, pests, public figures, and interests.
- Other modifying parameters can relate to (i) economics, (ii) parameters explicitly set by a developer or participant, and (iii) whether the participant is engaged in an interaction session. It is understood that other modifying parameters can be employed, without departing from the principle of the invention.
- the object identifiers in the base priority structure reference source objects.
- the priority component preferably provides priorities in steps, the steps comprising: 1) computing base priority numbers; 2) modifying the base priority numbers, so computed, according to modifying parameters; and 3) updating relevant priority structures based on the results of the modification.
- the priority component preferably is implemented so that object identifiers remain unchanged in relevant priority structures.
- the priority component preferably is implemented so that, in relevant priority structures, the source object's identifier is replaced by the identifiers of all its associated media objects.
- a participant preferably is enabled to use the Set-Priority control to select, for priority computations, modifying parameters associated with the participant's views and opinions.
- the views and opinions are directed to other participants/objects in the environment.
- the views and opinions typically include likes and dislikes.
- Friends ⁇ F ⁇ , F2 , ... , F m ⁇ is a list of objects (e.g., other participants and/or their avatars) that the participant likes
- Pests ⁇ S ⁇ , S 2 , ...
- SJ is a list of objects (e.g., other participants and/or their avatars) that the participant dislikes.
- the value off (sometimes referred to as a blocking factor) preferably increases with the degree to which the participant desires to monitor member i.
- the value of 5, (sometimes referred to as a blocking factor) increases with the degree to which the participant desires to block signals transmitted from member i.
- the numbers assigned to each of E, and S preferably are assigned automatically by the system. These automatically assigned numbers can be generated in various ways. As an example, the numbers can be assigned as defaults, e.g., developer or participant selected constants for each respective list.
- the numbers can be progressive, according to a predetermined calculus, responsive to the location of any given entry in the respective lists.
- the participant preferably is enabled to alter the assigned numbers, e.g., by employing the Set-Priority control.
- the numbers can be originally assigned by manual entry on the participant's part. For example, a participant can enter the originally-assigned priority of an object by using a slider in the Set-Priority control.
- the base (or otherwise modified) priority number P 2 becomes: 2J ⁇ — P 21 -f ⁇ - If object A, in structure R2C corresponds to S ⁇ in Pests, the base (or otherwise modified) priority number P 2 , becomes: P 2 ,
- Friends and Pests can be a single list. If a single list is used, the system preferably is enabled to assign a blocking factor b ⁇ to object A. This blocking factor contributes to the priority calculation according to the formula: P 2l — P 2l + b a .
- priority increases inversely with the priority number's value, (a) if b ⁇ > 0 (e.g., object A is in a Pests list), the larger the magnitude of b ⁇ , the lower A's priority and (b) if b ⁇ ⁇ 0 (e.g., object A is in a Friends list), the larger the absolute value of b ⁇ , the higher A's priority.
- a negative-valued blocking factor preferably can be employed to indicate that the participant desires to receive the signals from an object and, accordingly, can be used to request rendering enhancement.
- the modification of the priority of a source object A -based on a correlative entry in a Friends or Zo a Pests list— correspondingly modifies priority numbers of the media objects.
- Such modification of the source object's priority preferably works a global, uniform modification of the priorities of the associated media objects. It is also preferred, however, that a system supports direct modification of priorities as to selected media objects based on Friends and/or Pests lists. In this case, the priorities of selected media objects are modifiable other than via modification of the associated source object's priority.
- the priorities of the media objects may be modified (i) by adjusting the source object's correlative entry in a Friends or a Pests list or (ii) by entering and/or adjusting the entry of selected one or more of the media objects in a Friends or a Pests list.
- the source object's entry in a priority structure is (A, pcetate) and in the Friends list is (F, f) and (b) only the audio media object's priority is to be modified (e.g., by a blocking factor au ) using the Friends list
- the source object's entry in the Friends list preferably is replaced by entries for each of the audio, video and animation media objects as follows: (F,.m au d 10 f a u) (F,.m vldeo f) (F,.m amm fi), where the blocking factor for the audio media object is given by and
- Virtual environments generally support public figures that broadcast in one or more media to all participant hosts, e.g., the audio stream of a public speaker or an advertisement.
- a participant preferably is enabled to use the Set-Priority control to select, for priorities computation, a modifying parameter associated with public figures' broadcasts.
- the participant is enabled to select not only whether it desires to receive broadcasts from one or more categories of public figures, but also what media signals or broadcasts are received and at what priority.
- the participant selects by altering one or more default settings associated with such broadcasts.
- the resultant priority numbers of B follow the computation: As with Friends/Pests, in the case where priority increases inversely with the priority number's value, the larger the magnitude of b,'s positive value, the greater the priority number and, in turn, the lower the priority of signals from the public figure B, Conversely, the larger the absolute value of b policy ifb, is negative, the lesser the priority number and, in turn, the higher the priority.
- priority modification of a source object preferably is implemented the same as, or similar to, the modifying procedures previously described with Friends/Pests lists.
- priority modification of a source object correspondingly modifies priority numbers of the associated media objects, such corresponding modification preferably being global and uniform.
- priority modification can be effected directly as to selected media objects. In this case, while the priority numbers of only selected media objects are modified, the entry of the associated source object in a priority structure is replaced by all its media objects, including media objects that are not so modified.
- the priorities of the media objects may be modified (i) by adjusting the source object's priority in the Public priority structure or (ii) by entering and/or adjusting the entry of selected one or more of the media objects in the Public priority structure.
- the blocking factor is subtracted in this formula (cf, the formula when using a single Friends! Pests list) and providing that priority increases inversely with the priority number's value, (a) if b ⁇ x > 0, the blocking factor works to increase B,.m x 's priority (e.g., magnify signals) and (b) if b tx ⁇ 0, the blocking factor works to lower B,.m x 's priority (e.g., block signals).
- the Public priority structure (resultant or, if not modified, default) preferably is a part of the overall priority structures (base and/or modified) containing numbers associated with objects in the applicable attention space's sphere 57.
- a participant preferably is enabled to use the Set-Priority control to select a modifying parameter associated with having shared interests with other objects.
- participants use the Set-Priority control to instruct the system to compute a measure of interest similarity between themselves and other objects resident in the applicable attention space. This similarity measure preferably is employed to alter base priority numbers.
- a relatively high value for a similarity measure of a desirable object preferably is associated with a reduction in the avatar's priority number and, accordingly, results in enhanced priority.
- a relatively low value for a similarity measure of an undesirable object e.g., the avatar of a lawyer who is soliciting clients
- a relatively low-valued similarity measure can be associated with an object (e.g., a physicist's avatar), but nevertheless can result in reduction of the avatar's priority number, the reduction serving to enhance priority.
- interests are selected by one or more categories and types.
- the interests ultimately provide for increasing or decreasing priorities, or even leaving priorities unchanged.
- the base (or otherwise modified) priority number P associated with an object A, preferably is modified by the similarity measure, as follows: P, ⁇ -P, + w s ⁇ m *s*sim,.
- the priority number P is modified by the similarity measure, as follows: P, ⁇ — P, —w s ⁇ m *s*sim,.
- s is a positive scaling factor
- w s ⁇ m is a weighting factor.
- the scaling factor s is directed to bring the value of sim, into a comparable range as P,. For example, if -5 ⁇ P, ⁇ 10, s can be 10.
- the weighting factor w s ⁇ m preferably is determined — by the system and/or participant — to reflect the importance of the similarity measure relative to other factors in the overall priority number.
- priority modification for an object based on a similarity measure preferably is implemented the same as, or similar to, the modifying procedures previously described.
- priority modification of a source object correspondingly modifies priority numbers of the associated media objects, such corresponding modification preferably being global and uniform among the media objects.
- priority modification can be effected directly as to selected media objects.
- the system can enable a participant (e.g., via the Set-Priority control) to configure the similarity measure so that it contributes to priority modification of selected media objects, such objects being restricted to specified types (e.g., text and audio).
- the entry of the associated source object in a priority structure is replaced by all its media objects, including media objects that are not so modified.
- the computation provides a fair framework to take into account that (a) one or more mixtures of list types can appear in participant profiles, e.g., any subset of likes, dislikes, friends, pests and other supported lists, (b) one or more lists can contain any subset of concepts, such concepts being expressed in potentially divergent vocabulary of any language common to the participant's avatar and a resident object, and (c) the items in the lists can be weighted;
- the similarity measure's computation employs a cos ⁇ computation in multi-dimensional space, where ⁇ is the angle between the two profile vectors, the profile vectors being derived ultimately from profile structures.
- ⁇ is the angle between the two profile vectors, the profile vectors being derived ultimately from profile structures.
- the developer and/or participant selects the profile structures for the computation.
- the selected profile structures can be interest structures, such as the lists likes and dislikes.
- this example includes two lists of specific, complementary types, it is understood that other numbers of lists and/or other types of structures can be employed, including non-complementary lists, without departing from the principles of the invention.
- the selected interest structures preferably are organized to include (i) one or more topics t, (ii) associated weights w, the weights preferably being integers and, in any case, indicating the participant's interest level in the topic (e.g., a larger weight for an entry in a likes list indicates a higher favorable interest in the entry, whereas a larger weight for an entry in a dislikes list indicates a higher unfavorable interest in the entry).
- subscripts a and b associate the lists to participants A and B, respectively,
- subscripts / and d differentiate likes entries from dislikes entries and
- integer subscripts indicate different topics.
- a similarity computation preferably follows the steps:
- the merge step provides for topics in the interest list to comprise the union of the topics of the merged lists.
- the weights for the topics in both the likes and dislikes lists preferably are positive integers; however, because the two are complementary types, the weights for the topics in the dislikes list preferably are negated in the interest list. Accordingly, if a particular topic appears in both the likes and dislikes lists, the net weight w x! - w x is used in the interest list.
- Interest list a ⁇ (t a ⁇ , w a u) ... (t alm , w aim ) (t ad ⁇ , -w adl ) ... (t ad disturb , -w adn ) ⁇
- Interest list b ⁇ (t b u , w b u) ... (t a ⁇ P , w b ⁇ p ) (t bd ⁇ , - >MI) ⁇ ⁇ ⁇ (hdq , -w bd q)j
- U a ⁇ (tall) ⁇ ⁇ ⁇ (tal J adl) ⁇ ⁇ ⁇ (t a dn)( ll)( li) ⁇ ⁇ -(t b l p )(t bd2 ) ... (t bdq ) ⁇ where the subscript "a" in U a indicates the union is for A's computations.
- V a [W a li , ⁇ ⁇ ⁇ , W alm , -W ad l , ... , -W a dn , 0 , ... , 0 , 0 , ... ,0]
- V b [w b!2 , 0 . ... , 0, 0 , ... -W b di, 0, w b u, w b i3 , ... , w blp , -w bd2 , ... ,-w bda ]
- the value of the similarity measure ranges from -1 to 1.
- positive values preferably are inte ⁇ reted to indicate relatively higher similarity than is indicated by negative values and zero.
- positive values of greater magnitude preferably are inte ⁇ reted to indicate relatively higher similarity than is indicated by positive values of lesser magnitude.
- negative values of greater absolute values preferably are inte ⁇ reted to indicate higher dissimilarity than is indicated by negative values of lesser absolute values.
- a similarity measure valued at 1 indicates identical interest lists, i.e., identical topics with identical weights.
- a similarity measure valued at -1 indicates opposite interest lists.
- Opposite interest lists arise when the topics in a first participant's likes list are identical to the topics in the second participant's dislikes list and the topics in the first participant's dislikes list are identical to the topics in second participant's likes list, with the corresponding weights being identical.
- the similarity results for participants/avatars encountered by a participant are stored for a selected time period. Stored results can be employed in the event that implicated system resources are sufficiently loaded to preclude computation of a current similarity measure.
- Stored results for such employ reflect an assumption that profile structures tend to remain relatively unchanged over a time period (e.g., an empirical, calculated or selected time period) following the encounter associated with the stored result. Even so, it is preferred that stored results are not employed after the applicable time period has elapsed. It is also preferred that the similarity measure be re-computed once resources become available.
- Previously un-encountered participants may arrive in A's attention space, bearing a profile structure that is not reflected by the existing union.
- arrival preferably triggers the system to re-compute the similarity measures, taking into account the new profile structure.
- the system if the system resources are loaded, the system preferably employs the existing union to estimate the similarity measure between A and the new participant/avatar. It is also preferred that, when resources become available, the similarity measures are re-computed.
- the system can be implemented to provide A with an active-visitors data store for storing and updating the relevant profile structures of participants in consideration.
- the union is computed using the profile structures in this data store.
- a timer is associated with each participant's data store entry.
- a timer is associated with that participant, the timer being set to 0. Thereafter, if the participant leaves the space, the timer is incremented periodically.
- the participant's profile structure preferably is removed from the active-visitors data store. However, if the participant returns to the space before the removal threshold is exceeded, the participant's timer is reset to 0. Moreover, upon return, the participant's entry is updated, e.g, as to the participant's current profile structure. It is to be recognized that the detection of the participant's entry, exit and re-entry of the space can be variously implemented in the system 8, without departing from the principles of the invention.
- a participant has an associated avatar A, as well as an attention space in which avatars B and C are currently resident. Moreover, avatars B and C are the only avatars in A's active-visitors data store.
- V a [5, -A, -6, 0, 4, 0, 0, 5]
- V b [-2, -6, 5, 8, 0, 6, 0, 2]
- c [8, -1, -2, 0, 7, 0, 3, 5]
- the computations indicate that participant of avatar A has interests that are more similar to those of the participant of avatar C, than to those of the participant of avatar B.
- the results can also be inte ⁇ reted to indicate that avatars A and B are slightly dissimilar, whereas avatars A and C are relatively similar.
- the similarity measure preferably is implemented so that topics include source and/or media objects.
- the similarity measure preferably is implemented so that topics include concept objects.
- Concept objects generally are described in words and phrases.
- the similarity measure is subject to linguistic issues.
- the same topic is subject to being expressed in different words/phrases.
- a single word/phrase may refer to different topics.
- One solution to these issues is to employ a semantic network or thesaurus to identify synonyms and to map concept objects. This solution, however, tends to significantly increase computational complexity, particularly if it is used every time a word/phrase is entered by a participant.
- this solution is implemented as an interest specification tool 300, the tool comprising an interest specifier component 302 and a vocabulary specifier component 304.
- the rectangles in the Figure represent logic, the ovals represent data, and the faces represent participants 306 and 308.
- the participant 306 employs the interest specifier component 302 which component draws on the resources of a concept hierarchy 310 to provide a predefined vocabulary. If a participant 308 seeks to customize the vocabulary, the participant 308 employs the vocabulary specifier component 304 which component draws on the resources of a semantic network 312 and/or thesaurus 314. More specifically, the network 312 and thesaurus 314, through the vocabulary specifier component 304 enable the participant 308 both to map words provided by the participant 308 to a selected concept and to enter the concept at the specified location in the concept hierarchy 310. Once the concept is so entered, it preferably is available as well to non-mapping participant 306. Because vocabulary extensions preferably arise infrequently (e.g., due to the implementation of the interest specifier component 302) it is preferred to support the operation of the vocabulary specifier component 304 with the data associated in association with the semantic network 312 and/or thesaurus 314.
- the concept hierarchy 310 preferably contains concepts that represent interests most relevant in the respective application domain.
- participant 306 tries to set-up or modify interest lists 315, e.g., likes 316 and dislikes 318, the participant 306 is enabled to select (e.g., by dragging) a concept from the hierarchy 310 and supply it (e.g., by dropping) in the relevant list 316, 318.
- the participant 306 preferably is also enabled to move a concept object from one list to another, and/or to delete a concept object from a list.
- the concept hierarchy 310 preferably organizes its concepts in levels.
- the top level of the hierarchy 310 contains most general, root concepts. Traversing to lower levels of the hierarchy provides progressively specific concepts, the traversing preferably being subject to a selected total number of levels.
- the top level of the hierarchy may contain entertainment, recreation, and professional interests. Under entertainment reside movies, music and activities. Under recreation, there are travel and sports. Under music reside classic, rock, country, jazz, etc. Under rock reside soft, hard, heavy metal, etc.
- a concept in the above hierarchy 310 is selected, it is represented by concatenating the concepts along the path from the root concept to the selected concept. That is, the concepts at higher levels provide the context for a concept at the lower levels.
- each concept preferably is a weight indicating the participant's interest level (e.g., like or dislike).
- the weight preferably is subject to a default value which value is maintained in the hierarchy 310 and is modifiable by the participant when the concept is added to an interest list 315.
- the participant is provided with a selected interface (e.g., a slider and/or an editable text area adjacent to the concept display) for specifying the concepts' respective weights. It is understood that, although the interface can provide for controlling the range of selectable weight values, such control can be omitted without departing from the principles of the invention, particularly if the similarity computation, as implemented, has a range between -1 and 1. If a weight is not specified for a lower level concept, the weight of a specified next-higher level concept preferably is employed.
- one name can refer to plural participants and/or one participant can be referred to by different names.
- the system preferably offers a list of selectable participant for use, e.g., to complete friends, pests and other lists.
- priority numbers can be established (i.e., base priority numbers) or modified responsive to economic features associated with the virtual environment or an application thereof.
- an environment application can be implemented to comprise a cost structure that enables information providers (e.g., advertisers and underwriters) to raise the priority of their content-invested objects (hereafter sometimes referred to as an information source).
- the cost structure effectively degrades one or more blocking factors set by participants (e.g., public speakers, similarity, content) that might apply to such information source.
- participants e.g., public speakers, similarity, content
- the cost structure preferably also enables participants, as information consumers, to increase their blocking factors.
- the information provider and the participants are so enabled, but at a cost, whether through payment of money or fees, or though some other transaction.
- the cost structure provides n levels of pricing, the pricing levels reflected in two lists: ⁇ cj, c 2 , ... Cn ⁇ and ⁇ ei, e 2 , ... e n ⁇ .
- Each such list contains preferably n monotonically increasing, non-negative numbers.
- the values of coefficients c be set sufficiently large —relative to the values of coefficients e-- so as to reduce in part or entirely a participant's ability to eliminate rendering of advertisements/underwriters messages.
- some or all of the weights and blocking factors associated with previously described base and modifying priority numbers are treated as base factors to be adjusted based on a participant's coefficient ⁇ j prefatory to being used in the priority computation.
- ⁇ j a participant's coefficient
- e, 1.2 and the adjustment is by multiplication
- the adjustment increases the base factors by 20%.
- e / l and the adjustment is by multiplication
- assignment of this coefficient results in the use of unchanged base factors in the priority computation. This assignment may be applied to participants who pay the lowest fees.
- This special case can be implicated so that participants subject to this coefficient pay a fee so as to obtain any ability to establish or modify priorities and/or have any blocking power associated with weights and blocking factors.
- the information source is assigned an initial priority of -c, for information providers who pay the amount associated with coefficient c curry the information source is assigned an initial priority of -c,. Assuming lower- valued priority numbers (i.e., higher valued c,) conelate to higher priority, the more a provider pays for a respective information source, the higher the initial priority of the source.
- the source's initial priority typically is higher than the highest blocking factor any participant can have without paying, because, as stated above, ci » ei in the above lists.
- the participant's ability to reduce the priority of the information sources depends on the participant's choices respecting the cost structure.
- priority modification for an object based on cost structures preferably is implemented the same as, or similar to, the modifying procedures previously described.
- priority modification of a source object correspondingly modifies priority numbers of the associated media objects.
- all associated media objects preferably take these assignments.
- priority modification can be effected directly as to selected media objects.
- a cost structure can provide for assignment of different c's to different media objects of one source object (e.g. video and audio from an advertiser).
- a system 8 preferably is implemented to enhance social interaction among avatars. Accordingly, if avatars A and B are engaged in an interaction session, the system 8 preferably is implemented so that B's priority P B is highest among all objects in avatar A's attention space. Similarly, the system 8 preferably is implemented so that the members of an interacting group (whose avatars are also in A's focus cone) have higher priorities than the priorities of other non-member avatars.
- Enhanced social interaction may be variously accomplished using priorities. In one case, it is accomplished by assigning appropriate values to the previously-described modifying parameter P g .
- P g is employed, together with a weight factor, to reduce the priority numbers of objects within A's focus cone (e.g., the gaze cone) are decreased in value, thereby increasing the associated priority of such objects.
- an additional priority modifying parameter is assigned.
- This parameter designated as M max , provides for modifying the priority number P B assigned to B when interacting with A in a pair-wise interaction context, as follows: P B - P B - M max .
- M max can be computed using the formula:
- M max N R + N B + N w + N s
- N R is the maximum priority number assignable in a radial dimension of a relevant attention space
- N B is the maximum priority number assignable in an angular dimension of a relevant attention space
- N w is the maximum priority number assignable to objects based on A's profile
- N s is the maximum priority number assignable to objects based on shared interests.
- M max can be computed using other formulae, including, as an example, formulae that include more, less or different parameters (e.g., a subset of the parameters set forth above) or different operations.
- modification thereof based on M max may be comprehensive or selective (e.g., modify P ⁇ 's based on having correspondence to a parameter included in the selected M max formula).
- priority modification for an object based on social interaction preferably is implemented the same as, or similar to, the modifying procedures previously described.
- a system 8 preferably is implemented to enable developers to specify different priorities for different media objects associated with a single source object.
- the robot's entry in the R2C structure is ( ⁇ P B )
- the priority of mechanical sound becomes 0.5 P R
- the priority of music becomes 0.25P R
- the priority of animation becomes 0.33P R such that, after these modifications, the robot's entry (i.e., the source object entry) is replaced by ( ⁇ speec h PR) ( ⁇ mech 0.5P R ) ( ⁇ . mus ⁇ c 0.25P R ) ( ⁇ . amm 0.33P R ).
- a participant preferably is enabled to use the Set-Profile and/or Set-Priority controls to select, for priority computation, a separate (or modifying) parameter associated with the participant's interests in non-avatar objects' content.
- the participant is enabled to specify concept words and phrases and to instruct the system to increase (decrease) priority if an object's content relates to the so-specified concept.
- Each word can also have a weight associated with it to indicate the participant's interest level. For example, a larger weight w, associated with word, increases the priority of the object that contains the word.
- CON indicates the type of the structure
- Greater magnitude for positive- valued w preferably indicate that the participant has higher interest levels related to the concept t, and, therefore, desires to experience objects related to concept t, (e.g., through an enhanced rendering or communication).
- the priority numbers associated with an object subject to a concept t preferably have their priority number P x modified as follows: P x ⁇ — P x - w,.
- priority modification for an object based on concept structures preferably is implemented the same as, or similar to, the modifying procedures previously described.
- the three base priority structures are modified using various modifying parameters and, preferably, are supplemented by separate structures derived, as to objects, from information in participant profiles.
- the output from the priority computation(s) preferably contains one or more of the following structures, where '*' means zero or more indicated pairs: • Priority structures: ⁇ type-of-division (A, PJ* ⁇
- the structure RAW contains raw data, where R, is the distance between A and an object ARIC cos ⁇ , is cosine of the angle between vector , and 's relevant focus vector (e.g., a gaze vector), and u, is the unit focus vector relating A with A, (e.g., the relevant focus vector divided by the magnitude of that vector).
- the source object's entry in one or more relevant priority structures preferably is replaced by all its media objects. That is, after selected modifying parameters have been applied, the identifier and priority of source objects are replaced by identifiers of associated media objects and applicable priorities.
- the output from the priority computation(s) preferably contains one or more of the following structures, where '*' means zero or more indicated pairs.
- developers can choose which output structures are to be computed based on their needs, on the resources available and on the function for application.
- the output structures may be applied to rendering.
- selected output structures can be passed as input to associated rendering engines (e.g., audio rendering engine, video rendering engine, and/or avatar animation rendering engine) and can be used to control the rendering of objects.
- the selected output structures can also be passed as input to operating systems and/or network systems for data processing and communication. Examples of Using Priority Numbers in Rendering
- R2C's priority numbers P 2 j are assumed to have been modified using one or more modifying parameters (e.g, as described above or otherwise). It is understood that other base priority structures can be used, without departing from the principles of the invention. It is also understood that, although only a single partner is included in the interaction session, plural partners can be so included, without departing from the principles of the invention.
- Avatar Animation In one embodiment of operation, objects disposed outside the vision cone are not considered when assigning priorities based on the R2C structure. For such embodiment, then, the entry associated with A* is not considered. Moreover, considering only those avatars resident in the vision cone, structure R2C is sorted in the order of highest to lowest priority, i.e., by increasing priority numbers: ⁇ R2C (A 3 -10) (A 5 1) (A) 2) (A 2 3) ⁇ . If avatars are animated in 4 levels of detail, the most detailed animation level
- level 4 preferably is assigned to avatar A 3 , as such avatar has the highest priority.
- the least detailed animation level i.e., level 1
- intermediate levels e.g., level 3 and level 2
- M max is employed to elevate to highest priority an object(s) with which A is involved in pair-wise interaction (e.g., here A 3 ).
- This example also illustrates a system 8 implemented so that the members of an interacting group (whose avatars are also in A's focus cone) have higher priorities than the priorities of other, non-member avatars.
- P g is employed.
- the participant's perception of signals from group members e.g., A
- the avatar preferably is considered in the overall rendering determination.
- the system 8 of the example can employ the concept structure).
- AVs priority number drops from 6 to 1, which number corresponds to a priority above that of all other avatars, other than A 3 .
- the developer can yet choose not to render as it is outside the vision cone or, if choosing to render, can choose to render at level 3 or some lower (i.e., less detailed) level. If the rendering of A is placed at level 3, avatars A ⁇ , A 5 , and A 2 are rendered at some lower level of detail.
- a system 8 according to the invention can be implemented enabling objects to share a priority number.
- the system preferably provides one or more mechanisms for mediating the assignment of rendering levels.
- a mediating mechanism assigns a single intermediate rendering level (e.g., 3) to both priority-sharing objects.
- a mediating mechanism assigns rendering levels upon reference to additional output priority structures. As an example, if the R2C structure above had assigned priority number 2 to both Ai and A 2 , a mediating mechanism could be implemented to make reference to one of structures C and R2 to determine, by lower priority number, the object to be rendered at the higher priority.
- avatar A 3 is assigned both the highest resolution and the highest frame rate.
- Avatar A] is assigned the next level of resolution, followed by A 2 , then A 5 .
- Avatar A 5 is assigned the next level of frame rate.
- assignment of the next levels of frame rate is subject to mediation in that Aj and A 2 have the same priority number in structure C.
- various mediation mechanisms can be implemented. As an example, a mechanism can employ the concept structure in mediating such assignments.
- concept structure can also be used to selectively render avatars having selected features.
- a participant's concept structure can take the following form: ⁇ CON (blue eyes and red hair 10) ⁇ .
- the rendering engine triggers the appropriate media recognition technology to isolate avatars bearing such features and, as to isolated avatars and based on the priority weighting 10, the rendering engine preferably provides enhanced rendering for such avatars and/or features.
- system can be implemented to entirely block the rendering of audio associated with avatars disposed outside the sense cone.
- the system can be implemented not to so block such audio signals.
- the system controls the amplitude of the audio.
- the system can be implemented to employ the priority numbers of structure R2C so as to control objects' audio according to selected criteria, e.g., amplitude attenuation is by 3dB for every unit difference between the associated priority number and a reference value, wherein the reference value can be variously selected, including, for example, as (a) M max or (b) the value of the priority number of the highest-priority avatar under consideration by A .
- the sorted structure is: ⁇ R2C (A 3 -10) (A 5 1) (Ai 2) (A 2 3) ( * 6) ⁇
- a 3 's audio signal is not attenuated
- the audio signals of avatars A 5 , Ai, A , and A are attenuated by, respectively, -33dB,-36dB, -39dB, and -48dB.
- the interacting partner A 3 can be heard the loudest, followed by the interacting group member A 5 , followed by the others.
- the system can be implemented to use the R2C structure to select the audio streams to be rendered, while using the RA W structure to generate 3D sound.
- the system determines the content of the audio as part of the rendering decision.
- avatar A's concept structure has the following values: ⁇ CON (high blood pressure 5) (heart attack 3) (smoking 2) ⁇ .
- ⁇ CON high blood pressure 5
- heart attack 3 heart attack 3
- speech is anticipated and the system is implemented to provide a transcription mechanism.
- the transcription mechanism transcribes the audio into text. If the text is determined to contain the phrases set forth above or some effectively related values, the audio is passed to the rendering engine with enhanced priority.
- the system can also be implemented to control the amplitude of the audio, e.g., to increase the audio amplitude by ldB for every unit increase in the associated weight from some reference weight.
- the reference weight has a value of 0
- the amplitude of the audio stream containing high blood pressure is increased by 5 dB
- the amplitudes of the audio streams containing heart attack and smoking are increased, respectively, by 3 and 2 dB.
- Developers preferably are enabled to select how to implement priority rendering for sound, e.g., based on their needs and on the resources available. As an example, developers can determine whether to render a selected number of words (or sentences or paragraphs) around a concept phrase or even to render an entire stream.
- the priorities can be employed (i) to render objects, (ii) to participate in resolving conflicts among objects contending for attention and (iii) to manage system and network resources.
- the use of priorities can be to enhance immersion in the virtual environment.
- the system described above can be applied to senses other than vision. This application is particularly the case with those senses that have similar characteristics as vision, e.g., auditory and olfactory senses. However, it is to be understood that such senses may be characterized by different ranges and different sensitivities, e.g., sensitivities to the direction of attention focus.
- Each implemented sense preferably has at least one associated attention space.
- Each such attention space preferably is characterized by a sense sphere, sense cone, sense vector, focus sphere, focus cone, and focus vector.
- the radius of a sense sphere preferably is adjusted to reflect a perception range of the applicable sense.
- the direction of a focus vector represents the center of the attention.
- the length of the focus vector represents the distance of the attention focus, and the focus cone's internal angle of rotation determines the breadth of attention focus.
- a participant cannot perceive, when in first person view, signals of sources disposed outside a first person sense cone, which cone may fill part of or the entire sense sphere.
- the participant has enhanced ability to perceive the signals of sources disposed within a focus cone of the participant, as compared to the signals of sources disposed outside any such cone.
- a system can support priorities based on a plurality of senses.
- the computations as to each supported sense preferably are carried out independently.
- the system can be implemented to assign priorities to audio information sources using a hearing space for auditory signals, while assigning priorities to animation and/or video using one or more vision spaces.
- the Set-Attention control can be used, e.g., to set the radius for each sense/focus sphere, and the length and the angle for each sense/focus cone. It can also be used to choose attention spaces for use with each media type.
- Interaction in a two-dimensional (2D) virtual environment is a special case of the interactions in virtual environments of three or more dimensions.
- the priority component employs the projections of a focus cone and a focus vector in a two-dimensional (e.g., x-y) plane.
- the conditions and computations associated with priority computation apply in 2D cases as well, but with the components of the third dimension (e.g., z-axis) at 0.
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1019997009673A KR100705306B1 (en) | 1998-02-21 | 1999-02-18 | Priority-based virtual environment |
EP99902760A EP0978026B1 (en) | 1998-02-21 | 1999-02-18 | Priority-based virtual environment |
DE69904678T DE69904678T2 (en) | 1998-02-21 | 1999-02-18 | PRIORITY-BASED VIRTUAL ENVIRONMENT |
JP54230699A JP2001522499A (en) | 1998-02-21 | 1999-02-18 | Virtual environment based on priority |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/027,459 US6329986B1 (en) | 1998-02-21 | 1998-02-21 | Priority-based virtual environment |
US09/027,459 | 1998-02-21 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO1999042918A2 true WO1999042918A2 (en) | 1999-08-26 |
WO1999042918A3 WO1999042918A3 (en) | 1999-11-11 |
Family
ID=21837867
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB1999/000299 WO1999042918A2 (en) | 1998-02-21 | 1999-02-18 | Priority-based virtual environment |
Country Status (7)
Country | Link |
---|---|
US (1) | US6329986B1 (en) |
EP (1) | EP0978026B1 (en) |
JP (1) | JP2001522499A (en) |
KR (1) | KR100705306B1 (en) |
DE (1) | DE69904678T2 (en) |
TW (1) | TW421753B (en) |
WO (1) | WO1999042918A2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001312644A (en) * | 2000-04-28 | 2001-11-09 | Sony Corp | Information processor, information processing method and storage medium |
US20020169644A1 (en) * | 2000-05-22 | 2002-11-14 | Greene William S. | Method and system for implementing a management operations center in a global ecosystem of interrelated services |
KR20020097064A (en) * | 2002-09-18 | 2002-12-31 | 학교법인 한국정보통신학원 | Method for providing a scalable concurrency management in networked virtual environments |
WO2008088666A1 (en) * | 2007-01-16 | 2008-07-24 | Motorola, Inc. | System and method for managing interactions in a virtual environment |
WO2008088665A1 (en) * | 2007-01-16 | 2008-07-24 | Motorola, Inc. | Method and apparatus to facilitate multi-setting virtual reality experiences |
WO2023147044A1 (en) * | 2022-01-27 | 2023-08-03 | Meta Platforms Technologies, Llc | Scalable parallax system for rendering distant avatars, environments, and dynamic objects |
Families Citing this family (221)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4232232B2 (en) * | 1998-09-30 | 2009-03-04 | ソニー株式会社 | Information processing apparatus and method, and recording medium |
GB9902480D0 (en) * | 1999-02-05 | 1999-03-24 | Ncr Int Inc | Method and apparatus for advertising over a communications network |
US6664980B2 (en) * | 1999-02-26 | 2003-12-16 | Accenture Llp | Visual navigation utilizing web technology |
US6473851B1 (en) * | 1999-03-11 | 2002-10-29 | Mark E Plutowski | System for combining plurality of input control policies to provide a compositional output control policy |
US7389356B2 (en) * | 1999-12-15 | 2008-06-17 | Microsoft Corporation | Generalized differentiation methods and arrangements for adaptive multimedia communications |
US7210104B2 (en) * | 2000-02-16 | 2007-04-24 | Sega Corporation | Information display method and information display system for finding another user in a plurality of users based upon registered information |
KR20000030430A (en) * | 2000-02-29 | 2000-06-05 | 김양신 | Internet advertisement system using virtual space based on 3 dimension graphic |
US6672961B1 (en) * | 2000-03-16 | 2004-01-06 | Sony Computer Entertainment America Inc. | Computer system and method of displaying images |
US7561994B1 (en) * | 2000-06-29 | 2009-07-14 | Corsi Adam B | Method for a virtual pregnancy experience |
US6766376B2 (en) | 2000-09-12 | 2004-07-20 | Sn Acquisition, L.L.C | Streaming media buffering system |
US7788323B2 (en) * | 2000-09-21 | 2010-08-31 | International Business Machines Corporation | Method and apparatus for sharing information in a virtual environment |
JP2002109361A (en) * | 2000-09-28 | 2002-04-12 | Sanyo Electric Co Ltd | Method and device for displaying advertisement |
KR20030035138A (en) * | 2001-10-30 | 2003-05-09 | 한국전자통신연구원 | Transmit method of state message in client-server-based networked virtual environment |
US20030177019A1 (en) * | 2002-03-18 | 2003-09-18 | Santos Richard A. | System and method for visual representation of an electroni community |
US20040043373A1 (en) * | 2002-09-04 | 2004-03-04 | Kaiserman Jeffrey M. | System for providing computer-assisted development |
KR100721524B1 (en) * | 2002-10-16 | 2007-05-23 | 한국전자통신연구원 | Method and system for transforming adaptively visual contents according to user's symptom characteristics of low vision impairment and user's presentation preferences |
US8037150B2 (en) | 2002-11-21 | 2011-10-11 | Aol Inc. | System and methods for providing multiple personas in a communications environment |
US7636755B2 (en) | 2002-11-21 | 2009-12-22 | Aol Llc | Multiple avatar personalities |
US7272658B1 (en) | 2003-02-13 | 2007-09-18 | Adobe Systems Incorporated | Real-time priority-based media communication |
US7908554B1 (en) | 2003-03-03 | 2011-03-15 | Aol Inc. | Modifying avatar behavior based on user action or mood |
US20040179039A1 (en) | 2003-03-03 | 2004-09-16 | Blattner Patrick D. | Using avatars to communicate |
US7913176B1 (en) | 2003-03-03 | 2011-03-22 | Aol Inc. | Applying access controls to communications with avatars |
WO2005059678A2 (en) * | 2003-12-11 | 2005-06-30 | Chau Minh Vuong | System and method for providing identification and search information |
KR100452089B1 (en) | 2004-06-23 | 2004-10-13 | 엔에이치엔(주) | The image resource loading system and method which carries out loading of the object for renewal of a game screen |
US7388585B2 (en) * | 2004-09-20 | 2008-06-17 | My Virtual Reality Software | Method, system and device for efficient distribution of real time three dimensional computer modeled image scenes over a network |
US9652809B1 (en) | 2004-12-21 | 2017-05-16 | Aol Inc. | Using user profile information to determine an avatar and/or avatar characteristics |
US7793259B2 (en) * | 2005-03-02 | 2010-09-07 | Microsoft Corporation | System and method for managing user interaction data in a networked environment |
WO2006092647A1 (en) * | 2005-03-04 | 2006-09-08 | Nokia Corporation | Offering menu items to a user |
US7375678B2 (en) * | 2005-06-29 | 2008-05-20 | Honeywell International, Inc. | Displaying obstacles in perspective view |
US11403336B2 (en) | 2005-10-26 | 2022-08-02 | Cortica Ltd. | System and method for removing contextually identical multimedia content elements |
US9384196B2 (en) | 2005-10-26 | 2016-07-05 | Cortica, Ltd. | Signature generation for multimedia deep-content-classification by a large-scale matching system and method thereof |
US11216498B2 (en) | 2005-10-26 | 2022-01-04 | Cortica, Ltd. | System and method for generating signatures to three-dimensional multimedia data elements |
US10535192B2 (en) | 2005-10-26 | 2020-01-14 | Cortica Ltd. | System and method for generating a customized augmented reality environment to a user |
US8266185B2 (en) | 2005-10-26 | 2012-09-11 | Cortica Ltd. | System and methods thereof for generation of searchable structures respective of multimedia data content |
US9218606B2 (en) | 2005-10-26 | 2015-12-22 | Cortica, Ltd. | System and method for brand monitoring and trend analysis based on deep-content-classification |
US10742340B2 (en) | 2005-10-26 | 2020-08-11 | Cortica Ltd. | System and method for identifying the context of multimedia content elements displayed in a web-page and providing contextual filters respective thereto |
US10621988B2 (en) | 2005-10-26 | 2020-04-14 | Cortica Ltd | System and method for speech to text translation using cores of a natural liquid architecture system |
US10191976B2 (en) | 2005-10-26 | 2019-01-29 | Cortica, Ltd. | System and method of detecting common patterns within unstructured data elements retrieved from big data sources |
US9477658B2 (en) | 2005-10-26 | 2016-10-25 | Cortica, Ltd. | Systems and method for speech to speech translation using cores of a natural liquid architecture system |
US20160321253A1 (en) | 2005-10-26 | 2016-11-03 | Cortica, Ltd. | System and method for providing recommendations based on user profiles |
US10614626B2 (en) | 2005-10-26 | 2020-04-07 | Cortica Ltd. | System and method for providing augmented reality challenges |
US10949773B2 (en) | 2005-10-26 | 2021-03-16 | Cortica, Ltd. | System and methods thereof for recommending tags for multimedia content elements based on context |
US8326775B2 (en) | 2005-10-26 | 2012-12-04 | Cortica Ltd. | Signature generation for multimedia deep-content-classification by a large-scale matching system and method thereof |
US9767143B2 (en) | 2005-10-26 | 2017-09-19 | Cortica, Ltd. | System and method for caching of concept structures |
US11386139B2 (en) | 2005-10-26 | 2022-07-12 | Cortica Ltd. | System and method for generating analytics for entities depicted in multimedia content |
US9372940B2 (en) | 2005-10-26 | 2016-06-21 | Cortica, Ltd. | Apparatus and method for determining user attention using a deep-content-classification (DCC) system |
US10585934B2 (en) | 2005-10-26 | 2020-03-10 | Cortica Ltd. | Method and system for populating a concept database with respect to user identifiers |
US10848590B2 (en) | 2005-10-26 | 2020-11-24 | Cortica Ltd | System and method for determining a contextual insight and providing recommendations based thereon |
US10607355B2 (en) | 2005-10-26 | 2020-03-31 | Cortica, Ltd. | Method and system for determining the dimensions of an object shown in a multimedia content item |
US8312031B2 (en) | 2005-10-26 | 2012-11-13 | Cortica Ltd. | System and method for generation of complex signatures for multimedia data content |
US9953032B2 (en) | 2005-10-26 | 2018-04-24 | Cortica, Ltd. | System and method for characterization of multimedia content signals using cores of a natural liquid architecture system |
US10698939B2 (en) | 2005-10-26 | 2020-06-30 | Cortica Ltd | System and method for customizing images |
US10360253B2 (en) | 2005-10-26 | 2019-07-23 | Cortica, Ltd. | Systems and methods for generation of searchable structures respective of multimedia data content |
US11019161B2 (en) | 2005-10-26 | 2021-05-25 | Cortica, Ltd. | System and method for profiling users interest based on multimedia content analysis |
US10372746B2 (en) | 2005-10-26 | 2019-08-06 | Cortica, Ltd. | System and method for searching applications using multimedia content elements |
US11032017B2 (en) | 2005-10-26 | 2021-06-08 | Cortica, Ltd. | System and method for identifying the context of multimedia content elements |
US10635640B2 (en) | 2005-10-26 | 2020-04-28 | Cortica, Ltd. | System and method for enriching a concept database |
US9191626B2 (en) | 2005-10-26 | 2015-11-17 | Cortica, Ltd. | System and methods thereof for visual analysis of an image on a web-page and matching an advertisement thereto |
US10380267B2 (en) | 2005-10-26 | 2019-08-13 | Cortica, Ltd. | System and method for tagging multimedia content elements |
US8818916B2 (en) | 2005-10-26 | 2014-08-26 | Cortica, Ltd. | System and method for linking multimedia data elements to web pages |
US10193990B2 (en) | 2005-10-26 | 2019-01-29 | Cortica Ltd. | System and method for creating user profiles based on multimedia content |
US9031999B2 (en) | 2005-10-26 | 2015-05-12 | Cortica, Ltd. | System and methods for generation of a concept based database |
US10691642B2 (en) | 2005-10-26 | 2020-06-23 | Cortica Ltd | System and method for enriching a concept database with homogenous concepts |
US10380164B2 (en) | 2005-10-26 | 2019-08-13 | Cortica, Ltd. | System and method for using on-image gestures and multimedia content elements as search queries |
US10776585B2 (en) | 2005-10-26 | 2020-09-15 | Cortica, Ltd. | System and method for recognizing characters in multimedia content |
US10387914B2 (en) | 2005-10-26 | 2019-08-20 | Cortica, Ltd. | Method for identification of multimedia content elements and adding advertising content respective thereof |
US9646005B2 (en) | 2005-10-26 | 2017-05-09 | Cortica, Ltd. | System and method for creating a database of multimedia content elements assigned to users |
US11361014B2 (en) | 2005-10-26 | 2022-06-14 | Cortica Ltd. | System and method for completing a user profile |
US11003706B2 (en) | 2005-10-26 | 2021-05-11 | Cortica Ltd | System and methods for determining access permissions on personalized clusters of multimedia content elements |
US10380623B2 (en) | 2005-10-26 | 2019-08-13 | Cortica, Ltd. | System and method for generating an advertisement effectiveness performance score |
US11620327B2 (en) | 2005-10-26 | 2023-04-04 | Cortica Ltd | System and method for determining a contextual insight and generating an interface with recommendations based thereon |
US9747420B2 (en) | 2005-10-26 | 2017-08-29 | Cortica, Ltd. | System and method for diagnosing a patient based on an analysis of multimedia content |
US9529984B2 (en) | 2005-10-26 | 2016-12-27 | Cortica, Ltd. | System and method for verification of user identification based on multimedia content elements |
US10180942B2 (en) | 2005-10-26 | 2019-01-15 | Cortica Ltd. | System and method for generation of concept structures based on sub-concepts |
US11604847B2 (en) | 2005-10-26 | 2023-03-14 | Cortica Ltd. | System and method for overlaying content on a multimedia content element based on user interest |
US8633927B2 (en) | 2006-07-25 | 2014-01-21 | Nvidia Corporation | Re-render acceleration of frame with lighting change |
KR101396363B1 (en) * | 2006-09-04 | 2014-05-19 | 삼성전자주식회사 | Apparatus and Method for Searching Broadcasting Channel |
US10733326B2 (en) | 2006-10-26 | 2020-08-04 | Cortica Ltd. | System and method for identification of inappropriate multimedia content |
US7966567B2 (en) * | 2007-07-12 | 2011-06-21 | Center'd Corp. | Character expression in a geo-spatial environment |
US20080215975A1 (en) * | 2007-03-01 | 2008-09-04 | Phil Harrison | Virtual world user opinion & response monitoring |
US8756103B1 (en) | 2007-03-28 | 2014-06-17 | Qurio Holdings, Inc. | System and method of implementing alternative redemption options for a consumer-centric advertising system |
US8135018B1 (en) | 2007-03-29 | 2012-03-13 | Qurio Holdings, Inc. | Message propagation in a distributed virtual world |
US8116323B1 (en) | 2007-04-12 | 2012-02-14 | Qurio Holdings, Inc. | Methods for providing peer negotiation in a distributed virtual environment and related systems and computer program products |
US8601386B2 (en) | 2007-04-20 | 2013-12-03 | Ingenio Llc | Methods and systems to facilitate real time communications in virtual reality |
US20080263446A1 (en) * | 2007-04-20 | 2008-10-23 | Utbk, Inc. | Methods and Systems to Connect People to Services via Virtual Reality |
US20080262910A1 (en) * | 2007-04-20 | 2008-10-23 | Utbk, Inc. | Methods and Systems to Connect People via Virtual Reality for Real Time Communications |
US20080263459A1 (en) * | 2007-04-20 | 2008-10-23 | Utbk, Inc. | Methods and Systems to Determine Availability for Real Time Communications via Virtual Reality |
US20080263460A1 (en) * | 2007-04-20 | 2008-10-23 | Utbk, Inc. | Methods and Systems to Connect People for Virtual Meeting in Virtual Reality |
US8000328B1 (en) | 2007-05-22 | 2011-08-16 | Qurio Holdings, Inc. | Filtering messages in a distributed virtual world based on virtual space properties |
US8560387B2 (en) * | 2007-06-07 | 2013-10-15 | Qurio Holdings, Inc. | Systems and methods of providing collaborative consumer-controlled advertising environments |
US8600808B2 (en) * | 2007-06-07 | 2013-12-03 | Qurio Holdings, Inc. | Methods and systems of presenting advertisements in consumer-defined environments |
US8006231B2 (en) * | 2007-06-26 | 2011-08-23 | International Business Machines Corporation | Apparatus and method to automate the testing of a graphical user interface |
US9003304B2 (en) * | 2007-08-16 | 2015-04-07 | International Business Machines Corporation | Method and apparatus for moving an avatar in a virtual universe |
US8345049B2 (en) * | 2007-08-16 | 2013-01-01 | International Business Machine Corporation | Method and apparatus for predicting avatar movement in a virtual universe |
US7990387B2 (en) * | 2007-08-16 | 2011-08-02 | International Business Machines Corporation | Method and apparatus for spawning projected avatars in a virtual universe |
US9111285B2 (en) * | 2007-08-27 | 2015-08-18 | Qurio Holdings, Inc. | System and method for representing content, user presence and interaction within virtual world advertising environments |
KR101742256B1 (en) | 2007-09-26 | 2017-05-31 | 에이큐 미디어 인크 | Audio-visual navigation and communication |
US8041704B2 (en) | 2007-10-12 | 2011-10-18 | The Regents Of The University Of California | Searching for virtual world objects |
JP4985970B2 (en) * | 2007-10-24 | 2012-07-25 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Technology for controlling the display of objects |
US20090141023A1 (en) * | 2007-11-29 | 2009-06-04 | Brian Mark Shuster | Selective filtering of user input data in a multi-user virtual environment |
US8127235B2 (en) | 2007-11-30 | 2012-02-28 | International Business Machines Corporation | Automatic increasing of capacity of a virtual space in a virtual world |
CA2705400A1 (en) * | 2007-12-07 | 2009-07-02 | France Telecom | Method of managing user representations, corresponding managing entity, terminal and computer program |
US20090164919A1 (en) | 2007-12-24 | 2009-06-25 | Cary Lee Bates | Generating data for managing encounters in a virtual world environment |
US20090177976A1 (en) | 2008-01-09 | 2009-07-09 | Bokor Brian R | Managing and presenting avatar mood effects in a virtual world |
US20090192785A1 (en) * | 2008-01-29 | 2009-07-30 | Anna Carpenter Cavender | System and method for optimizing natural language descriptions of objects in a virtual environment |
KR100927009B1 (en) * | 2008-02-04 | 2009-11-16 | 광주과학기술원 | Haptic interaction method and system in augmented reality |
WO2009104564A1 (en) | 2008-02-20 | 2009-08-27 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Conversation server in virtual space, method for conversation and computer program |
US8230441B2 (en) | 2008-04-29 | 2012-07-24 | International Business Machines Corporation | Virtual world subgroup determination and segmentation for performance scalability |
US8819565B2 (en) * | 2008-05-14 | 2014-08-26 | International Business Machines Corporation | Describing elements in a virtual world based on changes since a previous encounter |
US8648875B2 (en) | 2008-05-14 | 2014-02-11 | International Business Machines Corporation | Differential resource applications in virtual worlds based on payment and account options |
US8671349B2 (en) * | 2008-05-15 | 2014-03-11 | International Business Machines Corporation | Virtual universe teleportation suggestion service |
US10380604B2 (en) | 2008-07-17 | 2019-08-13 | International Business Machines Corporation | Method and system for viral marketing within a virtual world |
US20100017283A1 (en) * | 2008-07-21 | 2010-01-21 | International Business Machines Corporation | Dynamic advertising systems and methods for virtual universes |
US10037565B2 (en) * | 2008-07-24 | 2018-07-31 | International Business Machines Corporation | System and method for prioritization of rendering policies in virtual environments |
US10699282B2 (en) * | 2008-08-05 | 2020-06-30 | International Business Machines Corporation | Method and system for automatic optimal advertising determination within a virtual universe |
US9256346B2 (en) | 2008-08-11 | 2016-02-09 | International Business Machines Corporation | Managing ephemeral locations in a virtual universe |
US10115113B2 (en) * | 2008-08-11 | 2018-10-30 | International Business Machines Corporation | Immersive advertisements in a virtual universe |
US20100036735A1 (en) * | 2008-08-11 | 2010-02-11 | International Business Machines Corporation | Triggering immersive advertisements in a virtual universe |
US8296246B2 (en) * | 2008-09-02 | 2012-10-23 | International Business Machines Corporation | Allocating virtual universe customer service |
US8315409B2 (en) * | 2008-09-16 | 2012-11-20 | International Business Machines Corporation | Modifications of audio communications in an online environment |
US8347235B2 (en) | 2008-09-26 | 2013-01-01 | International Business Machines Corporation | Method and system of providing information during content breakpoints in a virtual universe |
US20100088600A1 (en) * | 2008-10-07 | 2010-04-08 | Hamilton Ii Rick A | Redirection of an avatar |
US20100091015A1 (en) * | 2008-10-15 | 2010-04-15 | Robert Eric Heidel | Product, service, and market integration of three dimensional modeling/rendering software for the construction, remodeling, manufacturing, designing, buying, and/or selling of homes, businesses, structures, vehicles, and/or buildings |
US8051287B2 (en) | 2008-10-15 | 2011-11-01 | Adobe Systems Incorporated | Imparting real-time priority-based network communications in an encrypted communication session |
US8310378B2 (en) * | 2008-10-20 | 2012-11-13 | Honeywell International Inc. | Method and apparatus for displaying prioritized photo realistic features on a synthetic vision system |
US20100113159A1 (en) * | 2008-11-06 | 2010-05-06 | International Business Machines Corporation | Method and apparatus for partitioning virtual worlds using prioritized topic spaces in virtual world systems |
US8988421B2 (en) * | 2008-12-02 | 2015-03-24 | International Business Machines Corporation | Rendering avatar details |
US9741147B2 (en) * | 2008-12-12 | 2017-08-22 | International Business Machines Corporation | System and method to modify avatar characteristics based on inferred conditions |
US8681144B2 (en) * | 2008-12-19 | 2014-03-25 | International Business Machines Corporation | Prioritized rendering of objects in a virtual universe |
US9472014B2 (en) | 2008-12-19 | 2016-10-18 | International Business Machines Corporation | Alternative representations of virtual content in a virtual universe |
US8898574B2 (en) * | 2008-12-19 | 2014-11-25 | International Business Machines Corporation | Degrading avatar appearances in a virtual universe |
US9805492B2 (en) * | 2008-12-31 | 2017-10-31 | International Business Machines Corporation | Pre-fetching virtual content in a virtual universe |
US8126985B1 (en) | 2008-12-31 | 2012-02-28 | Qurio Holdings, Inc. | Prioritizing virtual object downloads in a distributed virtual environment |
US9633465B2 (en) | 2009-02-28 | 2017-04-25 | International Business Machines Corporation | Altering avatar appearances based on avatar population in a virtual universe |
US8570325B2 (en) * | 2009-03-31 | 2013-10-29 | Microsoft Corporation | Filter and surfacing virtual content in virtual worlds |
US9100435B2 (en) | 2009-04-02 | 2015-08-04 | International Business Machines Corporation | Preferred name presentation in online environments |
US8990306B2 (en) | 2009-05-28 | 2015-03-24 | International Business Machines Corporation | Pre-fetching items in a virtual universe based on avatar communications |
US8051195B1 (en) | 2009-06-04 | 2011-11-01 | Qurio Holdings, Inc. | Method and system for providing data streams in a virtual environment |
US8972870B2 (en) * | 2009-08-27 | 2015-03-03 | International Business Machines Corporation | Providing alternative representations of virtual content in a virtual universe |
US9256896B2 (en) * | 2009-08-27 | 2016-02-09 | International Business Machines Corporation | Virtual universe rendering based on prioritized metadata terms |
US9700804B2 (en) | 2009-09-29 | 2017-07-11 | International Business Machines Corporation | Method and system for accurate rating of avatars in a virtual environment |
US10108972B2 (en) * | 2009-12-01 | 2018-10-23 | International Business Machines Corporation | Evaluating advertising effectiveness in a virtual universe |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
US20120249797A1 (en) | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
US9285589B2 (en) | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
US9759917B2 (en) | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
US8467133B2 (en) | 2010-02-28 | 2013-06-18 | Osterhout Group, Inc. | See-through display with an optical assembly including a wedge-shaped illumination system |
US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
US20150309316A1 (en) | 2011-04-06 | 2015-10-29 | Microsoft Technology Licensing, Llc | Ar glasses with predictive control of external device based on event input |
US8477425B2 (en) | 2010-02-28 | 2013-07-02 | Osterhout Group, Inc. | See-through near-eye display glasses including a partially reflective, partially transmitting optical element |
US8488246B2 (en) | 2010-02-28 | 2013-07-16 | Osterhout Group, Inc. | See-through near-eye display glasses including a curved polarizing film in the image source, a partially reflective, partially transmitting optical element and an optically flat film |
US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
US8482859B2 (en) | 2010-02-28 | 2013-07-09 | Osterhout Group, Inc. | See-through near-eye display glasses wherein image light is transmitted to and reflected from an optically flat film |
CN102906623A (en) | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | Local advertising content on an interactive head-mounted eyepiece |
US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
US8472120B2 (en) | 2010-02-28 | 2013-06-25 | Osterhout Group, Inc. | See-through near-eye display glasses with a small scale image source |
US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
CA2807552A1 (en) | 2010-08-06 | 2012-02-09 | Moderna Therapeutics, Inc. | Engineered nucleic acids and methods of use thereof |
CN102446192A (en) * | 2010-09-30 | 2012-05-09 | 国际商业机器公司 | Method and device for estimating attention in virtual world |
US8937620B1 (en) | 2011-04-07 | 2015-01-20 | Google Inc. | System and methods for generation and control of story animation |
EP2705435B8 (en) | 2011-05-06 | 2017-08-23 | Magic Leap, Inc. | Massive simultaneous remote digital presence world |
US20120290949A1 (en) * | 2011-05-09 | 2012-11-15 | Idle Games | System and method for facilitating user interaction in a virtual space through unintended tangential effects to primary user interactions |
US8184069B1 (en) | 2011-06-20 | 2012-05-22 | Google Inc. | Systems and methods for adaptive transmission of data |
US9098069B2 (en) | 2011-11-16 | 2015-08-04 | Google Technology Holdings LLC | Display device, corresponding systems, and methods for orienting output on a display |
US20130159068A1 (en) | 2011-12-19 | 2013-06-20 | Kabam, Inc. | System and method for determining quality of service for actions to be performed in a virtual space |
US8988349B2 (en) | 2012-02-28 | 2015-03-24 | Google Technology Holdings LLC | Methods and apparatuses for operating a display in an electronic device |
US8947382B2 (en) * | 2012-02-28 | 2015-02-03 | Motorola Mobility Llc | Wearable display device, corresponding systems, and method for presenting output on the same |
US10147146B2 (en) * | 2012-03-14 | 2018-12-04 | Disney Enterprises, Inc. | Tailoring social elements of virtual environments |
US9755847B2 (en) * | 2012-12-19 | 2017-09-05 | Rabbit, Inc. | Method and system for sharing and discovery |
WO2015039276A1 (en) * | 2013-09-17 | 2015-03-26 | Empire Technology Development Llc | Automatic contacts sorting |
US9858720B2 (en) | 2014-07-25 | 2018-01-02 | Microsoft Technology Licensing, Llc | Three-dimensional mixed-reality viewport |
US9766460B2 (en) | 2014-07-25 | 2017-09-19 | Microsoft Technology Licensing, Llc | Ground plane adjustment in a virtual reality environment |
US10416760B2 (en) | 2014-07-25 | 2019-09-17 | Microsoft Technology Licensing, Llc | Gaze-based object placement within a virtual reality environment |
US9865089B2 (en) | 2014-07-25 | 2018-01-09 | Microsoft Technology Licensing, Llc | Virtual reality environment with real world objects |
US10451875B2 (en) | 2014-07-25 | 2019-10-22 | Microsoft Technology Licensing, Llc | Smart transparency for virtual objects |
US9904055B2 (en) | 2014-07-25 | 2018-02-27 | Microsoft Technology Licensing, Llc | Smart placement of virtual objects to stay in the field of view of a head mounted display |
US10311638B2 (en) | 2014-07-25 | 2019-06-04 | Microsoft Technology Licensing, Llc | Anti-trip when immersed in a virtual reality environment |
DE102015004749A1 (en) * | 2015-04-11 | 2016-10-13 | Audi Ag | Method for operating a virtual reality glasses and system with a virtual reality glasses |
US10799792B2 (en) * | 2015-07-23 | 2020-10-13 | At&T Intellectual Property I, L.P. | Coordinating multiple virtual environments |
US11037015B2 (en) | 2015-12-15 | 2021-06-15 | Cortica Ltd. | Identification of key points in multimedia data elements |
US11195043B2 (en) | 2015-12-15 | 2021-12-07 | Cortica, Ltd. | System and method for determining common patterns in multimedia content elements based on key points |
WO2019008581A1 (en) | 2017-07-05 | 2019-01-10 | Cortica Ltd. | Driving policies determination |
US11899707B2 (en) | 2017-07-09 | 2024-02-13 | Cortica Ltd. | Driving policies determination |
US11282133B2 (en) | 2017-11-21 | 2022-03-22 | International Business Machines Corporation | Augmented reality product comparison |
US10586360B2 (en) | 2017-11-21 | 2020-03-10 | International Business Machines Corporation | Changing view order of augmented reality objects based on user gaze |
US10565761B2 (en) | 2017-12-07 | 2020-02-18 | Wayfair Llc | Augmented reality z-stack prioritization |
US10206055B1 (en) * | 2017-12-28 | 2019-02-12 | Verizon Patent And Licensing Inc. | Methods and systems for generating spatialized audio during a virtual experience |
US10846544B2 (en) | 2018-07-16 | 2020-11-24 | Cartica Ai Ltd. | Transportation prediction system and method |
US10839694B2 (en) | 2018-10-18 | 2020-11-17 | Cartica Ai Ltd | Blind spot alert |
US20200133308A1 (en) | 2018-10-18 | 2020-04-30 | Cartica Ai Ltd | Vehicle to vehicle (v2v) communication less truck platooning |
US11181911B2 (en) | 2018-10-18 | 2021-11-23 | Cartica Ai Ltd | Control transfer of a vehicle |
US11126870B2 (en) | 2018-10-18 | 2021-09-21 | Cartica Ai Ltd. | Method and system for obstacle detection |
US11700356B2 (en) | 2018-10-26 | 2023-07-11 | AutoBrains Technologies Ltd. | Control transfer of a vehicle |
US10789535B2 (en) | 2018-11-26 | 2020-09-29 | Cartica Ai Ltd | Detection of road elements |
EP3671410B1 (en) * | 2018-12-19 | 2022-08-24 | Siemens Healthcare GmbH | Method and device to control a virtual reality display unit |
WO2020161816A1 (en) * | 2019-02-06 | 2020-08-13 | マクセル株式会社 | Mixed reality display device and mixed reality display device method |
US11643005B2 (en) | 2019-02-27 | 2023-05-09 | Autobrains Technologies Ltd | Adjusting adjustable headlights of a vehicle |
US11285963B2 (en) | 2019-03-10 | 2022-03-29 | Cartica Ai Ltd. | Driver-based prediction of dangerous events |
US11694088B2 (en) | 2019-03-13 | 2023-07-04 | Cortica Ltd. | Method for object detection using knowledge distillation |
US11132548B2 (en) | 2019-03-20 | 2021-09-28 | Cortica Ltd. | Determining object information that does not explicitly appear in a media unit signature |
US10776669B1 (en) | 2019-03-31 | 2020-09-15 | Cortica Ltd. | Signature generation and object detection that refer to rare scenes |
US10789527B1 (en) | 2019-03-31 | 2020-09-29 | Cortica Ltd. | Method for object detection using shallow neural networks |
US11488290B2 (en) | 2019-03-31 | 2022-11-01 | Cortica Ltd. | Hybrid representation of a media unit |
US11222069B2 (en) | 2019-03-31 | 2022-01-11 | Cortica Ltd. | Low-power calculation of a signature of a media unit |
US10796444B1 (en) | 2019-03-31 | 2020-10-06 | Cortica Ltd | Configuring spanning elements of a signature generator |
US11593662B2 (en) | 2019-12-12 | 2023-02-28 | Autobrains Technologies Ltd | Unsupervised cluster generation |
US10748022B1 (en) | 2019-12-12 | 2020-08-18 | Cartica Ai Ltd | Crowd separation |
US11590988B2 (en) | 2020-03-19 | 2023-02-28 | Autobrains Technologies Ltd | Predictive turning assistant |
US11827215B2 (en) | 2020-03-31 | 2023-11-28 | AutoBrains Technologies Ltd. | Method for training a driving related object detector |
US11756424B2 (en) | 2020-07-24 | 2023-09-12 | AutoBrains Technologies Ltd. | Parking assist |
JP7083380B2 (en) * | 2020-09-28 | 2022-06-10 | グリー株式会社 | Computer programs, methods and server equipment |
WO2022242854A1 (en) * | 2021-05-19 | 2022-11-24 | Telefonaktiebolaget Lm Ericsson (Publ) | Prioritizing rendering by extended reality rendering device responsive to rendering prioritization rules |
JP6995416B1 (en) * | 2021-06-09 | 2022-01-14 | クラスター株式会社 | Avatar output device, terminal device, avatar output method, and program |
US11756248B2 (en) * | 2021-08-23 | 2023-09-12 | International Business Machines Corporation | Computer graphics object rendering based on priority |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR970010089B1 (en) * | 1994-05-31 | 1997-06-21 | Samsung Aerospace Ind | Image input/output apparatus for encoding/decoding for video signal |
US5736982A (en) * | 1994-08-03 | 1998-04-07 | Nippon Telegraph And Telephone Corporation | Virtual space apparatus with avatars and speech |
KR100276549B1 (en) * | 1995-12-07 | 2000-12-15 | 이리마지리 쇼우이치로 | Image generation apparatus, image generation method, game machine using the method |
US5956039A (en) * | 1997-07-25 | 1999-09-21 | Platinum Technology Ip, Inc. | System and method for increasing performance by efficient use of limited resources via incremental fetching, loading and unloading of data assets of three-dimensional worlds based on transient asset priorities |
JP2007165063A (en) * | 2005-12-13 | 2007-06-28 | Samsung Sdi Co Ltd | Conductive particle, conductive paste, and electronic component |
-
1998
- 1998-02-21 US US09/027,459 patent/US6329986B1/en not_active Expired - Fee Related
-
1999
- 1999-02-18 KR KR1019997009673A patent/KR100705306B1/en not_active IP Right Cessation
- 1999-02-18 EP EP99902760A patent/EP0978026B1/en not_active Expired - Lifetime
- 1999-02-18 DE DE69904678T patent/DE69904678T2/en not_active Expired - Fee Related
- 1999-02-18 JP JP54230699A patent/JP2001522499A/en not_active Ceased
- 1999-02-18 WO PCT/IB1999/000299 patent/WO1999042918A2/en active IP Right Grant
- 1999-03-31 TW TW088105145A patent/TW421753B/en not_active IP Right Cessation
Non-Patent Citations (1)
Title |
---|
VRST'96 PROCEEDING OF THE ACM OF THE SYMPOSIUM ON VIRITUAL REALITY..., KITAMURA Y.: 'Viritual Object Manipulation Using Dynamically Selected Constraints with Peal-Time Collision Detection' July 1996,, pages 125 - 179, XP002919927 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001312644A (en) * | 2000-04-28 | 2001-11-09 | Sony Corp | Information processor, information processing method and storage medium |
US20020169644A1 (en) * | 2000-05-22 | 2002-11-14 | Greene William S. | Method and system for implementing a management operations center in a global ecosystem of interrelated services |
US8843386B2 (en) | 2000-05-22 | 2014-09-23 | Verizon Patent And Licensing Inc. | Method and system for realizing an avatar in a management operations center implemented in a global ecosystem of interrelated services |
US8856087B2 (en) | 2000-05-22 | 2014-10-07 | Verizon Patent And Licensing Inc. | Method and system for realizing a rendezvous service in a management operations center implemented in a global ecosystem of interrelated services |
KR20020097064A (en) * | 2002-09-18 | 2002-12-31 | 학교법인 한국정보통신학원 | Method for providing a scalable concurrency management in networked virtual environments |
WO2008088666A1 (en) * | 2007-01-16 | 2008-07-24 | Motorola, Inc. | System and method for managing interactions in a virtual environment |
WO2008088665A1 (en) * | 2007-01-16 | 2008-07-24 | Motorola, Inc. | Method and apparatus to facilitate multi-setting virtual reality experiences |
US7696992B2 (en) | 2007-01-16 | 2010-04-13 | Motorola, Inc. | Method and apparatus to facilitate multi-setting virtual reality experiences |
US7973786B2 (en) | 2007-01-16 | 2011-07-05 | Motorola Solutions, Inc. | System and method for managing interaction in a virtual environment |
WO2023147044A1 (en) * | 2022-01-27 | 2023-08-03 | Meta Platforms Technologies, Llc | Scalable parallax system for rendering distant avatars, environments, and dynamic objects |
Also Published As
Publication number | Publication date |
---|---|
DE69904678D1 (en) | 2003-02-06 |
EP0978026B1 (en) | 2003-01-02 |
WO1999042918A3 (en) | 1999-11-11 |
KR20010020126A (en) | 2001-03-15 |
DE69904678T2 (en) | 2003-09-25 |
EP0978026A2 (en) | 2000-02-09 |
KR100705306B1 (en) | 2007-04-11 |
TW421753B (en) | 2001-02-11 |
JP2001522499A (en) | 2001-11-13 |
US6329986B1 (en) | 2001-12-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0978026B1 (en) | Priority-based virtual environment | |
US6396509B1 (en) | Attention-based interaction in a virtual environment | |
KR100992013B1 (en) | Brokering of personalized rulesets for use in digital media character replacement | |
KR101527993B1 (en) | Shared virtual area communication environment based apparatus and methods | |
US20090037822A1 (en) | Context-aware shared content representations | |
US20090276707A1 (en) | Directed communication in a virtual environment | |
JP2001515630A (en) | Virtual environment viewpoint control | |
Kelling et al. | Impact of device, context of use, and content on viewing experience of 360-degree tourism video | |
JPWO2011126134A1 (en) | Real-time video collection / recognition / classification / processing / distribution server system | |
EP2568427A1 (en) | A computer-implemented method for displaying an individual timeline of a user of a social network, computer system and computer readable medium thereof | |
Jain et al. | A taxonomy of sounds in virtual reality | |
JP2023551476A (en) | Graphic interchange format file identification for inclusion in video game content | |
US11651108B1 (en) | Time access control in virtual environment application | |
US8645846B2 (en) | Accessibility in virtual worlds using tags | |
US20090210812A1 (en) | Accessibility in virtual worlds using tags | |
US9024977B2 (en) | Resizing objects in regions of virtual universes | |
Hofma et al. | Liquid workplaces: The potential implications of virtual reality on the workplace | |
JP2009193574A (en) | System and method for optimizing natural language description of object in virtual environment | |
JP6684306B2 (en) | Terminal device, video distribution device, program | |
US20230259249A1 (en) | Systems and methods for displaying and adjusting virtual objects based on interactive and dynamic content | |
US20120240199A1 (en) | Confidential presentations in virtual world infrastructure | |
US11717755B2 (en) | Content generation system and method | |
US20240031186A1 (en) | Architecture to control zones | |
US20240031530A1 (en) | Using zones in a three-dimensional virtual environment for limiting audio and video | |
Sakamoto et al. | Some experiences with developing intelligent Internet-of-Things |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): JP KR |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1999902760 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1019997009673 Country of ref document: KR |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 1999 542306 Kind code of ref document: A Format of ref document f/p: F |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
AK | Designated states |
Kind code of ref document: A3 Designated state(s): JP KR |
|
AL | Designated countries for regional patents |
Kind code of ref document: A3 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
WWP | Wipo information: published in national office |
Ref document number: 1999902760 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1019997009673 Country of ref document: KR |
|
WWG | Wipo information: grant in national office |
Ref document number: 1999902760 Country of ref document: EP |
|
WWG | Wipo information: grant in national office |
Ref document number: 1019997009673 Country of ref document: KR |