US20160030851A1 - Virtual world processing device and method - Google Patents

Virtual world processing device and method Download PDF

Info

Publication number
US20160030851A1
US20160030851A1 US14/825,784 US201514825784A US2016030851A1 US 20160030851 A1 US20160030851 A1 US 20160030851A1 US 201514825784 A US201514825784 A US 201514825784A US 2016030851 A1 US2016030851 A1 US 2016030851A1
Authority
US
United States
Prior art keywords
sensor
type
denotes
unit
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/825,784
Inventor
Seung Ju Han
Jae Joon Han
Won Chul BANG
Do Kyoon Kim
Sang Kyun Kim
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Industry Academy Cooperation Foundation of Myongji University
Original Assignee
Samsung Electronics Co Ltd
Industry Academy Cooperation Foundation of Myongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020090101434A external-priority patent/KR20100138702A/en
Application filed by Samsung Electronics Co Ltd, Industry Academy Cooperation Foundation of Myongji University filed Critical Samsung Electronics Co Ltd
Priority to US14/825,784 priority Critical patent/US20160030851A1/en
Publication of US20160030851A1 publication Critical patent/US20160030851A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/80Special adaptations for executing a specific game genre or game mode
    • A63F13/825Fostering virtual characters
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/211Input arrangements for video game devices characterised by their sensors, purposes or types using inertial sensors, e.g. accelerometers or gyroscopes
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/213Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/215Input arrangements for video game devices characterised by their sensors, purposes or types comprising means for detecting acoustic signals, e.g. using a microphone
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/216Input arrangements for video game devices characterised by their sensors, purposes or types using geographical information, e.g. location of the game device or player using GPS
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/217Input arrangements for video game devices characterised by their sensors, purposes or types using environment-related information, i.e. information generated otherwise than by the player, e.g. ambient temperature or humidity
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/218Input arrangements for video game devices characterised by their sensors, purposes or types using pressure sensors, e.g. generating a signal proportional to the pressure applied by the player
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/219Input arrangements for video game devices characterised by their sensors, purposes or types for aiming at specific areas on the display, e.g. light-guns
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/44Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment involving timing of operations, e.g. performing an action within a time slot
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/60Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
    • A63F13/65Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/50Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers
    • A63F2300/53Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers details of basic data processing
    • A63F2300/534Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers details of basic data processing for network load management, e.g. bandwidth optimization, latency reduction
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/69Involving elements of the real world in the game world, e.g. measurement in live races, real video

Definitions

  • Example embodiments relate to a method and apparatus for processing a virtual world, and more particularly to, a method and apparatus for applying information of a real world to a virtual world.
  • “Project Natal” may provide a user body motion capturing function, a face recognition function, and a voice recognition function by combining Microsoft's Xbox 360 game console with a separate sensor device consisting of a depth/color camera and a microphone array, thereby enabling a user to interact with a virtual world without a dedicated controller.
  • Sony Corporation introduced “Wand” which is an experience-type game motion controller. The “Wand” enables interaction with a virtual world through input of a motion trajectory of a controller by applying, to the PlayStation 3 game console, a location/direction sensing technology obtained by combining a color camera, a marker, and an ultrasonic sensor.
  • the interaction between a real world and a virtual world has two directions. In one direction, data information obtained by a sensor in the real world may be reflected to the virtual world. In the other direction, data information obtained from the virtual world may be reflected to the real world using an actuator.
  • Embodiments suggest a control system, control method, and commanding structure for applying data obtained through a sensor in the real world to the virtual world, to achieve the interaction between the real world and the virtual world.
  • a virtual world processing apparatus including a sensor to encode information collected from a real world into first metadata, an adaptation virtual world to real world (VR) unit to encode information on a virtual world into second metadata, and an adaptation real world to virtual world (RV) unit to generate information to be applied to the virtual world based on the first metadata and the second metadata and to encode the generated information into third metadata.
  • the adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of a binary form, and transmits the binary-encoded second metadata to the adaptation RV unit.
  • the binary-encoded second metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • the adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of extensible markup language (XML) form, and transmits the XML-encoded second metadata to the adaptation RV unit.
  • the adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of an XML form and encoding the XML-encoded data to data of a binary form, and transmits the binary-encoded second metadata to the adaptation RV unit.
  • the senor may generate the first metadata by encoding the information collected from the real world into data of a binary form, and transmits the binary-encoded first metadata to the adaptation RV unit.
  • the binary-encoded first metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • the sensor may generate the first metadata by encoding the information collected from the real world into data of an XML form, and transmit the XML-encoded first metadata to the adaptation RV unit.
  • the sensor may generate the first metadata by encoding the information collected from the real world into data of an XML form and encoding the XML-encoded data into data of a binary form, and transmit the binary-encoded first metadata to the adaptation RV unit.
  • a virtual world processing method including encoding information collected from a real world into first metadata, encoding information on a virtual world into second metadata, generating information to be applied to the virtual world based on the first metadata and the second metadata, and encoding the generated information into third metadata.
  • the encoding of the information collected from the real world into the first metadata may include generating the first metadata by encoding the information collected from the real world into data of a binary form comprising a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • the encoding of the information collected from the real world into the first metadata may include generating the first metadata by encoding the information collected from the real world into data of an XML form and encoding the XML-encoded data into data of a binary form.
  • the encoding of the information on the virtual world into the second metadata may include generating the second metadata by encoding the information on the virtual world into data of a binary form comprising a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • the encoding of the information on the virtual world into the second metadata may include generating the second metadata by encoding the information on the virtual world into data of an XML form and encoding the XML-encoded data to data of a binary form.
  • FIG. 1 is a view illustrating an operation of manipulating an object of a virtual world using a sensor according to example embodiments of the present invention
  • FIG. 2 is a view illustrating a system for manipulating an object of a virtual world using a sensor according to example embodiments
  • FIG. 3 is a view illustrating a virtual world processing apparatus for manipulating an object of a virtual world using a sensor, according to other example embodiments;
  • FIGS. 4 to 6 are views illustrating a sensor and an adaptation real world to virtual world (RV) unit according to example embodiments;
  • FIGS. 7 to 9 are views illustrating an adaptation virtual world to real world (VR) engine and an actuator according to example embodiments
  • FIG. 10 is a view illustrating a structure of a virtual world processing apparatus according to example embodiments.
  • FIG. 11 is a view illustrating a structure of a virtual world processing apparatus according to other embodiments.
  • FIG. 12 is a illustrating a sensor capability base type according to example embodiments.
  • FIG. 13 is a view illustrating syntax of a sensor capability base type according to example embodiments.
  • FIG. 14 is a view illustrating syntax of sensor capability base attributes according to example embodiments.
  • FIG. 15 is a view illustrating a sensor adaptation preference base type according to example embodiments.
  • FIG. 16 is a view illustrating syntax of a sensor adaptation preference base type according to example embodiments.
  • FIG. 17 is a view illustrating syntax of sensor adaptation preference base attributes according to example embodiments.
  • FIG. 18 is a view illustrating a sensed information base type according to example embodiments.
  • FIG. 19 is a flowchart illustrating a virtual world processing method according to example embodiments.
  • FIG. 20 is a view illustrating a flowchart of a virtual world processing method according to other example embodiments.
  • FIG. 21 is a view illustrating an operation of using a virtual world processing apparatus according to example embodiments.
  • object used herein may include an object, a thing, an avatar, and the like, implemented and expressed in a virtual world.
  • FIG. 1 illustrates an operation of manipulating an object 120 of a virtual world, using a sensor 100 according to example embodiments.
  • a user 110 of a real world may manipulate the object 120 of the virtual world using the sensor 100 .
  • the user 110 may input his or her motion, state, intention, shape, and the like through the sensor 100 .
  • the sensor 100 may transmit control information (CI) related to the motion, state, intention, shape, and the like of the user 110 , the CI included in a sensor signal, to a virtual world processing apparatus.
  • CI control information
  • the user 110 of the real world may include humans, animals, plants, inanimate objects such as articles, and even surrounding environment of the user 110 .
  • FIG. 2 illustrates a system for manipulating an object of a virtual world using a sensor according to example embodiments.
  • the signal that includes CI 201 related to the motion, state, intention, shape, and the like of a user of a real world 210 may be transmitted to a virtual world processing apparatus.
  • the CI 201 related to the motion, state, intention, shape, and the like of the user may include a sensor capability, a sensor adaptation preference, and sensed information, which will be described in detail with reference to FIG. 12 through FIG. 21 .
  • the virtual world processing apparatus may include an adaptation real world to virtual world (RV) 220 .
  • the adaptation RV 220 may be implemented by an RV engine.
  • the adaptation RV 220 may convert information of the real world 210 to information applicable to a virtual world 240 , using the CI 201 related to the motion, state, intention, shape, and the like of the user of the real world 210 , the CI 201 included in the sensor signal.
  • the adaptation RV 220 may convert virtual world information (VWI) 202 using the CI 201 related to the motion, state, intention, shape, and the like of the user of the real world 210 .
  • VWI virtual world information
  • the VWI 202 denotes information on the virtual world 240 .
  • the VWI 202 may include information on an object of the virtual world 240 or elements constituting the object.
  • the virtual world processing apparatus may transmit converted information 203 converted by the adaptation RV 220 to the virtual world 240 through adaptation real world to virtual world/virtual world to real world (RV/VR) 230 .
  • Table 1 illustrates structures shown in FIG. 2 .
  • FIG. 3 is a view illustrating a virtual world processing apparatus for manipulating an object of a virtual world using a sensor 250 according to other example embodiments.
  • the virtual world processing apparatus includes the sensor 250 , an adaptation RV unit 255 , and an adaptation VR unit 260 .
  • the sensor 250 may collect information on a motion, state, intention, shape, and the like of a user of a real world.
  • the information collected by the sensor 250 may include sensed information.
  • the senor 250 may include an input unit.
  • the input unit may be input with a sensor adaptation preference from the user of the real world.
  • the sensor 250 may encode the information collected from the real world, to first metadata.
  • the sensor 250 may transmit the first metadata to the adaptation RV unit 255 .
  • the sensor 250 may generate the first metadata by encoding the information collected from the real world into data of a binary form.
  • the binary-encoded first metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes of the binary encoding syntax.
  • the sensor 250 may transmit the binary-encoded first metadata to the adaptation RV unit 255 .
  • the sensor 250 may generate the first metadata by encoding the information collected from the real world into data of an extensible markup language (XML) form.
  • the sensor 250 may transmit the XML-encoded first metadata to the adaptation RV unit 255 .
  • the senor 250 may generate the first metadata by encoding the information collected from the real world into data of the XML form and encoding the XML-encoded data into data of the binary form.
  • the sensor 250 may transmit the binary-encoded first metadata to the adaptation RV unit 255 .
  • the senor 250 may include a metadata encoder 251 configured to encode the collected information into metadata.
  • the metadata encoder 251 may encode the collected information to the first metadata.
  • the sensor 250 may transmit the first metadata to the adaptation RV unit 255 .
  • a metadata decoder 261 included in the adaptation RV unit 255 may decode the first metadata received from the sensor 250 .
  • the metadata encoder 251 may include at least one of an XML encoder that encodes the collected information into XML data and a binary encoder that encodes the collected information into binary data.
  • the metadata decoder 251 may include at least one of an XML decoder that decodes received XML data and a binary decoder that decodes received binary data.
  • FIGS. 4 to 6 are views illustrating a sensor 271 and an adaptation RV unit according to example embodiments.
  • the senor 271 may include an XML encoder 272 .
  • the XML encoder 272 may encode information collected from the real world by the sensor 271 , for example sensed information, into metadata of the XML form.
  • the sensor 271 may transmit the metadata encoded by the XML encoder 272 to an adaptation RV unit 274 .
  • the adaptation RV unit 274 may include an XML decoder 273 .
  • the XML decoder 273 may decode the metadata received from the sensor 271 .
  • a sensor 275 may include an XML encoder 276 and a binary encoder 277 .
  • the XML encoder 276 may encode information collected from the real world by the sensor 275 , for example sensed information, into data of the XML form.
  • the binary encoder 277 may encode the data encoded by the XML encoder 276 into data of the binary form.
  • the sensor 275 may transmit the metadata encoded by the XML encoder 276 and the binary encoder 277 to an adaptation RV 280 .
  • the adaptation RV 280 may include a binary decoder 278 and an XML decoder 279 .
  • the binary decoder 278 may decode the metadata received from the sensor 275 into data of the XML form.
  • the XML decoder 279 may decode again the data decoded into the XML form by the binary decoder 278 .
  • a sensor 281 may include a binary encoder 282 .
  • the binary encoder 282 may encode information collected from the real world by the sensor 281 , for example sensed information, into metadata of the binary form.
  • the sensor 281 may transmit the metadata encoded by the binary encoder 282 to an adaptation RV 284 unit.
  • the adaptation RV 284 unit may include a binary decoder 283 .
  • the binary decoder 283 may decode the metadata received from the sensor 281 .
  • the adaptation VR unit 260 may encode information on the virtual world 265 into second metadata.
  • the adaptation VR unit 260 may generate the second metadata by encoding the information on the virtual world 265 into data of the binary form.
  • the binary-encoded second metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes of the binary encoding syntax.
  • the adaptation VR unit 260 may transmit the binary-encoded second metadata to the adaptation RV unit 255 .
  • the adaptation VR unit 260 may generate the second metadata by encoding information collected from the virtual world 265 into the data of the XML form.
  • the adaptation VR unit 260 may transmit the XML-encoded second metadata to the adaptation RV unit 255 .
  • the adaptation VR unit 260 may generate the second metadata by encoding information collected from the virtual world 265 into the data of the XML form and encoding the XML-encoded data to data of the binary form.
  • the adaptation VR unit 260 may transmit the binary-encoded second metadata to the adaptation RV unit 255 .
  • the adaptation VR unit 260 may include a metadata encoder 262 .
  • the metadata encoder 262 may encode the information on the virtual world 265 into the second metadata.
  • the metadata encoder 262 may include at least one of an XML encoder that encodes the information on the virtual world 265 into metadata of the XML form and a binary encoder that encodes the information on the virtual world 265 into metadata of the binary form.
  • the metadata encoder 262 may include an XML encoder.
  • the XML encoder may encode the information on the virtual world 265 into data of the XML form.
  • the metadata encoder 262 may include an XML encoder and a binary encoder.
  • the XML encoder may encode the information on the virtual world 265 into data of the XML form.
  • the binary encoder may encode the data of the XML form back to data of the binary form.
  • the metadata encoder 262 may include a binary encoder.
  • the binary encoder may encode the information on the virtual world 265 into data of the binary form.
  • the adaptation RV unit 255 may generate information to be applied to the virtual world 265 , based on the first metadata transmitted from the sensor 250 and the second metadata transmitted from the adaptation VR unit 260 .
  • the adaptation RV unit 255 may encode the generated information into third metadata.
  • the adaptation RV unit 255 may include a metadata decoder 256 , a metadata decoder 258 , and a metadata encoder 257 .
  • the adaptation RV unit 255 may generate information to be applied to the virtual world 265 , based on information decoded from the first metadata by the metadata decoder 256 and information decoded from the second metadata by the metadata decoder 258 .
  • the adaptation RV unit 255 may generate the information to be applied to the virtual world 265 , such that the information corresponds to virtual world object characteristics and sensed information included in the second metadata.
  • the metadata encoder 257 may encode the information, which is generated by the adaptation RV unit 255 and to be applied to the virtual world 265 , into the third metadata.
  • the adaptation RV unit 255 may transmit the third metadata to the adaptation VR unit 260 .
  • the adaptation VR unit 260 may include the metadata decoder 261 .
  • the metadata decoder 261 may decode the third metadata.
  • the adaptation VR unit 260 may convert attributes of an object of the virtual world 265 based on the decoded information and, in addition, apply the converted attributes to the virtual world 265 .
  • a virtual world processing system may transmit the information on the virtual world 265 to an actuator of the real world so that the information is reflected to the real world.
  • an actuator of the real world so that the information is reflected to the real world.
  • FIGS. 7 to 9 are views illustrating an adaptation VR engine 285 and an actuator 288 according to example embodiments.
  • the adaptation VR engine 285 may include an XML encoder 286 .
  • the adaptation VR engine 285 may transmit the information on the virtual world 265 to the actuator 288 of a real world so that the information on the virtual world 265 is reflected to the real world.
  • the adaptation VR unit 260 may collect information on a change in attributes of an object of the virtual world 265 , and transmit the collected information to the adaptation VR engine 285 .
  • the adaptation VR 285 may include the XML encoder 286 .
  • the XML encoder 286 may encode the information on the virtual world 265 into data of the XML form.
  • the adaptation VR engine 285 may transmit the data encoded by the XML encoder 286 to the actuator 288 .
  • the actuator 288 may include the XML decoder 287 .
  • the XML decoder 287 may decode the XML data received from the adaptation VR engine 285 .
  • the actuator 288 may operate corresponding to the information decoded by the XML decoder 287 .
  • an adaptation VR engine 290 may include an XML encoder 291 and a binary encoder 292 .
  • the adaptation VR unit 260 may collect information on a change in attributes of the object of the virtual world 265 , and transmit the collected information to the adaptation VR engine 290 .
  • the adaptation VR engine 290 may include the XML encoder 291 and the binary encoder 292 .
  • the XML encoder 291 may encode the information on the virtual world 265 into data of the XML form.
  • the binary encoder 292 may encode the data encoded by the XML encoder 291 back to data of the binary form.
  • the adaptation VR engine 290 may transmit the data encoded by the XML encoder 292 to an actuator 295 .
  • the actuator 295 may include a binary decoder 293 and an XML decoder 294 .
  • the binary decoder 293 may decode the binary data received from the adaptation VR engine 290 into data of the XML form.
  • the XML decoder 294 may decode the data decoded into the XML form by the binary decoder 293 .
  • the actuator 295 may operate corresponding to the information decoded by the XML decoder 294 .
  • an adaptation VR engine 296 may include a binary encoder 297 .
  • the adaptation VR unit 260 may collect information on a change in attributes of the object of the virtual world 265 , and transmit the collected information to the adaptation VR engine 296 .
  • the adaptation VR engine 296 may include a binary encoder 297 .
  • the binary encoder 297 may encode received information on the virtual world into data of the binary form.
  • the adaptation VR engine 296 may transmit the data encoded by the binary encoder 297 to an actuator 299 .
  • the actuator 299 may include a binary decoder 298 .
  • the binary decoder 298 may decode the binary data received from the adaptation VR engine 296 .
  • the actuator 299 may operate corresponding to the information decoded by the binary decoder 298 .
  • FIG. 10 is a view illustrating a structure of a virtual world processing apparatus 300 according to example embodiments.
  • the virtual world processing apparatus 300 includes a storage unit 310 and a processing unit 320 .
  • the storage unit 310 stores sensor capability related to a sensor.
  • the sensor may measure a motion, state, intention, shape, and the like of a user of a real world.
  • the sensor may be expressed as a sensory input device.
  • the sensor may be classified according to sensor types including (1) acoustic, sound, and vibration, (2) automotive and transportation, (3) chemical, (4) electric current, electric potential, magnetic, and radio, (5) environment and weather, (6) flow, (7) ionizing radiation, and subatomic particles, (8) navigation instruments, (9) position, angle, displacement, distance, speed, and acceleration, (10) optical, light, and imaging, (11) pressure, force, density, and level, (12) thermal, heat, and temperature, (13) proximity and presence, and (14) sensor technology.
  • Table 2 illustrates examples sensor according to the sensor types. Therefore, the sensors in Table 2 are suggested only as an embodiment but not limiting.
  • the microphone belonging to a sensor type (1) acoustic, sound, and vibration may collect voice of the user of the real world and ambient sounds of the user.
  • the speed sensor belonging to the sensor type (2) automotive and transportation may measure speed of the user of the real world and speed of an object such as a vehicle of the real world.
  • the oxygen sensor belonging to the sensor type (3) chemical may measure an oxygen ratio in ambient air around the user of the real world and an oxygen ratio in liquid around the user of the real world.
  • the metal detector belonging to the sensor type (4) electric current, electric potential, magnetic, and radio may detect metallic substances present in or around the user of the real world.
  • the rain sensor belonging to the sensor type (5) environment and weather may detect whether it is raining in the real world.
  • the flow sensor belonging to the sensor type (6) flow may measure a ratio of a fluid flow of the real world.
  • the scintillator belonging to the sensor type (7) ionizing radiation and subatomic particles may measure a ratio or radiation present in or around the user of the real world.
  • the variometer belonging to the sensor type (8) navigation instruments may measure a vertical movement speed of or around the user of the real world.
  • the odometer belonging to the sensor type (9) position, angle, displacement, distance, speed, and acceleration may measure a traveling distance of an object of the real world, such as a vehicle.
  • the phototransistor belonging to the sensor type (10) optical, light, and imaging may measure light of the real world.
  • the barometer belonging to the sensor type (11) pressure, force, density, and level may measure an atmospheric pressure of the real world.
  • the bolometer belonging to the sensor type (12) thermal, heat, and temperature may measure radiation rays of the real world.
  • the motion detector belonging to the sensor type (13) proximity and presence may measure a motion of the user of the real world.
  • the biosensor belonging to the sensor type (14) may measure biological characteristics of the user of the real world.
  • FIG. 11 illustrates a structure of a virtual world processing apparatus according to other example embodiments.
  • an input device 360 may be input with a sensor adaptation preference 361 by a user of a real world.
  • the input device 360 may be modularized and inserted in a sensor 370 or a virtual world processing apparatus 350 .
  • the sensor adaptation preference 361 will be described in further detail with reference to FIGS. 15 to 17 .
  • the sensor 370 may transmit a sensor capability 371 and sensed information 372 to the virtual world processing apparatus 350 .
  • the sensor capability 371 and the sensed information 372 will be described in further detail with reference to FIGS. 12 to 14 , and 18 .
  • the virtual world processing apparatus 350 may include a signal processing unit 351 and an adaptation unit 352 .
  • the signal processing unit 351 may receive the sensor capability 371 and the sensed information 372 , and perform signal-processing with respect to the sensor capability 371 and the sensed information 372 . Depending on embodiments, the signal processing unit 351 may filter and validate the sensor capability 371 and the sensed information 372 .
  • the adaptation unit 352 may receive the sensor adaptation preference 361 from the input device 360 . In addition, based on the received sensor adaptation preference 361 , the adaptation unit 352 may perform adaptation with respect to the information signal-processed by the signal processing unit 351 so that the information is applied to a virtual world 380 . In addition, the virtual world processing apparatus 350 may apply the information having undergone the adaptation by the adaptation unit 352 to the virtual world 380 .
  • the sensor capability denotes information on capability of a sensor.
  • a sensor capability base type denotes a base type of the sensor capability.
  • the sensor capability base type may be a base abstract type of the metadata related to a sensor capability commonly applied to all types of sensors, as part of metadata types related to the sensor capability.
  • FIG. 12 illustrates a sensor capability base type 400 according to example embodiments.
  • the sensor capability base 400 may include sensor capability base attributes 410 and any attributes 420 .
  • the sensor capability base attributes 410 denote a group of sensor capabilities basically included in the sensor capability base type 400 .
  • the any attributes 420 denote a group of additional sensor capabilities of a sensor.
  • the any attributes 420 may be unique additional sensor capabilities which are applicable to an arbitrary sensor.
  • the any attributes 420 may allow scalability for inclusion of attributes other than the base attributes.
  • FIG. 13 illustrates syntax 500 of a sensor capability base type according to example embodiments.
  • the syntax 500 of the sensor capability base type may include a diagram 510 , attributes 520 , and a source 530 .
  • the diagram 510 may include a diagram of the sensor capability base type.
  • the attributes 520 may include sensor capability base attributes and any attributes.
  • the source 530 may be a program representing the sensor capability base type using an extensible markup language (XML).
  • XML extensible markup language
  • FIG. 13 is suggested by way of example but not limiting.
  • FIG. 14 illustrates syntax 600 of sensor capability base attributes according to example embodiments.
  • the syntax 600 of the sensor capability base attributes may include a diagram 610 , attributes 620 , and a source 630 .
  • the diagram 610 may include a diagram of the sensor capability base attributes.
  • the attributes 620 may include a unit 601 , a maximum value 602 , a minimum value 603 , an offset 604 , a number of levels 605 , a sensitivity 606 , a signal to noise ratio (SNR) 607 , and an accuracy 608 .
  • SNR signal to noise ratio
  • the unit 601 is a unit of values measured by a sensor. Depending on embodiments, for example, when the sensor is a thermometer, the unit 601 may be Celsius (° C.) and Fahrenheit (° F.). When the sensor is a speed sensor, the unit 601 may be kilometers per hour (km/h) and meters per second (m/s).
  • the maximum value 602 and the minimum value 603 denote a maximum value and a minimum value measurable by the sensor, respectively.
  • the maximum value 602 may be 50° C. and the minimum value 603 may be 0° C.
  • the maximum value 602 and the minimum value 603 may be varied according to use and function of the sensor.
  • the offset 604 denotes an offset value added to a value measured by the sensor to obtain an absolute value.
  • the sensor may determine the offset 604 to a value making the speed zero. For example, when ⁇ 1 km/h is measured as speed of a vehicle of the real world, the offset 604 may be 1 km/h.
  • the number of levels 605 denotes a number of values measurable by the sensor. Depending on embodiments, for example, presuming that the sensor is a thermometer and the maximum value and the minimum value are 50° C. and 0° C., respectively, when the number of levels 605 is 5, the sensor may measure five values, that is, 10° C., 20° C., 30° C., 40° C., and 50° C. Even when temperature of the real world is 27° C., not only when 20° C., the temperature may be measured as 20° C. through round-down. Alternative, in this case, the temperature may be measured as 30° C. through roundup.
  • the sensitivity 606 denotes a minimum input value required for the sensor to measure an output value.
  • the sensor when the sensor is a thermometer and the sensitivity 606 is 1° C., the sensor may not measure a temperature change less than 1° C. but measure only the temperature change of at least 1° C.
  • the SNR 607 denotes a relative degree of a signal measured by the sensor with respect to a noise.
  • the SNR 607 of the sensor may be relatively small.
  • the accuracy 608 denotes an error of the sensor.
  • the accuracy 608 may be a measurement error caused by variation of a propagation speed of a sound according to temperature, humidity, and the like.
  • the accuracy 608 may be determined through a statistical error of the values already measured by the sensor.
  • the attributes 620 may further include a position.
  • the position denotes a position of the sensor.
  • the position of the sensor may be an armpit of the user of the real world.
  • the position may include longitude and latitude, and height and direction from a ground surface.
  • the unit 601 , the maximum value 602 , the minimum value 603 , the offset 604 , the number of levels 605 , the sensitivity 606 , the SNR 607 , the accuracy 608 , and the position, as the sensor capability base attributes, may be rearranged as shown in Table 3.
  • Unit 601 the unit of value maxValue the maximum value that the input device (sensor) can 602 provide. The terms will be different according to the individual device type. minValue the minimum value that the input device (sensor) can 603 provide. The terms will be different according to the individual device type. Offset 604 the number of value locations added to a base value in order to get to a specific absolute value. numOflevels the number of value levels that the device can provide 605 in between maximum and minimum value. Sensitivity the minimum magnitude of input signal required to 606 produce a specified output signal.
  • SNR 607 the ratio of a signal power to the noise power corrupting the signal Accuracy the degree of closeness of a measured quantity to its 608 actual value Position the position of the device from the user’s perspective according to the x-, y-, and z-axis
  • the source 630 may be a program representing the sensor capability base attributes using the XML.
  • a reference numeral 631 of the source 630 defines the maximum value 602 using the XML. According to the reference numeral 631 , the maximum value 602 has “float” type data and is optionally used.
  • a reference numeral 632 of the source 630 defines the minimum value 603 using the XML. According to the reference numeral 632 , the minimum value 603 has “float” type data and is optionally used.
  • a reference numeral 633 of the source 630 defines the number of levels 605 using the XML. According to the reference numeral 633 , the number of levels 605 has “onNegativeInteger” type data and is optionally used.
  • the source 630 shown in FIG. 14 is not limiting but only example embodiments.
  • the processing unit 320 may determine a first value received from the sensor based on the sensor capability, and transmit a second value corresponding to the first value to the virtual world.
  • the processing unit 320 may transmit the second value to the virtual world when the first value received from the sensor is less than or equal to a maximum value measurable by the sensor and greater than or equal to a minimum value measurable by the sensor.
  • the processing unit 320 may consider the first value as the maximum value and transmit the second value to the virtual world. Also, when the first value is less than the minimum value, the processing unit 320 may consider the first value as the minimum value and transmit the second value to the virtual world.
  • the virtual world processing apparatus 300 may further include a second storage unit (not shown) configured to store a sensor adaptation preference for manipulation of the first value received from the sensor.
  • the processing unit 320 may generate a third value from the first value based on the sensor capability, and generate the second value from the third value based on the sensor adaptation preference.
  • information on the motion, state, intention, shape, and the like of the user of the real world, which are measured through the sensor, may be directly reflected to the virtual world.
  • the senor may include a position sensor, an orientation sensor, an acceleration sensor, a light sensor, a sound sensor, a temperature sensor, a humidity sensor, a distance sensor, a motion sensor, an intelligent camera sensor, an ambient noise sensor, an atmospheric sensor, a velocity sensor, an angular velocity sensor, an angular acceleration sensor, a force sensor, a torque sensor, and a pressure sensor.
  • Source 1 denotes a sensor capability related to the position sensor using the XML.
  • a program source shown in Source 1 is only an example embodiment but does not limit the present invention.
  • a position sensor capability type is a tool for describing the sensor capability related to the position sensor.
  • the position sensor capability type may include sensor capability base attributes related to the position sensor.
  • the sensor capability base attributes related to the position sensor may include a range, a range type, an x maximum value, an x minimum value, a y maximum value, a y minimum value, a z maximum value, and a z minimum value.
  • the range denotes a range measurable by the position sensor.
  • the measurable range of the position sensor may be expressed using the range type and a global coordinate system.
  • An origin of the global coordinate may be located at a top left corner.
  • a right handed coordinate system may be applied as the global coordinate.
  • a positive direction of an x-axis may be a direction to a top right corner of a screen
  • a positive direction of a y-axis may be a gravity direction, that is, a bottomward direction of the screen
  • a positive direction of a z-axis may be a direction opposite to the user, that is, a direction into the screen.
  • the range type denotes a range of the global coordinate system according to the x-axis, the y-axis, and the z-axis.
  • the x maximum value denotes a maximum value on the x-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • the x minimum value denotes a minimum value on the x-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • the y maximum value denotes a maximum value on the y-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • the y minimum value denotes a minimum value on the y-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • the z maximum value denotes a maximum value on the z-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • the z minimum value denotes a minimum value on the z-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • Source 2 denotes a sensor capability related to an orientation sensor using the XML.
  • a program source shown in Source 2 is only an example embodiment but does not limit the present invention.
  • An orientation sensor capability type is a tool for describing the sensor capability related to the orientation sensor.
  • the orientation sensor capability type may include sensor capability base attributes related to the orientation sensor.
  • the sensor capability base attributes related to the orientation sensor may include an orientation range, an orientation range type, an x maximum value, an x minimum value, a y maximum value, a y minimum value, a z maximum value, and a z minimum value.
  • the range denotes a range measurable by the orientation sensor.
  • the measurable range of the orientation sensor may be expressed using the orientation range type and the global coordinate system.
  • the orientation range type denotes an orientation range of the global coordinate system according to the x-axis, the y-axis, and the z-axis.
  • the x maximum value denotes a maximum value on the x-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • the x minimum value denotes a minimum value on the x-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • the y maximum value denotes a maximum value on the y-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • the y minimum value denotes a minimum value on the y-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • the z maximum value denotes a maximum value on the z-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • the z minimum value denotes a minimum value on the z-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • Source 3 denotes a sensor capability related to an acceleration sensor using the XML.
  • a program source shown in Source 3 is only an example embodiment but does not limit the present invention.
  • An acceleration sensor capability type is a tool for describing the sensor capability related to the acceleration sensor.
  • the acceleration sensor capability type may include sensor capability base attributes related to the acceleration sensor.
  • the sensor capability base attributes related to the acceleration sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the acceleration sensor using a unit of acceleration, for example, m/s 2 .
  • the minimum value denotes a minimum value measurable by the acceleration sensor using a unit of acceleration, for example, m/s 2 .
  • Source 4 denotes a sensor capability related to a light sensor using the XML.
  • a program source shown in Source 4 is only an example embodiment but does not limit the present invention.
  • a light sensor capability type is a tool for describing the sensor capability related to the light sensor.
  • the light sensor capability type may include sensor capability base attributes related to the light sensor.
  • the sensor capability base attributes related to the light sensor may include a maximum value, a minimum value, a color, and a location.
  • the maximum value denotes a maximum value measurable by the light sensor using a unit of light intensity, for example, LUX.
  • the minimum value denotes a minimum value measurable by the light sensor using a unit of light intensity, for example, LUX.
  • the color denotes a color that may be provided by the light sensor.
  • the color may be an RGB color value.
  • the location denotes a location of the light sensor.
  • the location of the light sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 5 denotes a sensor capability related to a sound sensor using the XML.
  • a program source shown in Source 5 is only an example embodiment but does not limit the present invention.
  • a sound sensor capability type is a tool for describing the sensor capability related to the sound sensor.
  • the sound sensor capability type may include sensor capability base attributes related to the sound sensor.
  • the sensor capability base attributes related to the sound sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the sound sensor using a unit of sound intensity, for example, decibel (dB).
  • the minimum value denotes a minimum value measurable by the sound sensor using a unit of sound intensity, for example, dB.
  • Source 6 denotes a sensor capability related to a temperature sensor using the XML.
  • a program source shown in Source 6 is only an example embodiment but does not limit the present invention.
  • a temperature sensor capability type is a tool for describing the sensor capability related to the temperature sensor.
  • the temperature sensor capability type may include sensor capability base attributes related to the temperature sensor.
  • the sensor capability base attributes related to the temperature sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the temperature sensor using a unit of temperature, for example, ° C. and ° F.
  • the minimum value denotes a minimum value measurable by the temperature sensor using a unit of temperature, for example, ° C. and ° F.
  • the location denotes a location of the temperature sensor.
  • the location of the temperature sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 7 denotes a sensor capability related to a humidity sensor using the XML.
  • a program source shown in Source 7 is only an example embodiment but does not limit the present invention.
  • a humidity sensor capability type is a tool for describing the sensor capability related to the humidity sensor.
  • the humidity sensor capability type may include sensor capability base attributes related to the humidity sensor.
  • the humidity capability base attributes related to the humidity sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the humidity sensor using a unit of humidity, for example, percent (%).
  • the minimum value denotes a minimum value measurable by the humidity sensor using a unit of humidity, for example, %.
  • the location denotes a location of the humidity sensor.
  • the location of the humidity sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 8 denotes a sensor capability related to a distance sensor using the XML.
  • a program source shown in Source 8 is only an example embodiment but does not limit the present invention.
  • a distance sensor capability type is a tool for describing the sensor capability related to the distance sensor.
  • the distance sensor capability type may include sensor capability base attributes related to the distance sensor.
  • the sensor capability base attributes related to the distance sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the distance sensor using a unit of length, for example, meter.
  • the minimum value denotes a minimum value measurable by the distance sensor using a unit of length, for example, meter.
  • the location denotes a location of the distance sensor.
  • the location of the distance sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 9 denotes a sensor capability related to a motion sensor using the XML.
  • a program source shown in Source 9 is only an example embodiment but does not limit the present invention.
  • a motion sensor capability type is a tool for describing the sensor capability related to the motion sensor.
  • the motion sensor may be an integrated sensor of a plurality of sensors.
  • the motion sensor may integrally include a position sensor, a velocity sensor, an acceleration sensor, an orientation sensor, an angular velocity sensor, and an angular acceleration sensor.
  • the motion sensor capability type may include sensor capability base attributes related to the motion sensor.
  • the sensor capability base attributes related to the motion sensor may include a position capability, a velocity capability, an acceleration capability, an orientation capability, an angular velocity capability, and an angular acceleration capability.
  • the position capability denotes capability with respect to the position.
  • the velocity capability denotes capability with respect to the velocity.
  • the acceleration capability denotes capability with respect to the acceleration.
  • the orientation capability denotes capability with respect to the orientation.
  • the angular velocity capability denotes capability with respect to the angular velocity.
  • the angular acceleration capability denotes capability with respect to the angular acceleration.
  • Source 10 denotes a sensor capability related to an intelligent camera sensor using the XML.
  • a program source shown in Source 10 is only an example embodiment but does not limit the present invention.
  • An intelligent camera sensor capability type is a tool for describing the sensor capability related to the intelligent camera sensor.
  • the intelligent camera sensor capability type may include sensor capability base attributes related to the intelligent camera sensor.
  • the sensor capability base attributes related to the intelligent camera sensor may include a feature tracking status, an expression tracking status, a body movement tracking status, a maximum body feature point, a maximum face feature point, a tracked feature, tracked facial feature points, tracked body feature points, a feature type, a facial feature mask, and a body feature mask.
  • the feature tracking status denotes information on whether an intelligent camera is capable of tracking features.
  • the expression tracking status denotes information on whether the intelligent camera is capable of extracting animation related to a facial expression.
  • the body movement tracking status denotes information on whether the intelligent camera is capable of extracting animation related to a body.
  • the maximum body feature point denotes a maximum value of a body feature that can be tracked by the intelligent camera sensor.
  • the maximum face feature point denotes a maximum value of a face feature that can be tracked by the intelligent camera sensor.
  • the tracked feature denotes information on whether tracking of the body feature and the face feature is possible.
  • the tracked facial feature points denote information on whether the respective face features are activated or based on the facial feature mask.
  • the tracked body feature points denote information on whether the respective body features are activated or based on the body feature mask.
  • the feature type denotes a list of feature types.
  • the feature type may include 1. Face, 2. Body, and 3. Face and body.
  • the facial feature mask denotes a list of facial features.
  • the body feature mask denotes a list of body features.
  • Source 11 denotes a sensor capability related to an ambient noise sensor using the XML.
  • a program source shown in Source 11 is only an example embodiment but does not limit the present invention.
  • An ambient noise sensor capability type is a tool for describing the sensor capability related to the ambient noise sensor.
  • the ambient noise sensor capability type may include sensor capability base attributes related to the ambient noise sensor.
  • the sensor capability base attributes related to the ambient noise sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the ambient noise sensor.
  • a unit for the ambient noise sensor may be dB.
  • the minimum value denotes a minimum value measurable by the ambient noise sensor.
  • a unit for the ambient noise sensor may be dB.
  • the location denotes a location of the ambient noise sensor.
  • the location of the ambient noise sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 12 denotes a sensor capability related to an atmospheric pressure sensor using the XML.
  • a program source shown in Source 12 is only an example embodiment but does not limit the present invention.
  • An atmospheric pressure sensor capability type is a tool for describing the sensor capability related to the atmospheric pressure sensor.
  • the atmospheric pressure sensor capability type may include sensor capability base attributes related to the atmospheric pressure sensor.
  • the atmospheric pressure capability base attributes related to the atmospheric pressure sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the atmospheric pressure sensor using a unit of atmospheric pressure, for example, hectopascal (hPa).
  • the minimum value denotes a minimum value measurable by the atmospheric pressure sensor using a unit of atmospheric pressure, for example, hPa.
  • the location denotes a location of the atmospheric pressure sensor.
  • the location of the atmospheric pressure sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 13 denotes a sensor capability related to a velocity sensor using the XML.
  • a program source shown in Source 13 is only an example embodiment but does not limit the present invention.
  • a velocity sensor capability type is a tool for describing the sensor capability related to the velocity sensor.
  • the velocity sensor capability type may include sensor capability base attributes related to the velocity sensor.
  • the velocity capability base attributes related to the velocity sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the velocity sensor using a unit of velocity, for example, m/s.
  • the minimum value denotes a minimum value measurable by the velocity sensor using a unit of velocity, for example, m/s.
  • Source 14 denotes a sensor capability related to an angular velocity sensor using the XML.
  • a program source shown in Source 14 is only an example embodiment but does not limit the present invention.
  • An angular velocity sensor capability type is a tool for describing the sensor capability related to the angular velocity sensor.
  • the angular velocity sensor capability type may include sensor capability base attributes related to the angular velocity sensor.
  • the angular velocity capability base attributes related to the angular velocity sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the angular velocity sensor using a unit of angular velocity, for example, radian/s.
  • the minimum value denotes a minimum value measurable by the angular velocity sensor using a unit of angular velocity, for example, radian/s.
  • Source 15 denotes a sensor capability related to an angular acceleration sensor using the XML.
  • a program source shown in Source 15 is only an example embodiment but does not limit the present invention.
  • An angular acceleration sensor capability type is a tool for describing the sensor capability related to the angular acceleration sensor.
  • the angular acceleration sensor capability type may include sensor capability base attributes related to the angular acceleration sensor.
  • the angular acceleration capability base attributes related to the angular acceleration sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the angular acceleration sensor using a unit of angular acceleration, for example, radian/s 2 .
  • the minimum value denotes a minimum value measurable by the angular acceleration sensor using a unit of angular acceleration, for example, radian/s 2 .
  • Source 16 denotes a sensor capability related to a force sensor using the XML.
  • a program source shown in Source 16 is only an example embodiment but does not limit the present invention.
  • a force sensor capability type is a tool for describing the sensor capability related to the force sensor.
  • the force sensor capability type may include sensor capability base attributes related to the force sensor.
  • the force capability base attributes related to the force sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value measurable by the force sensor using a unit of force, for example, Newton (N).
  • the minimum value denotes a minimum value measurable by the force sensor using a unit of force, for example, N.
  • Source 17 denotes a sensor capability related to a torque sensor using the XML.
  • a program source shown in Source 17 is only an example embodiment but does not limit the present invention.
  • a torque sensor capability type is a tool for describing the sensor capability related to the torque sensor.
  • the torque sensor capability type may include sensor capability base attributes related to the torque sensor.
  • the torque capability base attributes related to the torque sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the torque sensor using a unit of torque, for example, Newton millimeter (N-mm).
  • the minimum value denotes a minimum value measurable by the torque sensor using a unit of torque, for example, N-mm.
  • Source 18 denotes a sensor capability related to a pressure sensor using the XML.
  • a program source shown in Source 18 is only an example embodiment but does not limit the present invention.
  • a pressure sensor capability type is a tool for describing the sensor capability related to the pressure sensor.
  • the pressure sensor capability type may include sensor capability base attributes related to the pressure sensor.
  • the pressure capability base attributes related to the pressure sensor may include a maximum value, a minimum value, and a location.
  • the maximum value denotes a maximum value measurable by the pressure sensor using a unit of pressure, for example, m/s.
  • the minimum value denotes a minimum value measurable by the pressure sensor using a unit of pressure, for example, m/s.
  • the sensor adaptation preference denotes information for manipulating a value received from a sensor.
  • a sensor adaptation preference base type denotes a base type of manipulation information of the user.
  • the sensor adaptation preference base type may be a base abstract type of the metadata related to a sensor adaptation preference commonly applied to all types of sensors, as part of metadata types related to the sensor adaptation preference.
  • FIG. 15 illustrates a sensor adaptation preference base type 700 according to example embodiments.
  • the sensor adaptation preference base type 700 may include sensor adaptation preference base attributes 710 and any attributes 720 .
  • the sensor adaptation preference base attributes 710 denote a group of sensor adaptation preferences basically included in the sensor adaptation preference base type 700 .
  • the any attributes 720 denote a group of additional sensor adaptation preferences.
  • the any attributes 720 may be unique additional sensor capabilities which are applicable to an arbitrary sensor.
  • the any attributes 420 may allow scalability for inclusion of attributes other than the base attributes.
  • FIG. 16 illustrates syntax 800 of a sensor adaptation preference base type according to example embodiments.
  • the syntax of the sensor adaptation preference base type may include a diagram 810 , attributes 820 , and a source 830 .
  • the diagram 810 may include a diagram of the sensor adaptation preference base type.
  • the attributes 820 may include sensor adaptation preference base attributes and any attributes.
  • the source 830 may be a program representing the sensor adaptation preference base type using an XML. However, the source 830 shown in FIG. 16 is suggested by way of example but not limiting.
  • FIG. 17 illustrates syntax 900 of sensor adaptation preference base attributes according to example embodiments.
  • the syntax 900 of the sensor adaptation preference base attributes may include a diagram 910 , attributes 920 , and a source 930 .
  • the diagram 910 may include a diagram of the sensor adaptation preference base attributes.
  • the attributes 920 may include a sensor reference identifier (ID) 901 , a sensor adaptation mode 902 , an activation state 903 , a unit 904 , a maximum value 905 , a minimum value 906 , and a number of levels 907 .
  • ID sensor reference identifier
  • the sensor reference ID 901 denotes information referencing an ID of an individual sensor that generates specific sensed information.
  • the sensor adaptation mode 902 denotes user preference information related to a method of adapting a sensor.
  • the sensor adaptation mode 902 may be a sensor adaptation preference related to an adaptation method that refines information on a motion, state, intention, shape, and the like of a user of a real world, measured through the sensor, and reflects the information to a virtual world.
  • a ‘strict’ value may denote a user preference that directly applies sensed information of the real world to the virtual world.
  • a ‘scalable’ value may denote a user preference that varies the sensed information of the real world according to the user preference and applies the sensed information to the virtual world.
  • the activation state information 903 denotes information on whether to activate the sensor in the virtual world. Depending on embodiments, the activation state information 903 may be a sensor adaptation preference that determines whether the sensor is in operation.
  • the unit 904 denotes a unit of a value used in the virtual world.
  • the unit 904 may be a pixel.
  • the unit 904 may be a unit of a value corresponding to the value received from the sensor.
  • the maximum value 905 and the minimum value 906 denote a maximum value and a minimum value used in the virtual world, respectively. Depending on embodiments, the maximum value 905 and the minimum value 906 may be the unit of the value corresponding to the value received from the sensor.
  • the number of levels 907 denotes a number of levels used in the virtual world. Depending on embodiments, the number of levels 907 may be a value for dividing levels between the maximum value and the minimum used in the virtual world.
  • the sensor reference ID 901 , the adaptation mode 902 , the activation state 903 , the unit 904 , the maximum value 905 , the minimum value 906 , and the number of levels 907 , as the sensor adaptation preference base attributes, may be rearranged as shown in Table 4.
  • SensorIdRef Refers the Id of an individual sensor that has generated 901 the specific sensed information Sensor the user’s preference on the adaptation method for the adaptation virtual world effect mode 902 Activate whether the effect shall be activated. a value of true 903 means the effect shall be activated and false means the effect shall be deactivated Unit 904 the unit of value maxValue the maximum desirable value of the effect in percentage 905 according to the max scale defined within the semantics definition of the individual effects minValue the minimum desirable value of the effect in percentage 906 according to the min scale defined within the semantics definition of the individual effects numOflevels the number of value levels that the device can provide in 907 between maximum and minimum value
  • the source 930 may be a program representing the sensor adaptation preference base attributes using the XML.
  • a reference numeral 931 defines the activation state 903 using the XML. According to the reference numeral 931 , the activation state 903 has “boolean” type data and is optionally used.
  • a reference numeral 932 defines the maximum value 905 using the XML. According to the reference numeral 932 , the maximum value 905 has “float” type data and is optionally used.
  • a reference number 933 defines minimum value 906 using the XML. According to the reference numeral 933 , the minimum value 906 has “float” type data and is optionally used.
  • a reference numeral 934 defines the number of levels 907 using the XML. According to the reference numeral 934 , the number of levels 907 has “onNegativeInteger” type data and is optionally used.
  • the source 930 illustrated in FIG. 17 is not limiting but only example embodiments.
  • Source 19 denotes a sensor adaptation preference related to a position sensor using the XML.
  • a program source shown in Source 19 is only an example embodiment but does not limit the present invention.
  • a position sensor type is a tool for describing the sensor adaptation preference related to the position sensor.
  • a position sensor capability type may include sensor adaptation preference base attributes related to the position sensor.
  • the sensor adaptation preference base attributes related to the position sensor may include a range and a number of levels.
  • the range denotes a range of a user preference with respect to position information measured by the position sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the position information measured by the position sensor.
  • Source 20 denotes a sensor adaptation preference related to an orientation sensor using the XML.
  • a program source shown in Source 20 is only an example embodiment but does not limit the present invention.
  • An orientation sensor type is a tool for describing the sensor adaptation preference related to the orientation sensor.
  • An orientation sensor capability type may include sensor adaptation preference base attributes related to the orientation sensor.
  • the sensor adaptation preference base attributes related to the orientation sensor may include an orientation range and a number of levels.
  • the orientation range denotes a range of a user preference with respect to orientation information measured by the orientation sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the orientation information measured by the orientation sensor.
  • Source 21 denotes a sensor adaptation preference related to an acceleration sensor using the XML.
  • a program source shown in Source 21 is only an example embodiment but does not limit the present invention.
  • An acceleration sensor type is a tool for describing the sensor adaptation preference related to the acceleration sensor.
  • An acceleration sensor capability type may include sensor adaptation preference base attributes related to the acceleration sensor.
  • the sensor adaptation preference base attributes related to the acceleration sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference related to acceleration information measured by the acceleration sensor.
  • the minimum value denotes a minimum value of the user preference related to the acceleration information measured by the acceleration sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the acceleration information measured by the acceleration sensor.
  • Source 22 denotes a sensor adaptation preference related to a light sensor using the XML.
  • a program source shown in Source 22 is only an example embodiment but does not limit the present invention.
  • a light sensor type is a tool for describing the sensor adaptation preference related to the light sensor.
  • a light sensor capability type may include sensor adaptation preference base attributes related to the light sensor.
  • the sensor adaptation preference base attributes related to the light sensor may include a maximum value, a minimum value, a number of levels, and an unfavorable color.
  • the maximum value denotes a maximum value of a user preference related to a value measured by the light sensor.
  • the minimum value denotes a minimum value of the user preference related to a value measured by the light sensor.
  • the number of levels denotes a number of levels of the user preference with respect to a value measured by the light sensor.
  • the unfavorable color denotes a list of unfavorable colors of the user, as RGB color values or a classification reference.
  • Source 23 denotes a sensor adaptation preference related to a sound sensor using the XML.
  • a program source shown in Source 23 is only an example embodiment but does not limit the present invention.
  • a sound sensor type is a tool for describing the sensor adaptation preference related to the sound sensor.
  • a sound sensor capability type may include sensor adaptation preference base attributes related to the sound sensor.
  • the sensor adaptation preference base attributes related to the sound sensor may include a maximum value and a minimum value.
  • the maximum value denotes a maximum value allowed by the user as a measured value of the sound sensor.
  • the minimum value denotes a minimum value allowed by the user as a measured value of the sound sensor.
  • Source 24 denotes a sensor adaptation preference related to a temperature sensor using the XML.
  • a program source shown in Source 24 is only an example embodiment but does not limit the present invention.
  • a temperature sensor type is a tool for describing the sensor adaptation preference related to the temperature sensor.
  • a temperature sensor capability type may include sensor adaptation preference base attributes related to the temperature sensor.
  • the sensor adaptation preference base attributes related to the temperature sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference related to temperature information measured by the temperature sensor.
  • the minimum value denotes a minimum value of the user preference related to the temperature information measured by the temperature sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the temperature information measured by the temperature sensor.
  • Source 25 denotes a sensor adaptation preference related to a humidity sensor using the XML.
  • a program source shown in Source 25 is only an example embodiment but does not limit the present invention.
  • a humidity sensor type is a tool for describing the sensor adaptation preference related to the humidity sensor.
  • a humidity sensor capability type may include sensor adaptation preference base attributes related to the humidity sensor.
  • the sensor adaptation preference base attributes related to the humidity sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference related to humidity information measured by the humidity sensor.
  • the minimum value denotes a minimum value of the user preference related to the humidity information measured by the humidity sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the humidity information measured by the humidity sensor.
  • Source 26 denotes a sensor adaptation preference related to a distance sensor using the XML.
  • a program source shown in Source 26 is only an example embodiment but does not limit the present invention.
  • a distance sensor type is a tool for describing the sensor adaptation preference related to the distance sensor.
  • a distance sensor capability type may include sensor adaptation preference base attributes related to the distance sensor.
  • the sensor adaptation preference base attributes related to the distance sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference related to length information measured by the distance sensor.
  • the minimum value denotes a minimum value of the user preference related to the length information measured by the distance sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the length information measured by the distance sensor.
  • Source 27 denotes a sensor adaptation preference related to a motion sensor using the XML.
  • a program source shown in Source 27 is only an example embodiment but does not limit the present invention.
  • a motion sensor capability type is a tool for describing the sensor adaptation preference related to the motion sensor.
  • the motion sensor capability type may include sensor adaptation preference base attributes related to the motion sensor.
  • the sensor adaptation preference base attributes related to the motion sensor may include a position preference, a velocity preference, an acceleration preference, an orientation preference, an angular velocity preference, and an angular acceleration preference.
  • the position preference denotes a user preference with respect to the position.
  • the velocity preference denotes a user preference with respect to the velocity.
  • the acceleration preference denotes a user preference with respect to the acceleration.
  • the orientation preference denotes a user preference with respect to the orientation.
  • the angular velocity preference denotes a user preference with respect to the angular velocity.
  • the angular acceleration preference denotes a user preference with respect to the angular acceleration.
  • Source 28 denotes a sensor adaptation preference related to an intelligent camera sensor using the XML.
  • a program source shown in Source 28 is only an example embodiment but does not limit the present invention.
  • An intelligent camera sensor capability type is a tool for describing the sensor adaptation preference related to the intelligent camera sensor.
  • the intelligent camera sensor capability type may include sensor adaptation preference base attributes related to the intelligent camera sensor.
  • the sensor adaptation preference base attributes related to the intelligent camera sensor may include a face feature tracking on, a body feature tracking on, a facial expression tracking on, a gesture tracking on, a face tracking map, and a body tracking map.
  • the face feature tracking denotes information on whether to activate a face feature tracking mode in which an intelligent camera sensor tracks features on a face of the user.
  • the body feature tracking denotes information on whether to activate a body feature tracking mode in which the intelligent camera sensor tracks features on a body of the user.
  • the facial expression tracking on denotes information on user preference with respect to tracking of a facial expression of the user by the intelligent camera sensor.
  • the gesture tracking on denotes information on user preference with respect to tracking of a gesture of the user by the intelligent camera sensor.
  • the face tracking map provides a Boolean map related to a face tracking map type.
  • the Boolean map provides face portions that the user wants to track.
  • the Boolean map according to the face tracking map type may provide eyes, a mouth, a noise, and ears as the face portions.
  • the body tracking map provides a Boolean map related to a body tracking map type.
  • the Boolean map provides body portions that the user wants to track.
  • the Boolean map according to the body tracking map type may provide a head, arms, hands, legs, feet, and a middle body as the body portions.
  • Source 29 denotes a sensor adaptation preference related to an ambient noise sensor using the XML.
  • a program source shown in Source 29 is only an example embodiment but does not limit the present invention.
  • An ambient noise sensor type is a tool for describing the sensor adaptation preference related to the ambient noise sensor.
  • An ambient noise sensor capability type may include sensor adaptation preference base attributes related to the ambient noise sensor.
  • the sensor adaptation preference base attributes related to the ambient noise sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to ambient noise information measured by the ambient noise sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the ambient noise information measured by the ambient noise sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the ambient noise information measured by the ambient noise sensor.
  • Source 30 denotes a sensor adaptation preference related to an atmospheric pressure sensor using the XML.
  • a program source shown in Source 30 is only an example embodiment but does not limit the present invention.
  • An atmospheric pressure sensor type is a tool for describing the sensor adaptation preference related to the atmospheric pressure sensor.
  • An atmospheric pressure sensor capability type may include sensor adaptation preference base attributes related to the atmospheric pressure sensor.
  • the sensor adaptation preference base attributes related to the atmospheric pressure sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to atmospheric pressure information measured by the atmospheric pressure sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the atmospheric pressure information measured by the atmospheric pressure sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the atmospheric pressure information measured by the atmospheric pressure sensor.
  • Source 31 denotes a sensor adaptation preference related to a velocity sensor using the XML.
  • a program source shown in Source 31 is only an example embodiment but does not limit the present invention.
  • a velocity sensor type is a tool for describing the sensor adaptation preference related to the velocity sensor.
  • a velocity sensor capability type may include sensor adaptation preference base attributes related to the velocity sensor.
  • the sensor adaptation preference base attributes related to the velocity sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to velocity information measured by the velocity sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the velocity information measured by the velocity sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the velocity information measured by the velocity sensor.
  • Source 32 denotes a sensor adaptation preference related to an angular velocity sensor using the XML.
  • a program source shown in Source 32 is only an example embodiment but does not limit the present invention.
  • An angular velocity sensor type is a tool for describing the sensor adaptation preference related to the angular velocity sensor.
  • An angular velocity sensor capability type may include sensor adaptation preference base attributes related to the angular velocity sensor.
  • the sensor adaptation preference base attributes related to the angular velocity sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to angular velocity information measured by the angular velocity sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the angular velocity information measured by the angular velocity sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the angular velocity information measured by the angular velocity sensor.
  • Source 33 denotes a sensor adaptation preference related to an angular acceleration sensor using the XML.
  • a program source shown in Source 33 is only an example embodiment but does not limit the present invention.
  • An angular acceleration sensor type is a tool for describing the sensor adaptation preference related to the angular acceleration sensor.
  • An angular acceleration sensor capability type may include sensor adaptation preference base attributes related to the angular acceleration sensor.
  • the sensor adaptation preference base attributes related to the angular acceleration sensor may include a maximum value, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to angular acceleration information measured by the angular acceleration sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the angular acceleration information measured by the angular acceleration sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the angular acceleration information measured by the angular acceleration sensor.
  • Source 34 denotes a sensor adaptation preference related to a force sensor using the XML.
  • a program source shown in Source 34 is only an example embodiment but does not limit the present invention.
  • a force sensor type is a tool for describing the sensor adaptation preference related to the force sensor.
  • a force sensor capability type may include sensor adaptation preference base attributes related to the force sensor.
  • the sensor adaptation preference base attributes related to the force sensor may include a maximum value and, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to force information measured by the force sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the force information measured by the force sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the force information measured by the force sensor.
  • Source 35 denotes a sensor adaptation preference related to a torque sensor using the XML.
  • a program source shown in Source 35 is only an example embodiment but does not limit the present invention.
  • a torque sensor type is a tool for describing the sensor adaptation preference related to the torque sensor.
  • a torque sensor capability type may include sensor adaptation preference base attributes related to the torque sensor.
  • the sensor adaptation preference base attributes related to the torque sensor may include a maximum value and, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to torque information measured by the torque sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the torque information measured by the torque sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the torque information measured by the torque sensor.
  • Source 36 denotes a sensor adaptation preference related to a pressure sensor using the XML.
  • a program source shown in Source 36 is only an example embodiment but does not limit the present invention.
  • a pressure sensor type is a tool for describing the sensor adaptation preference related to the pressure sensor.
  • a pressure sensor capability type may include sensor adaptation preference base attributes related to the pressure sensor.
  • the sensor adaptation preference base attributes related to the pressure sensor may include a maximum value and, a minimum value, and a number of levels.
  • the maximum value denotes a maximum value of a user preference with respect to pressure information measured by the pressure sensor.
  • the minimum value denotes a minimum value of the user preference with respect to the pressure information measured by the pressure sensor.
  • the number of levels denotes a number of levels of the user preference with respect to the pressure information measured by the pressure sensor.
  • the virtual world processing apparatus may include sensed information.
  • the sensed information denotes a command for controlling the sensor.
  • the sensed information may be a command for controlling the sensor so as to reflect the information on the motion, state, intention, shape, and the like of the user of the real world, measured through the sensor, to the virtual world.
  • the sensed information may serve as a root element for sensed information metadata.
  • FIG. 18 illustrates a sensed information base type 1000 according to example embodiments.
  • the sensed information base type 1000 may include sensed information base attributes 1010 and any attributes 1020 .
  • the sensed information base type 1000 may be a topmost type of a base type that may inherit an individual piece of the sensed information.
  • the sensed information base attributes 1010 denote a group of attributes for the commands.
  • the any attributes 1020 denote a group of additional sensed information.
  • the any attributes 1020 may be unique additional sensed information applicable to an arbitrary sensor.
  • the any attributes 1020 may allow scalability for inclusion of attributes other than the base attributes.
  • Source 37 may be a program denoting a sensed information base type using the XML. However, Source 37 is only an example embodiment but does not limit the present invention.
  • the sensed information base type 101 may include an ID 1011 , a sensor reference ID 1012 , a group ID 1013 , a priority 1014 , an activation state, and a linked list 1016 .
  • the ID 1011 denotes ID information for recognizing individual identity of the sensor.
  • the sensor reference ID 1012 may be information referencing an ID of the sensor that generates information included in particular sensed information.
  • the group ID 1013 denotes ID information for recognizing individual identity of a multi-sensor structure to which a particular sensor belongs.
  • the priority 1014 denotes information on a priority related to sensed information of sensors in the same group, the sensors sharing the same point at time when the sensed information is applied.
  • a value 1 may represent a highest priority and larger values may indicate lower priorities.
  • the activation state 1015 denotes information for determining whether the sensor is in operation.
  • the linked list 1016 denotes information on link data for grouping a plurality of sensors.
  • the linked list 1016 may be information on a multi-sensor structure for grouping the sensors by a method including reference information on IDs of neighboring sensors.
  • the sensed information base attributes 1010 may further include a value, a timestamp, and a life span.
  • the value denotes a measured value of the sensor.
  • the value may be received from the sensor.
  • the timestamp denotes information on a sensing time of the sensor.
  • the life span denotes information on a valid period of a sensor command.
  • the life span may be a second unit.
  • the sensed information base attributes including the ID, the sensor ID reference, the activation state, the value, the linked list, the timestamp, and the lifespan may be rearranged as shown in Table 5.
  • Linked grouping sensor structure that consists of a group of list sensors such that in each record there is a field that contains a reference (id) to the next sensor.
  • Source 37-2 may be a program representing the sensed information base attributes using the XML.
  • Source 38 denotes sensed information related to a position sensor using the XML.
  • a program source shown in Source 38 is only an example embodiment but does not limit the present invention.
  • a position sensor type is a tool for describing sensed information related to the position sensor.
  • the position sensor type may include attributes such as a timestamp, a life span, a position, a position value type, Px, Py, and Pz.
  • the timestamp denotes information on a sensing time of the position sensor.
  • the life span denotes information on a valid period of a command of the position sensor.
  • the life span may be a second unit.
  • the position denotes information on a 3-dimensional (3D) value of the position sensor, expressed by a unit of distance, for example, meter.
  • the position value type denotes a tool for indicating a 3D position vector.
  • the Px denotes information on an x-axis value of the position sensor.
  • the Py denotes information on a y-axis value of the position sensor.
  • the Pz denotes information on a z-axis value of the position sensor.
  • Source 38-2 and Source 38-3 denote binary representation syntax corresponding to the sensed information related to the position sensor according to the example embodiments.
  • Source 38-2 and Source 38-3 may include a binary encoding syntax the sensed information of the position sensor, the number of bits occupied by attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • the binary encoding syntax corresponding to the sensed information of the position sensor may include a priority denoting a priority of the position sensor.
  • the number of bits occupied by the priority may be 8 bits.
  • a priority mnemonic may be uimsbf which stands for ‘unsigned integer most significant bit first.’
  • the mnemonic may further include bslbf which stands for ‘bit string left bit first’ and Unicode transformation format-8 (UTF-8) which stands for a variable length letter encoding method for Unicode.
  • UTF-8 may use 1 to 4 bytes to express one Unicode letter.
  • Table 1-0 shows a mnemonic of data.
  • Bit string left bit first, where “left” is the order in which bits are written in ISO/IEC 15938-3.
  • Bit strings are generally written as a string of 1s and 0s within single quote marks, e.g. ‘1000 0001’. Blanks within a bit string are for ease of reading and have no significance. For convenience, large strings are occasionally written in hexadecimal, in which case conversion to a binary in the conventional manner will yield the value of the bit string. Thus, the left-most hexadecimal digit is first and in each hexadecimal digit the most significant of the four digits is first.
  • UTF-8 Binary string encoding defined in ISO 10646/IETF RFC 2279.
  • vluimsbf5 Variable length unsigned integer most significant bit first representation consisting of two parts.
  • the first part defines the number n of 4-bit bit fields used for the value representation, encoded by a sequence of n-1 “1” bits, followed by a “0” bit signaling its end.
  • the second part contains the value of the integer encoded using the number of bit fields specified in the first part.
  • uimsbf Unsigned integer, most significant bit first.
  • fsbf Float 32 bit
  • sign bit first The semantics of the bits within a float are specified in the IEEE Standard for Binary Floating Point Arithmetic (ANSI/IEEE Std 754-1985).
  • Table 1-1 shows descriptor components semantics of the position sensor according to example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • positionFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a position sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • position Describes the sensed value of the position sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 39 denotes sensed information related to an orientation sensor using the XML.
  • a program source shown in Source 39 is only an example embodiment but does not limit the present invention.
  • An orientation sensor type is a tool for describing sensed information related to the orientation sensor.
  • the orientation sensor type may include attributes such as a timestamp, a life span, an orientation, an orientation value type, Ox, Oy, and Oz.
  • the timestamp denotes information on a sensing time of the orientation sensor.
  • the life span denotes information on a valid period of a command of the orientation sensor.
  • the life span may be a second unit.
  • the orientation denotes information on a 3D value of the orientation sensor, expressed by a unit of orientation, for example, radian.
  • the orientation value type denotes a tool for indicating a 3D orientation vector.
  • the Ox denotes information on a value of an x-axis rotation angle of the orientation sensor.
  • the Oy denotes information on a value of a y-axis rotation angle of the orientation sensor.
  • the Oz denotes information on a value of a z-axis rotation angle of the orientation sensor.
  • Source 39-2 denotes a binary representation syntax corresponding to the sensed information related to the orientation sensor according the example embodiments.
  • Table 1-2 shows descriptor components semantics of the orientation sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • orientationFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to an orientation sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • orientation Describes the sensed value of the orientation sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 40 denotes sensed information related to an acceleration sensor using the XML.
  • a program source shown in Source 40 is only an example embodiment but does not limit the present invention.
  • An acceleration sensor type is a tool for describing sensed information related to the acceleration sensor.
  • the acceleration sensor type may include attributes such as a timestamp, a life span, an acceleration, an acceleration value type, Ax, Ay, and Az.
  • the timestamp denotes information on a sensing time of the acceleration sensor.
  • the life span denotes information on a valid period of a command of the acceleration sensor.
  • the life span may be a second unit.
  • the acceleration denotes information on a value of the acceleration sensor, expressed by a unit of acceleration, for example, m/s 2 .
  • the acceleration value type denotes a tool for indicating a 3D acceleration vector.
  • the Ax denotes information on an x-axis value of the acceleration sensor.
  • the Ay denotes information on a y-axis value of the acceleration sensor.
  • the Az denotes information on a z-axis value of the acceleration sensor.
  • Source 40-2 denotes a binary representation syntax corresponding to the sensed information related to the acceleration sensor according the example embodiments.
  • Table 1-3 shows descriptor components semantics of the acceleration sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • accelerationFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user- defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to an acceleration sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • acceleration Describes the sensed value of the acceleration sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 41 denotes sensed information related to a light sensor using the XML.
  • a program source shown in Source 41 is only an example embodiment but does not limit the present invention.
  • a light sensor type is a tool for describing sensed information related to the light sensor.
  • the light sensor type may include attributes such as a timestamp, a life span, a value, and a color.
  • the timestamp denotes information on a sensing time of the light sensor.
  • the life span denotes information on a valid period of a command of the light sensor.
  • the life span may be a second unit.
  • the value denotes information on a light sensor value expressed by a unit of light intensity, for example, LUX.
  • the color denotes a color that may be provided by the light sensor.
  • the color may be an RGB color value.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the light sensor according to the example embodiments.
  • Table 1-1-1 shows binary encoding of a unit CS corresponding to the light sensor according to the example embodiments.
  • Table 1-1-1 may include a unit type and a term ID of unit.
  • Table 1-1-1 shows binary encoding of a unit CS corresponding to the light sensor according to the example embodiments.
  • Table 1-1-1 may include a unit type and a term ID of unit.
  • Table 1-1 shows descriptor components semantics of the light sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • colorFlag This field, which is only present in the binary representation, signals the presence of color attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used. “00” means that the absolute time stamp type shall be used, “01” means that the clock tick time stamp type shall be used, and “10” means that the clock tick time delta stamp type shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6.
  • ID ID to identify the sensed information with respect to a light sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated. value Describes the sensed value of the light sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • Unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • color Describes the list of colors which the lighting device can sense as a reference to a classification scheme term or as RGB value.
  • a CS that may be used for this purpose is the ColorCS defined in A.2.3 of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 42 denotes sensed information related to a sound sensor using the XML.
  • a program source shown in Source 42 is only an example embodiment but does not limit the present invention.
  • a sound sensor command type is a tool for describing sensed information related to the sound sensor.
  • Source 43 denotes sensed information related to a temperature sensor using the XML.
  • a program source shown in Source 43 is only an example embodiment but does not limit the present invention.
  • a temperature sensor type is a tool for describing sensed information related to the temperature sensor.
  • the temperature sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the temperature sensor.
  • the life span denotes information on a valid period of a command of the temperature sensor.
  • the life span may be a second unit.
  • the value denotes information on a temperature sensor value expressed by a unit of temperature, for example, ° C. and ° F.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the temperature sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the temperature sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a temperature sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated. value Describes the sensed value of the temperature sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 44 denotes sensed information related to a humidity sensor using the XML.
  • a program source shown in Source 44 is only an example embodiment but does not limit the present invention.
  • a humidity sensor type is a tool for describing sensed information related to the humidity sensor.
  • the humidity sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the humidity sensor.
  • the life span denotes information on a valid period of a command of the humidity sensor.
  • the life span may be a second unit.
  • the value denotes information on a humidity sensor value expressed by a unit of humidity, for example, %.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the humidity sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the humidity sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a humidity sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • value Describes the sensed value of the humidity sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 45 denotes sensed information related to a distance sensor using the XML.
  • a program source shown in Source 45 is only an example embodiment but does not limit the present invention.
  • a distance sensor type is a tool for describing sensed information related to the distance sensor.
  • the distance sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the distance sensor.
  • the life span denotes information on a valid period of a command of the distance sensor.
  • the life span may be a second unit.
  • the value denotes information on a distance sensor value expressed by a unit of distance, for example, meter.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the distance sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the distance sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a distance sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated. value Describes the sensed value of the distance sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 46 denotes sensed information related to a motion sensor using the XML.
  • a program source shown in Source 46 is only an example embodiment but does not limit the present invention.
  • a motion sensor type is a tool for describing sensed information related to the length sensor.
  • the motion sensor type may include attributes such as an ID, an ID reference, a position, a velocity, an acceleration, an orientation, an angular velocity, and an angular acceleration.
  • the ID denotes ID information for recognizing individual identity of the motion sensor.
  • the ID reference denotes additional information related to the ID, the additional information for recognizing individual identity of the motion sensor.
  • the position denotes information on a position vector value of a unit of position, for example, meter.
  • the velocity denotes information on a velocity vector value of a unit of velocity, for example, m/s.
  • the acceleration denotes information on an acceleration vector value of a unit of acceleration, for example, m/s 2 .
  • the orientation denotes information on an orientation vector value of a unit of orientation, for example, radian.
  • the angular velocity denotes information on an angular velocity vector value of a unit of angular velocity, for example, radian/s.
  • the angular acceleration denotes information on an angular acceleration vector value of a unit of angular acceleration, for example, radian/s 2 .
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the motion sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the motion sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • positionFlag This field, which is only present in the binary representation, signals the presence of position value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • orientationFlag This field, which is only present in the binary representation, signals the presence of orientation value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • velocityFlag This field, which is only present in the binary representation, signals the presence of velocity value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • angularvelocityFlag This field, which is only present in the binary representation, signals the presence of angular velocity value attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • accelerationFlag This field, which is only present in the binary representation, signals the presence of acceleration value attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • angularaccelerationFlag This field, which is only present in the binary representation, signals the presence of angular acceleration value attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used. “00” means that the absolute time stamp type shall be used, “01” means that the clock tick time stamp type shall be used, and “10” means that the clock tick time delta stamp type shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6.
  • ID ID to identify the sensed information with respect to a motion sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted. A value of zero indicates the highest priority and larger values indicate lower priorities. The default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself. activate Describes whether the sensor is activated.
  • a value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • position Describes the sensed position value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • orientation Describes the sensed orientation value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • velocity Describes the sensed velocity value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • angularvelocity Describes the sensed angular velocity value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • acceleration Describes the sensed acceleration value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • angularacceleration Describes the sensed angular acceleration value of the motion sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • Source 47 denotes sensed information related to an intelligent camera sensor using the XML.
  • a program source shown in Source 47 is only an example embodiment but does not limit the present invention.
  • An intelligent camera sensor type is a tool for describing sensed information related to the intelligent camera sensor.
  • the intelligent camera sensor type may include a facial animation ID, a body animation ID, a face feature, and a body feature.
  • the facial animation ID denotes an ID referencing an animation clip with respect to a facial expression.
  • the body animation ID denotes an ID referencing an animation clip with respect to a body.
  • the face feature denotes information on a 3D position of each face feature sensed by the intelligent camera sensor.
  • the body feature denotes information on a 3D position of each body feature sensed by the intelligent camera sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the intelligent camera sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the intelligent camera sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • FacialIDFlag This field, which is only present in the binary representation, signals the presence of the facial animation ID. A value of “1” means the facial animation ID mode shall be used and “0” means the facial animation ID mode shall not be used.
  • BodyIDFlag This field, which is only present in the binary representation, signals the presence of the body animation ID. A value of “1” means the body animation ID mode shall be used and “0” means the body animation ID mode shall not be used.
  • FaceFeatureFlag This field, which is only present in the binary representation, signals the presence of the face features. A value of “1” means the face feature tracking mode shall be used and “0” means the face feature tracking mode shall not be used.
  • BodyFeatureFlag This field, which is only present in the binary representation, signals the presence of the body features. A value of “1” means the body feature tracking mode shall be used and “0” means the body feature tracking mode shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used. “00” means that the absolute time stamp type shall be used, “01” means that the clock tick time stamp type shall be used, and “10” means that the clock tick time delta stamp type shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6.
  • ID ID to identify the sensed information with respect to a intelligent camera sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted. A value of zero indicates the highest priority and larger values indicate lower priorities. The default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself. activate Describes whether the sensor is activated.
  • FacialAnimationIDLength This field, which is only present in the binary representation, specifies the length of the following FacialAnimationID attribute. FacialAnimationID Describes the ID referencing the facial expression animation clip. BodyAnimationIDLength This field, which is only present in the binary representation, specifies the length of the following BodyAnimationID attribute. BodyAnimationID Describes the ID referencing the body animation clip. NumOfFaceFeature This field, which is only present in the binary representation, specifies the number of face feature points. FaceFeature Describes the 3D position of each of the face feature points detected by the camera.
  • the order of the elements corresponds to the order of the face feature points defined at the featureControl for face in 2.2.15 of ISO/IEC_23005-4 NumOfBodyFeature This field, which is only present in the binary representation, specifies the number of body feature points.
  • BodyFeature Describes the 3D position of each of the body feature points detected by the camera.
  • the order of the elements corresponds to the order of the body feature points defined at the featureControl for body in 2.2.14 of ISO/IEC_23005-4.
  • Source 48 denotes sensed information related to an ambient noise sensor using the XML.
  • a program source shown in Source 48 is only an example embodiment but does not limit the present invention.
  • An ambient noise sensor type is a tool for describing sensed information related to the ambient noise sensor.
  • the ambient noise sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the ambient noise sensor.
  • the life span denotes information on a valid period of a command of the ambient noise sensor.
  • the life span may be a second unit.
  • the value denotes information on an ambient noise sensor value expressed by a unit of sound intensity, for example, dB.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the ambient noise sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the ambient noise sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • lifespanFlag This field, which is only present in the binary representation, signals the presence of the life span attribute. A value of “1” means the lifespan shall be used and “0” means the lifespan shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute. A value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used. “00” means that the absolute time stamp type shall be used, “01” means that the clock tick time stamp type shall be used, and “10” means that the clock tick time delta stamp type shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a temperature sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • lifespan Describes the duration taken to measure the information based on the timestamp.
  • value Describes the sensed value of the ambient noise sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 49 denotes sensed information related to an atmospheric pressure sensor using the XML.
  • a program source shown in Source 49 is only an example embodiment but does not limit the present invention.
  • An atmospheric pressure sensor type is a tool for describing sensed information related to the atmospheric pressure sensor.
  • the atmospheric pressure sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the atmospheric pressure sensor.
  • the life span denotes information on a valid period of a command of the atmospheric pressure sensor.
  • the life span may be a second unit.
  • the value denotes information on an atmospheric pressure sensor value expressed by a unit of atmospheric pressure, for example, hPa.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the atmospheric pressure sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the atmospheric pressure sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to an atmospheric pressure sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated. value Describes the sensed value of the atmospheric pressure sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 50 denotes sensed information related to a velocity sensor using the XML.
  • a program source shown in Source 50 is only an example embodiment but does not limit the present invention.
  • a velocity sensor type is a tool for describing sensed information related to the velocity sensor.
  • the velocity sensor type may include attributes such as a timestamp, a life span, a velocity, a velocity value type, Vx, Vy, and Vz.
  • the timestamp denotes information on a sensing time of the velocity sensor.
  • the life span denotes information on a valid period of a command of the velocity sensor.
  • the life span may be a second unit.
  • the velocity denotes information on a velocity sensor value expressed by a unit of velocity, for example, m/s.
  • the velocity value type denotes a tool for indicating a 3D velocity vector.
  • the Vx denotes information on an x-axis value of the velocity sensor.
  • the Vy denotes information on a y-axis value of the velocity sensor.
  • the Vz denotes information on a z-axis value of the velocity sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the velocity sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the velocity sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • velocityFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a velocity sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • velocity Describes the sensed value of the velocity sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 51 denotes sensed information related to an angular velocity sensor using the XML.
  • a program source shown in Source 51 is only an example embodiment but does not limit the present invention.
  • An angular velocity sensor type is a tool for describing sensed information related to the angular velocity sensor.
  • the angular velocity sensor type may include attributes such as a timestamp, a life span, an angular velocity, an angular velocity value type, AVx, AVy, and AVz.
  • the timestamp denotes information on a sensing time of the angular velocity sensor.
  • the life span denotes information on a valid period of a command of the angular velocity sensor.
  • the life span may be a second unit.
  • the angular velocity denotes information on an angular velocity sensor value expressed by a unit of angular velocity, for example, radian.
  • the angular velocity value type denotes a tool for indicating a 3D angular velocity vector.
  • the AVx denotes information on a value of an x-axis rotation angular velocity of the angular velocity sensor.
  • the AVy denotes information on a value of a y-axis rotation angular velocity of the angular velocity sensor.
  • the AVz denotes information on a value of a z-axis rotation angular velocity of the angular velocity sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the angular velocity sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the angular velocity sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • angularvelocityFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to an angular velocity sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • angularvelocity Describes the sensed value of the angular velocity sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 52 denotes sensed information related to an angular acceleration sensor using the XML.
  • a program source shown in Source 52 is only an example embodiment but does not limit the present invention.
  • An angular acceleration sensor type is a tool for describing sensed information related to the angular acceleration sensor.
  • the angular acceleration sensor type may include attributes such as a timestamp, a life span, an angular acceleration, an angular acceleration value type, an AAx, an AAy, and an AAz.
  • the timestamp denotes information on a sensing time of the angular acceleration sensor.
  • the life span denotes information on a valid period of a command of the angular acceleration sensor.
  • the life span may be a second unit.
  • the angular acceleration denotes information on an angular acceleration sensor value expressed by a unit of angular acceleration, for example, radian/s 2 .
  • the angular acceleration value type denotes a tool for indicating a 3D angular acceleration vector.
  • the AAx denotes information on an x-axis value of the angular acceleration sensor.
  • the AAy denotes information on a y-axis value of the angular acceleration sensor.
  • the AAz denotes information on a z-axis value of the angular acceleration sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the angular acceleration sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the angular acceleration sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • angularaccelerationFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to an angular acceleration sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • angularacceleration Describes the sensed value of the angular acceleration sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 53 denotes sensed information related to a force sensor using the XML.
  • a program source shown in Source 53 is only an example embodiment but does not limit the present invention.
  • a force sensor type is a tool for describing sensed information related to the force sensor.
  • the force sensor type may include attributes such as a timestamp, a life span, a force, a force value type, FSx, FSy, and FSz.
  • the timestamp denotes information on a sensing time of the force sensor.
  • the life span denotes information on a valid period of a command of the force sensor.
  • the life span may be a second unit.
  • the force denotes information on a force sensor value expressed by a unit of force, for example, N.
  • the force value type denotes a tool for indicating a 3D force vector.
  • the FSx denotes information on an x-axis force value of the force sensor.
  • the FSy denotes information on a y-axis force value of the force sensor.
  • the FSz denotes information on a z-axis force value of the force sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the force sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the force sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • forceFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a force sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • force Describes the sensed value of the force sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 54 denotes sensed information related to a torque sensor using the XML.
  • a program source shown in Source 54 is only an example embodiment but does not limit the present invention.
  • a torque sensor type is a tool for describing sensed information related to the torque sensor.
  • the torque sensor type may include attributes such as a timestamp, a life span, a torque, a torque value type, TSx, TSy, and TSz.
  • the timestamp denotes information on a sensing time of the torque sensor.
  • the life span denotes information on a valid period of a command of the torque sensor.
  • the life span may be a second unit.
  • the torque denotes information on a torque sensor value expressed by a unit of torque, for example, N-mm.
  • the torque value type denotes a tool for indicating a 3D torque vector.
  • the TSx denotes information on an x-axis torque value of the torque sensor.
  • the TSy denotes information on a y-axis torque value of the torque sensor.
  • the TSz denotes information on a z-axis torque value of the force sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the torque sensor according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the torque sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • torqueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user- defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a torque sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated.
  • torque Describes the sensed value of the torque sensor in 3D with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 55 denotes sensed information related to a pressure sensor using the XML.
  • a program source shown in Source 55 is only an example embodiment but does not limit the present invention.
  • a pressure sensor type is a tool for describing sensed information related to the pressure sensor.
  • the pressure sensor type may include attributes such as a timestamp, a life span, and a value.
  • the timestamp denotes information on a sensing time of the pressure sensor.
  • the life span denotes information on a valid period of a command of the pressure sensor.
  • the life span may be a second unit.
  • the value denotes information on a pressure sensor value expressed by a unit of pressure, for example, N/mm 2 .
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the torque pressure according to the example embodiments.
  • Table 1-3 shows descriptor components semantics of the pressure sensor according to the example embodiments.
  • TimeStampFlag This field, which is only present in the binary representation, signals the presence of the timestamp element. A value of “1” means the timestamp shall be used and “0” means the timestamp shall not be used.
  • IDFlag This field, which is only present in the binary representation, signals the presence of the ID attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • sensorIdRefFlag This field, which is only present in the binary representation, signals the presence of the sensor ID reference attribute. A value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • linkedlistFlag This field, which is only present in the binary representation, signals the presence of the linked list attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • groupIDFlag This field, which is only present in the binary representation, signals the presence of the group ID attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • priorityFlag This field, which is only present in the binary representation, signals the presence of the priority attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • activateFlag This field, which is only present in the binary representation, signals the presence of the activation attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • valueFlag This field, which is only present in the binary representation, signals the presence of sensor value attribute.
  • a value of “1” means the attribute shall be used and “0” means the attribute shall not be used.
  • unitFlag This field, which is only present in the binary representation, signals the presence of unit attribute.
  • a value of “1” means the user-defined unit shall be used and “0” means the user-defined unit shall not be used.
  • TimeStampType This field, which is only present in the binary representation, describes which time stamp scheme shall be used.
  • AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC 23005-6.
  • ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time delta between the present and the past time, is defined in A.2.3 of ISO/IEC 23005-6. ID ID to identify the sensed information with respect to a pressure sensor.
  • sensorIdRefLength This field, which is only present in the binary representation, specifies the length of the following sensorIdRef attribute.
  • sensorIdRef References a sensor that has generated the information included in this specific sensed information.
  • linkedlistLength This field, which is only present in the binary representation, specifies the length of the following linkedlist attribute.
  • linkedlist Identifier for the next sensor of the multi-sensor structure that consists of a group of sensors in a way that each record contains a reference to the ID of the next sensor.
  • groupIDLength This field, which is only present in the binary representation, specifies the length of the following groupID attribute.
  • priority Describes a priority for sensed information with respect to other sensed information sharing the same point in time when the sensed information becomes adapted.
  • a value of zero indicates the highest priority and larger values indicate lower priorities.
  • the default value of the priority is zero. If there is more than one sensed information with the same priority, the order of process can be determined by the Adaptation engine itself.
  • activate Describes whether the sensor is activated. A value of “1” means the sensor is activated and “0” means the sensor is deactivated. value Describes the sensed value of the pressure sensor with respect to the default unit if the unit is not defined. Otherwise, use the unit type defined in the sensor capability.
  • unit Specifies the unit of the sensed value, if a unit other than the default unit is used, as a reference to a classification scheme term provided by UnitCS defined in xxx of ISO/IEC 23005-6 and use the binary representation defined above.
  • Source 0-0 may be a program representing a timestamp type using the XML.
  • the program source shown in Source 0-0 is only an example embodiment but does not limit the present invention.
  • Source 41-2 denotes a binary representation syntax related to the timestamp type according to the example embodiments.
  • FIG. 19 is a flowchart illustrating a virtual world processing method according to example embodiments.
  • the virtual world processing method may store a sensor capability related to capability of a sensor in operation 1110 .
  • a first value received from the sensor may be determined based on the sensor capability and a second value corresponding to the first value may be transmitted to the virtual world.
  • the sensor capability may include a maximum value and a minimum value measurable by the sensor.
  • the virtual world processing method may transmit the second value corresponding to the first value to the virtual world.
  • the sensor capability may include a unit of the first value measured by the sensor.
  • the sensor capability may include an offset value added to the first value measured by the sensor to obtain an absolute value.
  • the sensor capability may further include a number of values measurable by the sensor.
  • the sensor capability may further include a minimum input value required for the sensor to measure an output value.
  • the sensor capability may further include an SNR of the sensor.
  • the sensor capability may further include an error of the sensor.
  • the sensor capability may further include a position of the sensor.
  • the virtual world processing method may further include an operation (not shown) of storing a sensor adaptation preference for manipulation of the first value received from the sensor.
  • the operation of transmitting the first value may include generating a third value from the first value based on the sensor capability and generating the second value from the third value based on the sensor adaptation preference.
  • the sensor adaptation preference may include information on a method of applying the sensor adaptation preference to the first value.
  • the sensor adaptation preference may further include information on whether to activate the sensor in the virtual world.
  • the sensor adaptation preference may further include a unit of the second value used in the virtual world.
  • the sensor adaptation preference may further include a maximum value and a minimum value of the second value used in the virtual world.
  • the sensor adaptation preference may further include a number of the second values used in the virtual world.
  • FIG. 20 illustrates a flowchart of a virtual world processing method according to other example embodiments.
  • the virtual world processing method may perform initial setting to be input with information of a real world from a sensor in operation 1210 .
  • the initial setting may be an operation of activating the sensor.
  • the virtual world processing method may store a sensor capability as information on capability of the sensor and a sensor adaptation preference as information for manipulation of a value received from the sensor, in operation 1220 .
  • the virtual world processing method may measure information on a motion, state, intention, shape, and the like of a user of the real world through eth sensor, in operation 1230 .
  • operation 1230 may be repeated until the information is measured.
  • preprocessing with respect to the information may be performed in operation 1240 .
  • the virtual world processing method may control the sensor using sensed information which is a command for controlling the sensor in operation 1250 .
  • An adaptation RV may determine a first value received from the sensor based on the sensor capability and transmit a second value corresponding to the first value to a virtual world, in operation 1260 .
  • a third value may be generated from the first value based on the sensor capability, the second value may be generated from the third value based on the sensor adaptation preference, and the second value may be transmitted to the virtual world.
  • FIG. 21 illustrates an operation of using a virtual world processing apparatus according to example embodiments.
  • a user 1310 of a real world may input his or her intention through a sensor 1301 .
  • the sensor 1301 may include a motion sensor configured to measure a motion of the user 1310 of the real world, and remote pointers attached to ends of arms and legs of the user 1310 and configured to measure directions and positions indicated by the ends of the arms and legs.
  • a sensor signal may be transmitted to the virtual world processing apparatus, the sensor signal which includes CI 1302 related to an arm opening motion, a still standing state, positions of hands and feet, an open angle of a hand, and the like of the user 1310 .
  • the CI 1302 may include a sensor capability, a sensor adaptation preference, and sensed information.
  • the CI 1302 may include position information of the arms and the legs of the user 1310 , expressed by X real , Y real , and Z real denoting values on an x-axis, y-axis, and z-axis and ⁇ Xreal , ⁇ Yreal , and ⁇ Zreal denoting angles with respect to the x-axis, y-axis, and z-axis.
  • the virtual world processing apparatus may include an RV engine 1320 .
  • the RV engine 1320 may convert information of the real world to information applicable to a virtual world, using the CI 1302 included in the sensor signal.
  • the RV engine 1320 may convert VWI 1303 using the CI 1302 .
  • the VWI 1303 denotes information on the virtual world.
  • the VWI 1303 may include information on an object of the virtual world or elements constituting the object.
  • the VWI 1303 may include virtual world object information 1304 and avatar information 1305 .
  • the virtual world object information 1304 denotes information on the object of the virtual world.
  • the virtual world object information 1304 may include an object ID denoting ID information for recognizing identity of the object of the virtual world, and an object control and scale denoting information for controlling a state, size, and the like of the object of the virtual world.
  • the virtual world processing apparatus may control the virtual world object information 1304 and the avatar information 1305 by a control command.
  • the control command may include commands such as generation, disappearance, copy, and the like.
  • the virtual world processing apparatus may generate the commands by selecting information to be manipulated from the virtual world object information 1304 and the avatar information 1305 , along with the control command, and designating an ID corresponding to the selected information.
  • Source 56 denotes a method of constructing the control command using an XML.
  • a program source shown in Source 56 is only an example embodiment but does not limit the present invention.
  • the RV engine 1320 may convert the VWI 1303 by applying information on the arm opening motion, the still standing state, the positions of hands and feet, the open angle of a hand, and the like, using the CI 1302 .
  • the RV engine 1320 may transmit information 1306 on the converted VWI to the virtual world.
  • the information 1306 on the converted VWI may include position information of arms and legs of an avatar of the virtual world, expressed by X virtual , Y virtual , and Z virtual denoting values on the x-axis, y-axis, and z-axis and ⁇ Xvirtual , ⁇ Yvirtual , and ⁇ Zvirtual denoting angles with respect to the x-axis, y-axis, and z-axis.
  • the information 1306 may include information on a size of the object of the virtual world, expressed by a scale(s,d,h) virtual denoting a width value, a height value, and a depth value of the object.
  • the avatar in a virtual world 1330 of before transmission of the information 1306 , the avatar is holding the object.
  • the avatar of the virtual world may scale up the object.
  • the CI 1302 related to the arm opening motion, the still standing state, the positions of hands and feet, the open angle of a hand, and the like may be generated through the sensor 1301 .
  • the RV engine 1320 may convert the CI 1302 related to the user 1310 of the virtual world, which is data measured in the real world, to the information applicable to the virtual world.
  • the converted information may be applied to a structure of information related to the avatar and the object of the virtual world. Therefore, the motion of holding and enlarging the object may be reflected to the avatar, and the object may be enlarged.
  • Example embodiments include computer-readable media including program instructions to implement various operations embodied by a computer.
  • the media may also include, alone or in combination with the program instructions, data files, data structures, tables, and the like.
  • the media and program instructions may be those specially designed and constructed for the purposes of example embodiments, or they may be of the kind well known and available to those having skill in the computer software arts.
  • Examples of computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD ROM disks; magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM).
  • ROM read-only memory devices
  • RAM random access memory
  • Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.
  • the described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described example embodiments, or vice versa.

Abstract

Disclosed are a virtual world processing device and method. By way of example, data collected from the real world is converted to binary form data which is then transmitted, or is converted to XML data, or the converted XML data is further converted to binary form data which is then transmitted, thereby allowing the data transmission rate to be increased and a low bandwidth to be used, and, in the case of a data-receiving adaptation RV engine, the complexity of the adaptation RV engine can be reduced as there is no need to include an XML parser.

Description

  • This application is a National Phase Application, under 35 U.S.C. 371, of International Application No. PCT/KR2010/004087, filed Jun. 23, 2010, which claimed the benefit of priority to Korean Application No. 10-2009-0057313 filed Jun. 25, 2009; Korean Application No. 10-2009-0101434 filed Oct. 23, 2009; U.S. Provisional Application No. 61/255,636 filed Oct. 28, 2009; Korean Application No. 10-2009-0104474 filed Oct. 30, 2009; Korean Application No. 10-2010-0003607 filed Jan. 14, 2010; and Korean Application No. 10-2010-0006119 filed Jan. 22, 2010, the disclosures of which are incorporated herein by reference.
  • TECHNICAL FIELD
  • Example embodiments relate to a method and apparatus for processing a virtual world, and more particularly to, a method and apparatus for applying information of a real world to a virtual world.
  • BACKGROUND ART
  • Currently, an interest in experience-type games has been increasing. Microsoft Corporation introduced “Project Natal” at the “E3 2009” Press Conference. “Project Natal” may provide a user body motion capturing function, a face recognition function, and a voice recognition function by combining Microsoft's Xbox 360 game console with a separate sensor device consisting of a depth/color camera and a microphone array, thereby enabling a user to interact with a virtual world without a dedicated controller. Also, Sony Corporation introduced “Wand” which is an experience-type game motion controller. The “Wand” enables interaction with a virtual world through input of a motion trajectory of a controller by applying, to the PlayStation 3 game console, a location/direction sensing technology obtained by combining a color camera, a marker, and an ultrasonic sensor.
  • The interaction between a real world and a virtual world has two directions. In one direction, data information obtained by a sensor in the real world may be reflected to the virtual world. In the other direction, data information obtained from the virtual world may be reflected to the real world using an actuator. Embodiments suggest a control system, control method, and commanding structure for applying data obtained through a sensor in the real world to the virtual world, to achieve the interaction between the real world and the virtual world.
  • DISCLOSURE OF INVENTION Technical Goals Technical Solutions
  • According to example embodiments, there is provided a virtual world processing apparatus including a sensor to encode information collected from a real world into first metadata, an adaptation virtual world to real world (VR) unit to encode information on a virtual world into second metadata, and an adaptation real world to virtual world (RV) unit to generate information to be applied to the virtual world based on the first metadata and the second metadata and to encode the generated information into third metadata. The adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of a binary form, and transmits the binary-encoded second metadata to the adaptation RV unit. Here, the binary-encoded second metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes. The adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of extensible markup language (XML) form, and transmits the XML-encoded second metadata to the adaptation RV unit. The adaptation VR unit may generate the second metadata by encoding the information on the virtual world into data of an XML form and encoding the XML-encoded data to data of a binary form, and transmits the binary-encoded second metadata to the adaptation RV unit.
  • According to another aspect of example embodiments, the sensor may generate the first metadata by encoding the information collected from the real world into data of a binary form, and transmits the binary-encoded first metadata to the adaptation RV unit. The binary-encoded first metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • The sensor may generate the first metadata by encoding the information collected from the real world into data of an XML form, and transmit the XML-encoded first metadata to the adaptation RV unit.
  • The sensor may generate the first metadata by encoding the information collected from the real world into data of an XML form and encoding the XML-encoded data into data of a binary form, and transmit the binary-encoded first metadata to the adaptation RV unit. According to still another aspect, there is provided a virtual world processing method including encoding information collected from a real world into first metadata, encoding information on a virtual world into second metadata, generating information to be applied to the virtual world based on the first metadata and the second metadata, and encoding the generated information into third metadata.
  • The encoding of the information collected from the real world into the first metadata may include generating the first metadata by encoding the information collected from the real world into data of a binary form comprising a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • The encoding of the information collected from the real world into the first metadata may include generating the first metadata by encoding the information collected from the real world into data of an XML form and encoding the XML-encoded data into data of a binary form.
  • The encoding of the information on the virtual world into the second metadata may include generating the second metadata by encoding the information on the virtual world into data of a binary form comprising a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • The encoding of the information on the virtual world into the second metadata may include generating the second metadata by encoding the information on the virtual world into data of an XML form and encoding the XML-encoded data to data of a binary form.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a view illustrating an operation of manipulating an object of a virtual world using a sensor according to example embodiments of the present invention;
  • FIG. 2 is a view illustrating a system for manipulating an object of a virtual world using a sensor according to example embodiments;
  • FIG. 3 is a view illustrating a virtual world processing apparatus for manipulating an object of a virtual world using a sensor, according to other example embodiments;
  • FIGS. 4 to 6 are views illustrating a sensor and an adaptation real world to virtual world (RV) unit according to example embodiments;
  • FIGS. 7 to 9 are views illustrating an adaptation virtual world to real world (VR) engine and an actuator according to example embodiments;
  • FIG. 10 is a view illustrating a structure of a virtual world processing apparatus according to example embodiments;
  • FIG. 11 is a view illustrating a structure of a virtual world processing apparatus according to other embodiments;
  • FIG. 12 is a illustrating a sensor capability base type according to example embodiments;
  • FIG. 13 is a view illustrating syntax of a sensor capability base type according to example embodiments;
  • FIG. 14 is a view illustrating syntax of sensor capability base attributes according to example embodiments;
  • FIG. 15 is a view illustrating a sensor adaptation preference base type according to example embodiments;
  • FIG. 16 is a view illustrating syntax of a sensor adaptation preference base type according to example embodiments;
  • FIG. 17 is a view illustrating syntax of sensor adaptation preference base attributes according to example embodiments;
  • FIG. 18 is a view illustrating a sensed information base type according to example embodiments;
  • FIG. 19 is a flowchart illustrating a virtual world processing method according to example embodiments;
  • FIG. 20 is a view illustrating a flowchart of a virtual world processing method according to other example embodiments; and
  • FIG. 21 is a view illustrating an operation of using a virtual world processing apparatus according to example embodiments.
  • BEST MODE FOR CARRYING OUT THE INVENTION
  • Reference will now be made in detail to example embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. Example embodiments are described below in order to explain example embodiments by referring to the figures.
  • A term ‘object’ used herein may include an object, a thing, an avatar, and the like, implemented and expressed in a virtual world.
  • Hereinafter, the example embodiments will be described with reference to the accompanying drawings.
  • FIG. 1 illustrates an operation of manipulating an object 120 of a virtual world, using a sensor 100 according to example embodiments.
  • Referring to FIG. 1, a user 110 of a real world may manipulate the object 120 of the virtual world using the sensor 100. The user 110 may input his or her motion, state, intention, shape, and the like through the sensor 100. The sensor 100 may transmit control information (CI) related to the motion, state, intention, shape, and the like of the user 110, the CI included in a sensor signal, to a virtual world processing apparatus.
  • Depending on embodiments, the user 110 of the real world may include humans, animals, plants, inanimate objects such as articles, and even surrounding environment of the user 110.
  • FIG. 2 illustrates a system for manipulating an object of a virtual world using a sensor according to example embodiments.
  • Referring to FIG. 2, the signal that includes CI 201 related to the motion, state, intention, shape, and the like of a user of a real world 210, the CI 201 input through a sensor as a real world device, may be transmitted to a virtual world processing apparatus. Depending on embodiments, the CI 201 related to the motion, state, intention, shape, and the like of the user may include a sensor capability, a sensor adaptation preference, and sensed information, which will be described in detail with reference to FIG. 12 through FIG. 21.
  • The virtual world processing apparatus may include an adaptation real world to virtual world (RV) 220. The adaptation RV 220 may be implemented by an RV engine. The adaptation RV 220 may convert information of the real world 210 to information applicable to a virtual world 240, using the CI 201 related to the motion, state, intention, shape, and the like of the user of the real world 210, the CI 201 included in the sensor signal.
  • Depending on embodiments, the adaptation RV 220 may convert virtual world information (VWI) 202 using the CI 201 related to the motion, state, intention, shape, and the like of the user of the real world 210.
  • The VWI 202 denotes information on the virtual world 240. For example, the VWI 202 may include information on an object of the virtual world 240 or elements constituting the object.
  • The virtual world processing apparatus may transmit converted information 203 converted by the adaptation RV 220 to the virtual world 240 through adaptation real world to virtual world/virtual world to real world (RV/VR) 230.
  • Table 1 illustrates structures shown in FIG. 2.
  • TABLE 1
    SIDC Sensory input device VWI Virtual world
    capabilities. information
    Another expression of
    sensor capability
    USIP User sensory input SODC Sensory output device
    preferences. capabilities
    Another expression of
    sensor adaptation
    preference
    SIDCmd Sensory input device USOP User sensory output
    commands preferences
    Another expression of
    sensed information
    VWC Virtual world capabilities SODCmd Sensory output device
    commands
    VWP Virtual world preferences SEM Sensory effect
    metadata
    VWEM Virtual world effect SI Sensory information
    metadata
  • FIG. 3 is a view illustrating a virtual world processing apparatus for manipulating an object of a virtual world using a sensor 250 according to other example embodiments.
  • Referring to FIG. 3, the virtual world processing apparatus includes the sensor 250, an adaptation RV unit 255, and an adaptation VR unit 260.
  • The sensor 250 may collect information on a motion, state, intention, shape, and the like of a user of a real world. The information collected by the sensor 250 may include sensed information.
  • Depending on embodiments, the sensor 250 may include an input unit. The input unit may be input with a sensor adaptation preference from the user of the real world.
  • The sensor 250 may encode the information collected from the real world, to first metadata. The sensor 250 may transmit the first metadata to the adaptation RV unit 255.
  • The sensor 250 may generate the first metadata by encoding the information collected from the real world into data of a binary form. The binary-encoded first metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes of the binary encoding syntax. The sensor 250 may transmit the binary-encoded first metadata to the adaptation RV unit 255.
  • Also, the sensor 250 may generate the first metadata by encoding the information collected from the real world into data of an extensible markup language (XML) form. The sensor 250 may transmit the XML-encoded first metadata to the adaptation RV unit 255.
  • In addition, the sensor 250 may generate the first metadata by encoding the information collected from the real world into data of the XML form and encoding the XML-encoded data into data of the binary form. The sensor 250 may transmit the binary-encoded first metadata to the adaptation RV unit 255.
  • Depending on embodiments, the sensor 250 may include a metadata encoder 251 configured to encode the collected information into metadata.
  • The metadata encoder 251 may encode the collected information to the first metadata. The sensor 250 may transmit the first metadata to the adaptation RV unit 255. A metadata decoder 261 included in the adaptation RV unit 255 may decode the first metadata received from the sensor 250.
  • Depending on embodiments, the metadata encoder 251 may include at least one of an XML encoder that encodes the collected information into XML data and a binary encoder that encodes the collected information into binary data. The metadata decoder 251 may include at least one of an XML decoder that decodes received XML data and a binary decoder that decodes received binary data.
  • Hereinafter, encoding of the information collected by the sensor 250 and decoding of the data received by the adaptation RV unit 255 according to example embodiments will be described with reference to FIGS. 4 to 6.
  • FIGS. 4 to 6 are views illustrating a sensor 271 and an adaptation RV unit according to example embodiments.
  • Referring to FIG. 4, the sensor 271 may include an XML encoder 272. The XML encoder 272 may encode information collected from the real world by the sensor 271, for example sensed information, into metadata of the XML form.
  • The sensor 271 may transmit the metadata encoded by the XML encoder 272 to an adaptation RV unit 274.
  • The adaptation RV unit 274 may include an XML decoder 273. The XML decoder 273 may decode the metadata received from the sensor 271.
  • Referring to FIG. 5, a sensor 275 according to present example embodiments may include an XML encoder 276 and a binary encoder 277. The XML encoder 276 may encode information collected from the real world by the sensor 275, for example sensed information, into data of the XML form. The binary encoder 277 may encode the data encoded by the XML encoder 276 into data of the binary form.
  • In addition, the sensor 275 may transmit the metadata encoded by the XML encoder 276 and the binary encoder 277 to an adaptation RV 280.
  • The adaptation RV 280 may include a binary decoder 278 and an XML decoder 279. The binary decoder 278 may decode the metadata received from the sensor 275 into data of the XML form. The XML decoder 279 may decode again the data decoded into the XML form by the binary decoder 278.
  • Referring to FIG. 6, a sensor 281 according to present example embodiments may include a binary encoder 282. The binary encoder 282 may encode information collected from the real world by the sensor 281, for example sensed information, into metadata of the binary form.
  • The sensor 281 may transmit the metadata encoded by the binary encoder 282 to an adaptation RV 284 unit.
  • The adaptation RV 284 unit may include a binary decoder 283. The binary decoder 283 may decode the metadata received from the sensor 281.
  • Referring back to FIG. 3, the adaptation VR unit 260 may encode information on the virtual world 265 into second metadata.
  • The adaptation VR unit 260 may generate the second metadata by encoding the information on the virtual world 265 into data of the binary form. The binary-encoded second metadata may include a binary encoding syntax, a number of bits of attributes of the binary encoding syntax, and a mnemonic of the attributes of the binary encoding syntax. The adaptation VR unit 260 may transmit the binary-encoded second metadata to the adaptation RV unit 255.
  • Also, the adaptation VR unit 260 may generate the second metadata by encoding information collected from the virtual world 265 into the data of the XML form. The adaptation VR unit 260 may transmit the XML-encoded second metadata to the adaptation RV unit 255.
  • Furthermore, the adaptation VR unit 260 may generate the second metadata by encoding information collected from the virtual world 265 into the data of the XML form and encoding the XML-encoded data to data of the binary form. The adaptation VR unit 260 may transmit the binary-encoded second metadata to the adaptation RV unit 255.
  • Depending on embodiments, the adaptation VR unit 260 may include a metadata encoder 262. The metadata encoder 262 may encode the information on the virtual world 265 into the second metadata.
  • Depending on embodiments, the metadata encoder 262 may include at least one of an XML encoder that encodes the information on the virtual world 265 into metadata of the XML form and a binary encoder that encodes the information on the virtual world 265 into metadata of the binary form.
  • Depending on embodiments, the metadata encoder 262 may include an XML encoder. The XML encoder may encode the information on the virtual world 265 into data of the XML form.
  • In addition, the metadata encoder 262 may include an XML encoder and a binary encoder. The XML encoder may encode the information on the virtual world 265 into data of the XML form. The binary encoder may encode the data of the XML form back to data of the binary form.
  • Also, the metadata encoder 262 may include a binary encoder. The binary encoder may encode the information on the virtual world 265 into data of the binary form.
  • The adaptation RV unit 255 may generate information to be applied to the virtual world 265, based on the first metadata transmitted from the sensor 250 and the second metadata transmitted from the adaptation VR unit 260. Here, the adaptation RV unit 255 may encode the generated information into third metadata.
  • Depending on embodiments, the adaptation RV unit 255 may include a metadata decoder 256, a metadata decoder 258, and a metadata encoder 257.
  • The adaptation RV unit 255 may generate information to be applied to the virtual world 265, based on information decoded from the first metadata by the metadata decoder 256 and information decoded from the second metadata by the metadata decoder 258. Here, the adaptation RV unit 255 may generate the information to be applied to the virtual world 265, such that the information corresponds to virtual world object characteristics and sensed information included in the second metadata.
  • The metadata encoder 257 may encode the information, which is generated by the adaptation RV unit 255 and to be applied to the virtual world 265, into the third metadata. In addition, the adaptation RV unit 255 may transmit the third metadata to the adaptation VR unit 260.
  • The adaptation VR unit 260 may include the metadata decoder 261. The metadata decoder 261 may decode the third metadata. The adaptation VR unit 260 may convert attributes of an object of the virtual world 265 based on the decoded information and, in addition, apply the converted attributes to the virtual world 265.
  • A virtual world processing system may transmit the information on the virtual world 265 to an actuator of the real world so that the information is reflected to the real world. Hereinafter, example embodiments in which the information on the virtual world 265 is reflected to the real world will be described in detail with reference to FIGS. 7 to 9.
  • FIGS. 7 to 9 are views illustrating an adaptation VR engine 285 and an actuator 288 according to example embodiments.
  • Referring to FIG. 7, the adaptation VR engine 285 may include an XML encoder 286. The adaptation VR engine 285, as an example embodiment of the adaptation RV unit 255 shown in FIG. 3, may transmit the information on the virtual world 265 to the actuator 288 of a real world so that the information on the virtual world 265 is reflected to the real world.
  • The adaptation VR unit 260 may collect information on a change in attributes of an object of the virtual world 265, and transmit the collected information to the adaptation VR engine 285. The adaptation VR 285 may include the XML encoder 286. The XML encoder 286 may encode the information on the virtual world 265 into data of the XML form. In addition, the adaptation VR engine 285 may transmit the data encoded by the XML encoder 286 to the actuator 288.
  • The actuator 288 may include the XML decoder 287. The XML decoder 287 may decode the XML data received from the adaptation VR engine 285.
  • The actuator 288 may operate corresponding to the information decoded by the XML decoder 287.
  • Referring to FIG. 8, an adaptation VR engine 290 may include an XML encoder 291 and a binary encoder 292.
  • The adaptation VR unit 260 may collect information on a change in attributes of the object of the virtual world 265, and transmit the collected information to the adaptation VR engine 290. The adaptation VR engine 290 may include the XML encoder 291 and the binary encoder 292. The XML encoder 291 may encode the information on the virtual world 265 into data of the XML form. The binary encoder 292 may encode the data encoded by the XML encoder 291 back to data of the binary form. In addition, the adaptation VR engine 290 may transmit the data encoded by the XML encoder 292 to an actuator 295.
  • The actuator 295 may include a binary decoder 293 and an XML decoder 294. The binary decoder 293 may decode the binary data received from the adaptation VR engine 290 into data of the XML form. The XML decoder 294 may decode the data decoded into the XML form by the binary decoder 293.
  • The actuator 295 may operate corresponding to the information decoded by the XML decoder 294.
  • Referring to FIG. 9, an adaptation VR engine 296 may include a binary encoder 297.
  • The adaptation VR unit 260 may collect information on a change in attributes of the object of the virtual world 265, and transmit the collected information to the adaptation VR engine 296. The adaptation VR engine 296 may include a binary encoder 297. The binary encoder 297 may encode received information on the virtual world into data of the binary form. In addition, the adaptation VR engine 296 may transmit the data encoded by the binary encoder 297 to an actuator 299.
  • The actuator 299 may include a binary decoder 298. The binary decoder 298 may decode the binary data received from the adaptation VR engine 296.
  • The actuator 299 may operate corresponding to the information decoded by the binary decoder 298.
  • FIG. 10 is a view illustrating a structure of a virtual world processing apparatus 300 according to example embodiments.
  • Referring to FIG. 10, the virtual world processing apparatus 300 includes a storage unit 310 and a processing unit 320.
  • The storage unit 310 stores sensor capability related to a sensor.
  • The sensor may measure a motion, state, intention, shape, and the like of a user of a real world. The sensor may be expressed as a sensory input device. Depending on embodiments, the sensor may be classified according to sensor types including (1) acoustic, sound, and vibration, (2) automotive and transportation, (3) chemical, (4) electric current, electric potential, magnetic, and radio, (5) environment and weather, (6) flow, (7) ionizing radiation, and subatomic particles, (8) navigation instruments, (9) position, angle, displacement, distance, speed, and acceleration, (10) optical, light, and imaging, (11) pressure, force, density, and level, (12) thermal, heat, and temperature, (13) proximity and presence, and (14) sensor technology.
  • Table 2 illustrates examples sensor according to the sensor types. Therefore, the sensors in Table 2 are suggested only as an embodiment but not limiting.
  • TABLE 2
    sensor type list of sensors
    (1) acoustic, sound, geophone
    vibration hydrophone
    lace sensor, a guitar pickup
    microphone
    seismometer
    accelerometer
    (2) automotive, crank sensor
    transportation curb feeler
    defect detector
    map sensor
    parking sensors
    parktronic
    radar gun
    speedometer
    speed sensor
    throttle position sensor
    variable reluctance sensor
    wheel speed sensor
    (3) chemical breathalyzer
    carbon dioxide sensor
    carbon monoxide detector
    catalytic bead sensor
    chemical field-effect transistor
    electronic nose
    electrolyte-insulator-semiconductor sensor
    hydrogen sensor
    infrared point sensor
    ion-selective electrode
    nondispersive infrared sensor
    microwave chemistry sensor
    nitrogen oxide sensor
    optode
    oxygen sensor
    pellistor
    pH glass electrode
    potentiometric sensor
    redox electrode
    smoke detector
    zinc oxide nanorod sensor
    (4) electric current, ammeter
    electric potential, current sensor
    magnetic, radio galvanometer
    hall effect sensor
    hall probe
    leaf electroscope
    magnetic anomaly detector
    magnetometer
    metal detector
    multimeter
    ohmmeter
    voltmeter
    watt-hour meter
    (5) environment, fish counter
    weather gas detector
    hygrometer
    pyranometer
    pyrgeometer
    rain gauge
    rain sensor
    seismometers
    (6) flow air flow meter
    flow sensor
    gas meter
    mass flow sensor
    water meter
    (7) ionizing radiation, bubble chamber
    subatomic particles cloud chamber
    geiger counter
    neutron detection
    particle detector
    scintillation counter
    scintillator
    wire chamber
    (8) navigation air speed indicator
    instruments altimeter
    attitude indicator
    fluxgate compass
    gyroscope
    inertial reference unit
    magnetic compass
    MHD sensor
    ring laser gyroscope
    turn coordinator
    variometer
    vibrating structure gyroscope
    yaw rate sensor
    (9) position, angle, accelerometer
    displacement, inclinometer
    distance, speed, laser rangefinder
    acceleration linear encoder
    linear variable differential transformer (LVDT)
    liquid capacitive inclinometers
    odometer
    piezoelectric accelerometer
    position sensor
    rotary encoder
    rotary variable differential transformer
    selsyn
    tachometer
    (10) optical, light, charge-coupled device
    imaging colorimeter
    infra-red sensor
    LED as light sensor
    nichols radiometer
    fiber optic sensors
    photodiode
    photomultiplier tubes
    phototransistor
    photoelectric sensor
    photoionization detector
    photomultiplier
    photoresistor
    photoswitch
    phototube
    proximity sensor
    scintillometer
    shack-Hartmann
    wavefront sensor
    (11) pressure, force, anemometer
    density, level bhangmeter
    barograph
    barometer
    hydrometer
    Level sensor
    Load cell
    magnetic level gauge
    oscillating U-tube
    pressure sensor
    piezoelectric sensor
    pressure gauge
    strain gauge
    torque sensor
    viscometer
    (12) thermal, heat, bolometer
    temperature calorimeter
    heat flux sensor
    infrared thermometer
    microbolometer
    microwave radiometer
    net radiometer
    resistance temperature detector
    resistance thermometer
    thermistor
    thermocouple
    thermometer
    (13) proximity, alarm sensor
    presence bedwetting alarm
    motion detector
    occupancy sensor
    passive infrared sensor
    reed switch
    stud finder
    triangulation sensor
    touch switch
    wired glove
    (14) sensor active pixel sensor
    technology machine vision
    biochip
    biosensor
    capacitance probe
    catadioptric sensor
    carbon paste electrode
    displacement receiver
    electromechanical film
    electro-optical sensor
    image sensor
    inductive sensor
    intelligent sensor
    lab-on-a-chip
    leaf sensor
    RADAR
    sensor array
    sensor node
    soft sensor
    staring array
    transducer
    ultrasonic sensor
    video sensor
  • For example, the microphone belonging to a sensor type (1) acoustic, sound, and vibration may collect voice of the user of the real world and ambient sounds of the user. The speed sensor belonging to the sensor type (2) automotive and transportation may measure speed of the user of the real world and speed of an object such as a vehicle of the real world. The oxygen sensor belonging to the sensor type (3) chemical may measure an oxygen ratio in ambient air around the user of the real world and an oxygen ratio in liquid around the user of the real world. The metal detector belonging to the sensor type (4) electric current, electric potential, magnetic, and radio may detect metallic substances present in or around the user of the real world. The rain sensor belonging to the sensor type (5) environment and weather may detect whether it is raining in the real world. The flow sensor belonging to the sensor type (6) flow may measure a ratio of a fluid flow of the real world. The scintillator belonging to the sensor type (7) ionizing radiation and subatomic particles may measure a ratio or radiation present in or around the user of the real world. The variometer belonging to the sensor type (8) navigation instruments may measure a vertical movement speed of or around the user of the real world. The odometer belonging to the sensor type (9) position, angle, displacement, distance, speed, and acceleration may measure a traveling distance of an object of the real world, such as a vehicle. The phototransistor belonging to the sensor type (10) optical, light, and imaging may measure light of the real world. The barometer belonging to the sensor type (11) pressure, force, density, and level may measure an atmospheric pressure of the real world. The bolometer belonging to the sensor type (12) thermal, heat, and temperature may measure radiation rays of the real world. The motion detector belonging to the sensor type (13) proximity and presence may measure a motion of the user of the real world. The biosensor belonging to the sensor type (14) may measure biological characteristics of the user of the real world.
  • FIG. 11 illustrates a structure of a virtual world processing apparatus according to other example embodiments.
  • Referring to FIG. 11, an input device 360 according to the present embodiments may be input with a sensor adaptation preference 361 by a user of a real world. Depending on embodiments, the input device 360 may be modularized and inserted in a sensor 370 or a virtual world processing apparatus 350. The sensor adaptation preference 361 will be described in further detail with reference to FIGS. 15 to 17.
  • The sensor 370 may transmit a sensor capability 371 and sensed information 372 to the virtual world processing apparatus 350. The sensor capability 371 and the sensed information 372 will be described in further detail with reference to FIGS. 12 to 14, and 18.
  • The virtual world processing apparatus 350 may include a signal processing unit 351 and an adaptation unit 352.
  • The signal processing unit 351 may receive the sensor capability 371 and the sensed information 372, and perform signal-processing with respect to the sensor capability 371 and the sensed information 372. Depending on embodiments, the signal processing unit 351 may filter and validate the sensor capability 371 and the sensed information 372.
  • The adaptation unit 352 may receive the sensor adaptation preference 361 from the input device 360. In addition, based on the received sensor adaptation preference 361, the adaptation unit 352 may perform adaptation with respect to the information signal-processed by the signal processing unit 351 so that the information is applied to a virtual world 380. In addition, the virtual world processing apparatus 350 may apply the information having undergone the adaptation by the adaptation unit 352 to the virtual world 380.
  • The sensor capability denotes information on capability of a sensor.
  • A sensor capability base type denotes a base type of the sensor capability. Depending on embodiments, the sensor capability base type may be a base abstract type of the metadata related to a sensor capability commonly applied to all types of sensors, as part of metadata types related to the sensor capability.
  • Hereinafter, the sensor capability and the sensor capability base type will be described in detail with reference to FIGS. 12 to 14.
  • FIG. 12 illustrates a sensor capability base type 400 according to example embodiments.
  • Referring to FIG. 12, the sensor capability base 400 may include sensor capability base attributes 410 and any attributes 420.
  • The sensor capability base attributes 410 denote a group of sensor capabilities basically included in the sensor capability base type 400.
  • The any attributes 420 denote a group of additional sensor capabilities of a sensor. The any attributes 420 may be unique additional sensor capabilities which are applicable to an arbitrary sensor. The any attributes 420 may allow scalability for inclusion of attributes other than the base attributes.
  • FIG. 13 illustrates syntax 500 of a sensor capability base type according to example embodiments.
  • Referring to FIG. 13, the syntax 500 of the sensor capability base type may include a diagram 510, attributes 520, and a source 530.
  • The diagram 510 may include a diagram of the sensor capability base type.
  • The attributes 520 may include sensor capability base attributes and any attributes.
  • The source 530 may be a program representing the sensor capability base type using an extensible markup language (XML). However, the source 530 shown in FIG. 13 is suggested by way of example but not limiting.
  • FIG. 14 illustrates syntax 600 of sensor capability base attributes according to example embodiments.
  • Referring to FIG. 14, the syntax 600 of the sensor capability base attributes may include a diagram 610, attributes 620, and a source 630.
  • The diagram 610 may include a diagram of the sensor capability base attributes.
  • The attributes 620 may include a unit 601, a maximum value 602, a minimum value 603, an offset 604, a number of levels 605, a sensitivity 606, a signal to noise ratio (SNR) 607, and an accuracy 608.
  • The unit 601 is a unit of values measured by a sensor. Depending on embodiments, for example, when the sensor is a thermometer, the unit 601 may be Celsius (° C.) and Fahrenheit (° F.). When the sensor is a speed sensor, the unit 601 may be kilometers per hour (km/h) and meters per second (m/s).
  • The maximum value 602 and the minimum value 603 denote a maximum value and a minimum value measurable by the sensor, respectively. Depending on embodiments, for example, when the sensor is a thermometer, the maximum value 602 may be 50° C. and the minimum value 603 may be 0° C. Even in the same type of sensor, for example, the thermometer, the maximum value 602 and the minimum value 603 may be varied according to use and function of the sensor.
  • The offset 604 denotes an offset value added to a value measured by the sensor to obtain an absolute value. Depending on embodiments, for example, presuming that the sensor is a speed sensor and a user or an object of a real world stays still, when a value other than zero is measured as speed, the sensor may determine the offset 604 to a value making the speed zero. For example, when −1 km/h is measured as speed of a vehicle of the real world, the offset 604 may be 1 km/h.
  • The number of levels 605 denotes a number of values measurable by the sensor. Depending on embodiments, for example, presuming that the sensor is a thermometer and the maximum value and the minimum value are 50° C. and 0° C., respectively, when the number of levels 605 is 5, the sensor may measure five values, that is, 10° C., 20° C., 30° C., 40° C., and 50° C. Even when temperature of the real world is 27° C., not only when 20° C., the temperature may be measured as 20° C. through round-down. Alternative, in this case, the temperature may be measured as 30° C. through roundup.
  • The sensitivity 606 denotes a minimum input value required for the sensor to measure an output value. Depending on embodiments, for example, when the sensor is a thermometer and the sensitivity 606 is 1° C., the sensor may not measure a temperature change less than 1° C. but measure only the temperature change of at least 1° C.
  • The SNR 607 denotes a relative degree of a signal measured by the sensor with respect to a noise. Depending on embodiments, presuming that the sensor is a microphone to measure and a vocal sound of the user of the real world is to be measured, when an ambient noise is large, the SNR 607 of the sensor may be relatively small.
  • The accuracy 608 denotes an error of the sensor. Depending on embodiments, when the sensor is a microphone, the accuracy 608 may be a measurement error caused by variation of a propagation speed of a sound according to temperature, humidity, and the like. Alternatively, the accuracy 608 may be determined through a statistical error of the values already measured by the sensor.
  • Depending on embodiments, the attributes 620 may further include a position. The position denotes a position of the sensor. When the sensor is a thermometer, the position of the sensor may be an armpit of the user of the real world. The position may include longitude and latitude, and height and direction from a ground surface.
  • The unit 601, the maximum value 602, the minimum value 603, the offset 604, the number of levels 605, the sensitivity 606, the SNR 607, the accuracy 608, and the position, as the sensor capability base attributes, may be rearranged as shown in Table 3.
  • TABLE 3
    Name Definition
    Unit
    601 the unit of value
    maxValue the maximum value that the input device (sensor) can
    602 provide. The terms will be different according to
    the individual device type.
    minValue the minimum value that the input device (sensor) can
    603 provide. The terms will be different according to
    the individual device type.
    Offset 604 the number of value locations added to a base value in
    order to get to a specific absolute value.
    numOflevels the number of value levels that the device can provide
    605 in between maximum and minimum value.
    Sensitivity the minimum magnitude of input signal required to
    606 produce a specified output signal.
    SNR 607 the ratio of a signal power to the noise power
    corrupting the signal
    Accuracy the degree of closeness of a measured quantity to its
    608 actual value
    Position the position of the device from the user’s perspective
    according to the x-, y-, and z-axis
  • The source 630 may be a program representing the sensor capability base attributes using the XML.
  • A reference numeral 631 of the source 630 defines the maximum value 602 using the XML. According to the reference numeral 631, the maximum value 602 has “float” type data and is optionally used.
  • A reference numeral 632 of the source 630 defines the minimum value 603 using the XML. According to the reference numeral 632, the minimum value 603 has “float” type data and is optionally used.
  • A reference numeral 633 of the source 630 defines the number of levels 605 using the XML. According to the reference numeral 633, the number of levels 605 has “onNegativeInteger” type data and is optionally used.
  • However, the source 630 shown in FIG. 14 is not limiting but only example embodiments.
  • Referring to FIG. 10 again, the processing unit 320 may determine a first value received from the sensor based on the sensor capability, and transmit a second value corresponding to the first value to the virtual world.
  • Depending on embodiments, the processing unit 320 may transmit the second value to the virtual world when the first value received from the sensor is less than or equal to a maximum value measurable by the sensor and greater than or equal to a minimum value measurable by the sensor.
  • Depending on embodiments, when the first value received from the sensor is greater than the maximum value, the processing unit 320 may consider the first value as the maximum value and transmit the second value to the virtual world. Also, when the first value is less than the minimum value, the processing unit 320 may consider the first value as the minimum value and transmit the second value to the virtual world.
  • The virtual world processing apparatus 300 may further include a second storage unit (not shown) configured to store a sensor adaptation preference for manipulation of the first value received from the sensor. The processing unit 320 may generate a third value from the first value based on the sensor capability, and generate the second value from the third value based on the sensor adaptation preference.
  • Depending on embodiments, information on the motion, state, intention, shape, and the like of the user of the real world, which are measured through the sensor, may be directly reflected to the virtual world.
  • Hereinafter, the sensor capability will be described in relation to specific embodiments of the sensor. Although not limited to those embodiments, the sensor may include a position sensor, an orientation sensor, an acceleration sensor, a light sensor, a sound sensor, a temperature sensor, a humidity sensor, a distance sensor, a motion sensor, an intelligent camera sensor, an ambient noise sensor, an atmospheric sensor, a velocity sensor, an angular velocity sensor, an angular acceleration sensor, a force sensor, a torque sensor, and a pressure sensor.
  • Source 1 denotes a sensor capability related to the position sensor using the XML. However, a program source shown in Source 1 is only an example embodiment but does not limit the present invention.
  • SOURCE 1
     <!-- ################################################ -->
     <!-- Position Sensor capability type -->
     <!-- ################################################-->
     <complexType name=″PositionSensorCapabilityType″>
      <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <sequence>
         <element name=″range″
    type=″cid:RangeType″/>
        </sequence>
       </extension>
     </complexContent>
     </complexType>
     <complexType name=″RangeType″>
      <sequence>
       <element name=″XminValue″ type=″float″>
       <element name=″XmaxValue″ type=″float″>
       <element name=″YminValue″ type=″float″>
       <element name=″YmaxValue″ type=″float″>
       <element name=″ZminValue″ type=″float″>
       <element name=″ZmaxValue″ type=″float″>
      </sequence>
     </complexType>
  • A position sensor capability type is a tool for describing the sensor capability related to the position sensor.
  • The position sensor capability type may include sensor capability base attributes related to the position sensor.
  • The sensor capability base attributes related to the position sensor may include a range, a range type, an x maximum value, an x minimum value, a y maximum value, a y minimum value, a z maximum value, and a z minimum value.
  • The range denotes a range measurable by the position sensor. For example, the measurable range of the position sensor may be expressed using the range type and a global coordinate system.
  • An origin of the global coordinate may be located at a top left corner. A right handed coordinate system may be applied as the global coordinate. In the global coordinate, a positive direction of an x-axis may be a direction to a top right corner of a screen, a positive direction of a y-axis may be a gravity direction, that is, a bottomward direction of the screen, and a positive direction of a z-axis may be a direction opposite to the user, that is, a direction into the screen.
  • The range type denotes a range of the global coordinate system according to the x-axis, the y-axis, and the z-axis.
  • The x maximum value denotes a maximum value on the x-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • The x minimum value denotes a minimum value on the x-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • The y maximum value denotes a maximum value on the y-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • The y minimum value denotes a minimum value on the y-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • The z maximum value denotes a maximum value on the z-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • The z minimum value denotes a minimum value on the z-axis, measurable by the position sensor using a unit of a position coordinate, for example, meter.
  • Source 2 denotes a sensor capability related to an orientation sensor using the XML. However, a program source shown in Source 2 is only an example embodiment but does not limit the present invention.
  • SOURCE 2
     <!-- ################################################ -->
     <!-- Orientation Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″OrientationSensorCapabilityType″>
     <complexContent>
      <extension base=″cid:SensorCapabilityBaseType″>
       <sequence>
        <element           name=″orientationrange″
    type=″cid:OrientationRangeType″/>
       </sequence>
      </extension>
     </complexContent>
     </complexType>
     <complexType name=″OrientationRangeType″>
     <sequence>
      <element name=″XMinRotation″ type=″float″>
      <element name=″XMaxRotation″ type=″float″>
      <element name=″YMinRotation″ type=″float″>
      <element name=″YMaxRotation″ type=″float″>
      <element name=″ZMinRotation″ type=″float″>
      <element name=″ZMaxRotation″ type=″float″>
     </sequence>
     </complexType>
  • An orientation sensor capability type is a tool for describing the sensor capability related to the orientation sensor.
  • The orientation sensor capability type may include sensor capability base attributes related to the orientation sensor.
  • The sensor capability base attributes related to the orientation sensor may include an orientation range, an orientation range type, an x maximum value, an x minimum value, a y maximum value, a y minimum value, a z maximum value, and a z minimum value.
  • The range denotes a range measurable by the orientation sensor. For example, the measurable range of the orientation sensor may be expressed using the orientation range type and the global coordinate system.
  • The orientation range type denotes an orientation range of the global coordinate system according to the x-axis, the y-axis, and the z-axis.
  • The x maximum value denotes a maximum value on the x-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • The x minimum value denotes a minimum value on the x-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • The y maximum value denotes a maximum value on the y-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • The y minimum value denotes a minimum value on the y-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • The z maximum value denotes a maximum value on the z-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • The z minimum value denotes a minimum value on the z-axis, measurable by the orientation sensor using a unit of an orientation coordinate, for example, radian.
  • Source 3 denotes a sensor capability related to an acceleration sensor using the XML. However, a program source shown in Source 3 is only an example embodiment but does not limit the present invention.
  • SOURCE 3
    <!-- ################################################ -->
    <!-- Acceleration Sensor capability type -->
    <!-- ################################################ -->
    <complexType name=″AccelerationSensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • An acceleration sensor capability type is a tool for describing the sensor capability related to the acceleration sensor.
  • The acceleration sensor capability type may include sensor capability base attributes related to the acceleration sensor.
  • The sensor capability base attributes related to the acceleration sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the acceleration sensor using a unit of acceleration, for example, m/s2.
  • The minimum value denotes a minimum value measurable by the acceleration sensor using a unit of acceleration, for example, m/s2.
  • Source 4 denotes a sensor capability related to a light sensor using the XML. However, a program source shown in Source 4 is only an example embodiment but does not limit the present invention.
  • SOURCE 4
     <!-- ################################################ -->
     <!-- Light Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″LightSensorCapabilityType″>
     <complexContent>
      <extension base=″cid:SensorCapabilityBaseType″>
       <sequence>
        <element   name=″color″   type=″cid:colorType″
    minOccurs=″0″
     maxOccurs=″unbounded″/>
       </sequence>
       <attribute name=″location″ type=″mpeg7:termReferenceType″
    use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • A light sensor capability type is a tool for describing the sensor capability related to the light sensor.
  • The light sensor capability type may include sensor capability base attributes related to the light sensor.
  • The sensor capability base attributes related to the light sensor may include a maximum value, a minimum value, a color, and a location.
  • The maximum value denotes a maximum value measurable by the light sensor using a unit of light intensity, for example, LUX.
  • The minimum value denotes a minimum value measurable by the light sensor using a unit of light intensity, for example, LUX.
  • The color denotes a color that may be provided by the light sensor. For example, the color may be an RGB color value.
  • The location denotes a location of the light sensor. For example, the location of the light sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 5 denotes a sensor capability related to a sound sensor using the XML. However, a program source shown in Source 5 is only an example embodiment but does not limit the present invention.
  • SOURCE 5
    <!--######################## -->
    <!-- Sound Sensor capability type -->
    <!--######################## -->
    <complexType name=″SoundSensorCapabilityType″>
     <complexContent>
      <extension base=″sidc:CapabilityBaseType″/>
     </complexContent>
    </complexType>
  • A sound sensor capability type is a tool for describing the sensor capability related to the sound sensor.
  • The sound sensor capability type may include sensor capability base attributes related to the sound sensor.
  • The sensor capability base attributes related to the sound sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the sound sensor using a unit of sound intensity, for example, decibel (dB).
  • The minimum value denotes a minimum value measurable by the sound sensor using a unit of sound intensity, for example, dB.
  • Source 6 denotes a sensor capability related to a temperature sensor using the XML. However, a program source shown in Source 6 is only an example embodiment but does not limit the present invention.
  • SOURCE 6
     <!-- ################################################ -->
     <!-- Temperature Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″TemperatureSensorCapabilityType″>
     <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <attribute name=″location″ type=″mpeg7:term
    ReferenceType″use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • A temperature sensor capability type is a tool for describing the sensor capability related to the temperature sensor.
  • The temperature sensor capability type may include sensor capability base attributes related to the temperature sensor.
  • The sensor capability base attributes related to the temperature sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the temperature sensor using a unit of temperature, for example, ° C. and ° F.
  • The minimum value denotes a minimum value measurable by the temperature sensor using a unit of temperature, for example, ° C. and ° F.
  • The location denotes a location of the temperature sensor. For example, the location of the temperature sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 7 denotes a sensor capability related to a humidity sensor using the XML. However, a program source shown in Source 7 is only an example embodiment but does not limit the present invention.
  • SOURCE 7
     <!-- ################################################ -->
     <!-- Humidity Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″ HumiditySensorCapabilityType″>
     <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <attribute name=″location″ type=″mpeg7:term
    ReferenceType″use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • A humidity sensor capability type is a tool for describing the sensor capability related to the humidity sensor.
  • The humidity sensor capability type may include sensor capability base attributes related to the humidity sensor.
  • The humidity capability base attributes related to the humidity sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the humidity sensor using a unit of humidity, for example, percent (%).
  • The minimum value denotes a minimum value measurable by the humidity sensor using a unit of humidity, for example, %.
  • The location denotes a location of the humidity sensor. For example, the location of the humidity sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 8 denotes a sensor capability related to a distance sensor using the XML. However, a program source shown in Source 8 is only an example embodiment but does not limit the present invention.
  • SOURCE 8
     <!-- ################################################ -->
     <!-- Distance Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″LengthSensorCapabilityType″>
     <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <attribute name=″location″ type=″mpeg7:
    termReferenceType″use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • A distance sensor capability type is a tool for describing the sensor capability related to the distance sensor.
  • The distance sensor capability type may include sensor capability base attributes related to the distance sensor.
  • The sensor capability base attributes related to the distance sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the distance sensor using a unit of length, for example, meter.
  • The minimum value denotes a minimum value measurable by the distance sensor using a unit of length, for example, meter.
  • The location denotes a location of the distance sensor. For example, the location of the distance sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 9 denotes a sensor capability related to a motion sensor using the XML. However, a program source shown in Source 9 is only an example embodiment but does not limit the present invention.
  • SOURCE 9
     <!-- ################################################ -->
     <!-- Motion Sensor capability type -->
     <!-- ################################################ -->
     <complexType name=″MotionSensorCapabilityType″>
      <sequence>
       <element  name=″positioncapability″
    type=″cid:PositionSensorCapabilityType″ minOccurs=″0″/>
       <element name=″orientationcapability″
    type=″cid:OrientationSensorCapabilityType″ minOccurs=″0″/>
       <element  name=″velocitycapability″
    type=″cid:VelocitySensorCapabilityType″ minOccurs=″0″/>
       <element name=″angularvelocitycapability″
    type=″cid:AngularVelocitySensorCapabilityType″ minOccurs=″0″/>
       <element  name=″accelerationrange″
    type=″cid:AccelerationSensorCapabilityType″ minOccurs=″0″/>
       <element name=″angularaccelerationcapability″
    type=″cid:AngularAccelerationSensorCapabilityType″ minOccurs=″0″/>
      </sequence>
     </complexType>
  • A motion sensor capability type is a tool for describing the sensor capability related to the motion sensor.
  • The motion sensor may be an integrated sensor of a plurality of sensors. For example, the motion sensor may integrally include a position sensor, a velocity sensor, an acceleration sensor, an orientation sensor, an angular velocity sensor, and an angular acceleration sensor.
  • The motion sensor capability type may include sensor capability base attributes related to the motion sensor.
  • The sensor capability base attributes related to the motion sensor may include a position capability, a velocity capability, an acceleration capability, an orientation capability, an angular velocity capability, and an angular acceleration capability.
  • The position capability denotes capability with respect to the position.
  • The velocity capability denotes capability with respect to the velocity.
  • The acceleration capability denotes capability with respect to the acceleration.
  • The orientation capability denotes capability with respect to the orientation.
  • The angular velocity capability denotes capability with respect to the angular velocity.
  • The angular acceleration capability denotes capability with respect to the angular acceleration.
  • Source 10 denotes a sensor capability related to an intelligent camera sensor using the XML. However, a program source shown in Source 10 is only an example embodiment but does not limit the present invention.
  • SOURCE 10
      <!-- ################################################ -->
      <!-- Intelligent Camera CapabilityType  -->
      <!-- ################################################ -->
      <complexType name=″IntelligentCameraCapabilityType″>
      <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <sequence>
         <element name=″FeatureTrackingStatus″
    type=″boolean″ minOccurs=″0″/>
         <element
    name=″FacialExpressionTrackingStatus″ type=″boolean″ minOccurs=″0″/>
         <element name=″GestureTrackingStatus″
    type=″boolean″ minOccurs=″0″/>  <element
    name=″maxBodyFeaturePoint″ type=″float″ minOccurs=″0″/>
         <element name=″maxFaceFeaturePoint″
    type=″float″ minOccurs=″0″/>
         <element  name=″TrackedFeature″
    type=″cid:FeatureType″/>
         <element
    name=″TrackedFacialFeaturePoints″ type=″cid:FacialFeatureMask″/>
         <element name=″TrackedBodyFeaturePoints″
    type=″cid:BodyFeatureMask″/>
        </sequence>
       </extension>
      </complexContent>
     </complexType>
     <complexType name=″FeatureType″>
      <sequence>
       <element name=″Face″ type=″boolean″/>
       <element name=″Body″ type=″boolean″/>
       <element name=″Both″ type=″boolean″/>
      </sequence>
     </complexType>
     <complexType name=″FacialFeatureMask″>
      <sequence>
       <element   name=″FaceFeaturePoint″   type=″boolean″
    minOccurs=″60″ maxOccurs=″200″/>
       </sequence>
      </complexType>
      <complexType name=″BodyFeatureMask″>
       <sequence>
        <element   name=″BodyFeaturePoint″  type=″boolean″
    minOccurs=″60″ maxOccurs=″200″/>
       </sequence>
      </complexType>
  • An intelligent camera sensor capability type is a tool for describing the sensor capability related to the intelligent camera sensor.
  • The intelligent camera sensor capability type may include sensor capability base attributes related to the intelligent camera sensor.
  • The sensor capability base attributes related to the intelligent camera sensor may include a feature tracking status, an expression tracking status, a body movement tracking status, a maximum body feature point, a maximum face feature point, a tracked feature, tracked facial feature points, tracked body feature points, a feature type, a facial feature mask, and a body feature mask.
  • The feature tracking status denotes information on whether an intelligent camera is capable of tracking features.
  • The expression tracking status denotes information on whether the intelligent camera is capable of extracting animation related to a facial expression.
  • The body movement tracking status denotes information on whether the intelligent camera is capable of extracting animation related to a body.
  • The maximum body feature point denotes a maximum value of a body feature that can be tracked by the intelligent camera sensor.
  • The maximum face feature point denotes a maximum value of a face feature that can be tracked by the intelligent camera sensor.
  • The tracked feature denotes information on whether tracking of the body feature and the face feature is possible.
  • The tracked facial feature points denote information on whether the respective face features are activated or based on the facial feature mask.
  • The tracked body feature points denote information on whether the respective body features are activated or based on the body feature mask.
  • The feature type denotes a list of feature types. For example, the feature type may include 1. Face, 2. Body, and 3. Face and body.
  • The facial feature mask denotes a list of facial features.
  • The body feature mask denotes a list of body features.
  • Source 11 denotes a sensor capability related to an ambient noise sensor using the XML. However, a program source shown in Source 11 is only an example embodiment but does not limit the present invention.
  • SOURCE 11
     <!-- ################################################ -->
     <!-- Ambient noise Sensor capability type  -->
     <!-- ################################################-->
     <complexType name=″AmbientNoiseSensorCapabilityType″>
     <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <attribute name=″location″ type=″mpeg7:term
    ReferenceType″use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • An ambient noise sensor capability type is a tool for describing the sensor capability related to the ambient noise sensor.
  • The ambient noise sensor capability type may include sensor capability base attributes related to the ambient noise sensor.
  • The sensor capability base attributes related to the ambient noise sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the ambient noise sensor. For example, a unit for the ambient noise sensor may be dB.
  • The minimum value denotes a minimum value measurable by the ambient noise sensor. For example, a unit for the ambient noise sensor may be dB.
  • The location denotes a location of the ambient noise sensor. For example, the location of the ambient noise sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 12 denotes a sensor capability related to an atmospheric pressure sensor using the XML. However, a program source shown in Source 12 is only an example embodiment but does not limit the present invention.
  • SOURCE 12
     <!-- ################################################ -->
     <!-- Atmospheric Pressure Sensor capability type -
    ->
     <!-- ################################################ -->
     <complexType name=″AtmosphericPressureSensorCapabilityType″>
     <complexContent>
       <extension base=″cid:SensorCapabilityBaseType″>
        <attribute name=″location″ type=″mpeg7:term
    ReferenceType″use=″optional″/>
      </extension>
     </complexContent>
     </complexType>
  • An atmospheric pressure sensor capability type is a tool for describing the sensor capability related to the atmospheric pressure sensor.
  • The atmospheric pressure sensor capability type may include sensor capability base attributes related to the atmospheric pressure sensor.
  • The atmospheric pressure capability base attributes related to the atmospheric pressure sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the atmospheric pressure sensor using a unit of atmospheric pressure, for example, hectopascal (hPa).
  • The minimum value denotes a minimum value measurable by the atmospheric pressure sensor using a unit of atmospheric pressure, for example, hPa.
  • The location denotes a location of the atmospheric pressure sensor. For example, the location of the atmospheric pressure sensor may be expressed using the global coordinate according to the x-axis, the y-axis, and the z-axis.
  • Source 13 denotes a sensor capability related to a velocity sensor using the XML. However, a program source shown in Source 13 is only an example embodiment but does not limit the present invention.
  • SOURCE 13
    <!-- ################################################ -->
    <!-- Velocity Sensor capability type -->
    <!-- ################################################ -->
    <complexType name=″VelocitySensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • A velocity sensor capability type is a tool for describing the sensor capability related to the velocity sensor.
  • The velocity sensor capability type may include sensor capability base attributes related to the velocity sensor.
  • The velocity capability base attributes related to the velocity sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the velocity sensor using a unit of velocity, for example, m/s.
  • The minimum value denotes a minimum value measurable by the velocity sensor using a unit of velocity, for example, m/s.
  • Source 14 denotes a sensor capability related to an angular velocity sensor using the XML. However, a program source shown in Source 14 is only an example embodiment but does not limit the present invention.
  • SOURCE 14
    <!-- ################################################ -->
    <!-- Angular Velocity Sensor capability type -->
    <!-- ################################################ -->
    <complexType name=″AngularVelocitySensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • An angular velocity sensor capability type is a tool for describing the sensor capability related to the angular velocity sensor.
  • The angular velocity sensor capability type may include sensor capability base attributes related to the angular velocity sensor.
  • The angular velocity capability base attributes related to the angular velocity sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the angular velocity sensor using a unit of angular velocity, for example, radian/s.
  • The minimum value denotes a minimum value measurable by the angular velocity sensor using a unit of angular velocity, for example, radian/s.
  • Source 15 denotes a sensor capability related to an angular acceleration sensor using the XML. However, a program source shown in Source 15 is only an example embodiment but does not limit the present invention.
  • SOURCE 15
     <!-- ################################################ -->
     <!-- Angular Acceleration Sensor capability type --
    >
     <!-- ################################################ -->
     <complexType name=″AngularAccelerationSensorCapabilityType″>
     <complexContent>
      <extension base=″cid:SensorCapabilityBaseType″>
      </extension>
     </complexContent>
     </complexType>
  • An angular acceleration sensor capability type is a tool for describing the sensor capability related to the angular acceleration sensor.
  • The angular acceleration sensor capability type may include sensor capability base attributes related to the angular acceleration sensor.
  • The angular acceleration capability base attributes related to the angular acceleration sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the angular acceleration sensor using a unit of angular acceleration, for example, radian/s2.
  • The minimum value denotes a minimum value measurable by the angular acceleration sensor using a unit of angular acceleration, for example, radian/s2.
  • Source 16 denotes a sensor capability related to a force sensor using the XML. However, a program source shown in Source 16 is only an example embodiment but does not limit the present invention.
  • SOURCE 16
    <!-- ################################################ -->
    <!-- Force Sensor capability type -->
    <!-- ################################################ -->
    <complexType name=″ForceSensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • A force sensor capability type is a tool for describing the sensor capability related to the force sensor.
  • The force sensor capability type may include sensor capability base attributes related to the force sensor.
  • The force capability base attributes related to the force sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value measurable by the force sensor using a unit of force, for example, Newton (N).
  • The minimum value denotes a minimum value measurable by the force sensor using a unit of force, for example, N.
  • Source 17 denotes a sensor capability related to a torque sensor using the XML. However, a program source shown in Source 17 is only an example embodiment but does not limit the present invention.
  • SOURCE 17
    <!-- ################################################-->
    <!-- Torque Sensor capability type -->
    <!-- ################################################-->
    <complexType name=″ForceSensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • A torque sensor capability type is a tool for describing the sensor capability related to the torque sensor.
  • The torque sensor capability type may include sensor capability base attributes related to the torque sensor.
  • The torque capability base attributes related to the torque sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the torque sensor using a unit of torque, for example, Newton millimeter (N-mm).
  • The minimum value denotes a minimum value measurable by the torque sensor using a unit of torque, for example, N-mm.
  • Source 18 denotes a sensor capability related to a pressure sensor using the XML. However, a program source shown in Source 18 is only an example embodiment but does not limit the present invention.
  • SOURCE 18
    <!-- ################################################ -->
    <!-- Pressure Sensor capability type -->
    <!-- ################################################ -->
    <complexType name=″PressureSensorCapabilityType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
     </extension>
    </complexContent>
    </complexType>
  • A pressure sensor capability type is a tool for describing the sensor capability related to the pressure sensor.
  • The pressure sensor capability type may include sensor capability base attributes related to the pressure sensor.
  • The pressure capability base attributes related to the pressure sensor may include a maximum value, a minimum value, and a location.
  • The maximum value denotes a maximum value measurable by the pressure sensor using a unit of pressure, for example, m/s.
  • The minimum value denotes a minimum value measurable by the pressure sensor using a unit of pressure, for example, m/s.
  • Hereinafter, the sensor adaptation preference will be described in detail.
  • The sensor adaptation preference denotes information for manipulating a value received from a sensor.
  • A sensor adaptation preference base type denotes a base type of manipulation information of the user. Depending on embodiments, the sensor adaptation preference base type may be a base abstract type of the metadata related to a sensor adaptation preference commonly applied to all types of sensors, as part of metadata types related to the sensor adaptation preference.
  • Hereinafter, the sensor adaptation preference and the sensor adaptation preference base type will be described in detail with reference to FIGS. 15 through 17.
  • FIG. 15 illustrates a sensor adaptation preference base type 700 according to example embodiments.
  • Referring to FIG. 15, the sensor adaptation preference base type 700 may include sensor adaptation preference base attributes 710 and any attributes 720.
  • The sensor adaptation preference base attributes 710 denote a group of sensor adaptation preferences basically included in the sensor adaptation preference base type 700.
  • The any attributes 720 denote a group of additional sensor adaptation preferences. The any attributes 720 may be unique additional sensor capabilities which are applicable to an arbitrary sensor. The any attributes 420 may allow scalability for inclusion of attributes other than the base attributes.
  • FIG. 16 illustrates syntax 800 of a sensor adaptation preference base type according to example embodiments.
  • Referring to FIG. 16, the syntax of the sensor adaptation preference base type may include a diagram 810, attributes 820, and a source 830.
  • The diagram 810 may include a diagram of the sensor adaptation preference base type.
  • The attributes 820 may include sensor adaptation preference base attributes and any attributes.
  • The source 830 may be a program representing the sensor adaptation preference base type using an XML. However, the source 830 shown in FIG. 16 is suggested by way of example but not limiting.
  • FIG. 17 illustrates syntax 900 of sensor adaptation preference base attributes according to example embodiments.
  • Referring to FIG. 17, the syntax 900 of the sensor adaptation preference base attributes may include a diagram 910, attributes 920, and a source 930.
  • The diagram 910 may include a diagram of the sensor adaptation preference base attributes.
  • The attributes 920 may include a sensor reference identifier (ID) 901, a sensor adaptation mode 902, an activation state 903, a unit 904, a maximum value 905, a minimum value 906, and a number of levels 907.
  • The sensor reference ID 901 denotes information referencing an ID of an individual sensor that generates specific sensed information.
  • The sensor adaptation mode 902 denotes user preference information related to a method of adapting a sensor. Depending on embodiments, the sensor adaptation mode 902 may be a sensor adaptation preference related to an adaptation method that refines information on a motion, state, intention, shape, and the like of a user of a real world, measured through the sensor, and reflects the information to a virtual world. For example, a ‘strict’ value may denote a user preference that directly applies sensed information of the real world to the virtual world. A ‘scalable’ value may denote a user preference that varies the sensed information of the real world according to the user preference and applies the sensed information to the virtual world.
  • The activation state information 903 denotes information on whether to activate the sensor in the virtual world. Depending on embodiments, the activation state information 903 may be a sensor adaptation preference that determines whether the sensor is in operation.
  • The unit 904 denotes a unit of a value used in the virtual world. For example, the unit 904 may be a pixel. Also, the unit 904 may be a unit of a value corresponding to the value received from the sensor.
  • The maximum value 905 and the minimum value 906 denote a maximum value and a minimum value used in the virtual world, respectively. Depending on embodiments, the maximum value 905 and the minimum value 906 may be the unit of the value corresponding to the value received from the sensor.
  • The number of levels 907 denotes a number of levels used in the virtual world. Depending on embodiments, the number of levels 907 may be a value for dividing levels between the maximum value and the minimum used in the virtual world.
  • The sensor reference ID 901, the adaptation mode 902, the activation state 903, the unit 904, the maximum value 905, the minimum value 906, and the number of levels 907, as the sensor adaptation preference base attributes, may be rearranged as shown in Table 4.
  • TABLE 4
    Name Definition
    SensorIdRef Refers the Id of an individual sensor that has generated
    901 the specific sensed information
    Sensor the user’s preference on the adaptation method for the
    adaptation virtual world effect
    mode
    902
    Activate whether the effect shall be activated. a value of true
    903 means the effect shall be activated and false means the
    effect shall be deactivated
    Unit 904 the unit of value
    maxValue the maximum desirable value of the effect in percentage
    905 according to the max scale defined within the semantics
    definition of the individual effects
    minValue the minimum desirable value of the effect in percentage
    906 according to the min scale defined within the semantics
    definition of the individual effects
    numOflevels the number of value levels that the device can provide in
    907 between maximum and minimum value
  • The source 930 may be a program representing the sensor adaptation preference base attributes using the XML.
  • A reference numeral 931 defines the activation state 903 using the XML. According to the reference numeral 931, the activation state 903 has “boolean” type data and is optionally used.
  • A reference numeral 932 defines the maximum value 905 using the XML. According to the reference numeral 932, the maximum value 905 has “float” type data and is optionally used.
  • A reference number 933 defines minimum value 906 using the XML. According to the reference numeral 933, the minimum value 906 has “float” type data and is optionally used.
  • A reference numeral 934 defines the number of levels 907 using the XML. According to the reference numeral 934, the number of levels 907 has “onNegativeInteger” type data and is optionally used.
  • However, the source 930 illustrated in FIG. 17 is not limiting but only example embodiments.
  • Hereinafter, the sensor adaptation preference will be described in relation to specific embodiments of the sensor.
  • Source 19 denotes a sensor adaptation preference related to a position sensor using the XML. However, a program source shown in Source 19 is only an example embodiment but does not limit the present invention.
  • SOURCE 19
    <!-- ################################################ -->
    <!—Position Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=″PositionSensorPrefType″>
    <complexContent>
     <extension base=″cid:SensorCapabilityBaseType″>
      <sequence>
       <element name=″range″ type=″cid:RangeType″/>
      </sequence>
     </extension>
    </complexContent>
    </complexType>
  • A position sensor type is a tool for describing the sensor adaptation preference related to the position sensor.
  • A position sensor capability type may include sensor adaptation preference base attributes related to the position sensor.
  • The sensor adaptation preference base attributes related to the position sensor may include a range and a number of levels.
  • The range denotes a range of a user preference with respect to position information measured by the position sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the position information measured by the position sensor.
  • Source 20 denotes a sensor adaptation preference related to an orientation sensor using the XML. However, a program source shown in Source 20 is only an example embodiment but does not limit the present invention.
  • SOURCE 20
     <!-- ################################################ -->
     <!-- Orientation Sensor Preference type -->
     <!-- ################################################ -->
     <complexType name=″OrientationSensorPrefType″>
     <complexContent>
      <extension base=cid:SensorCapabilityBaseType/>
     <sequence>
            <element        name=″orientationrange″
    type=″cid:OrientationRangeType″/>
          </sequence>
     </extension>
     </complexContent>
     </complexType>
  • An orientation sensor type is a tool for describing the sensor adaptation preference related to the orientation sensor.
  • An orientation sensor capability type may include sensor adaptation preference base attributes related to the orientation sensor.
  • The sensor adaptation preference base attributes related to the orientation sensor may include an orientation range and a number of levels.
  • The orientation range denotes a range of a user preference with respect to orientation information measured by the orientation sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the orientation information measured by the orientation sensor.
  • Source 21 denotes a sensor adaptation preference related to an acceleration sensor using the XML. However, a program source shown in Source 21 is only an example embodiment but does not limit the present invention.
  • SOURCE 21
    <!-- ################################################ -->
    <!-- Acceleration Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=″AccelerationSensorPrefType″>
    <complexContent>
     <extension base=cid:SensorCapabilityBaseType/>
    </complexContent>
    </complexType>
  • An acceleration sensor type is a tool for describing the sensor adaptation preference related to the acceleration sensor.
  • An acceleration sensor capability type may include sensor adaptation preference base attributes related to the acceleration sensor.
  • The sensor adaptation preference base attributes related to the acceleration sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference related to acceleration information measured by the acceleration sensor.
  • The minimum value denotes a minimum value of the user preference related to the acceleration information measured by the acceleration sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the acceleration information measured by the acceleration sensor.
  • Source 22 denotes a sensor adaptation preference related to a light sensor using the XML. However, a program source shown in Source 22 is only an example embodiment but does not limit the present invention.
  • SOURCE 22
     <!-- ################################################ -->
     <!-- Light Sensor Preference type  -->
     <!-- ################################################ -->
     <complexType name=″LightSensorPrefType″>
     <complexContent>
      <extension base=″cid:UserSensorPreferenceBaseType″>
       <sequence>
        <element name=″color″ type=″cid:colorType″
    minOccurs=″0″
     maxOccurs=″unbounded″/>
       </sequence>
      </extension>
     </complexContent>
     </complexType>
  • A light sensor type is a tool for describing the sensor adaptation preference related to the light sensor.
  • A light sensor capability type may include sensor adaptation preference base attributes related to the light sensor.
  • The sensor adaptation preference base attributes related to the light sensor may include a maximum value, a minimum value, a number of levels, and an unfavorable color.
  • The maximum value denotes a maximum value of a user preference related to a value measured by the light sensor.
  • The minimum value denotes a minimum value of the user preference related to a value measured by the light sensor.
  • The number of levels denotes a number of levels of the user preference with respect to a value measured by the light sensor.
  • The unfavorable color denotes a list of unfavorable colors of the user, as RGB color values or a classification reference.
  • Source 23 denotes a sensor adaptation preference related to a sound sensor using the XML. However, a program source shown in Source 23 is only an example embodiment but does not limit the present invention.
  • SOURCE 23
    <!--######################## -->
    <!-- USIPV Sound Sensor type -->
    <!--######################## -->
    <complexType name=″SoundSensorType″>
     <complexContent>
      <extension base=″usip:PreferenceBaseType″/>
     </complexContent>
    </complexType>
  • A sound sensor type is a tool for describing the sensor adaptation preference related to the sound sensor.
  • A sound sensor capability type may include sensor adaptation preference base attributes related to the sound sensor.
  • The sensor adaptation preference base attributes related to the sound sensor may include a maximum value and a minimum value.
  • The maximum value denotes a maximum value allowed by the user as a measured value of the sound sensor.
  • The minimum value denotes a minimum value allowed by the user as a measured value of the sound sensor.
  • Source 24 denotes a sensor adaptation preference related to a temperature sensor using the XML. However, a program source shown in Source 24 is only an example embodiment but does not limit the present invention.
  • SOURCE 24
    <!-- ################################################ -->
    <!-- Temperature Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=″TemperatureSensorPrefType″>
    <complexContent>
     <extension base=″cid:UserSensorPreferenceBaseType″/>
    </complexContent>
    </complexType>
  • A temperature sensor type is a tool for describing the sensor adaptation preference related to the temperature sensor.
  • A temperature sensor capability type may include sensor adaptation preference base attributes related to the temperature sensor.
  • The sensor adaptation preference base attributes related to the temperature sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference related to temperature information measured by the temperature sensor.
  • The minimum value denotes a minimum value of the user preference related to the temperature information measured by the temperature sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the temperature information measured by the temperature sensor.
  • Source 25 denotes a sensor adaptation preference related to a humidity sensor using the XML. However, a program source shown in Source 25 is only an example embodiment but does not limit the present invention.
  • SOURCE 25
    <!-- ################################################ -->
    <!-- Humidity Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=″HumiditySensorPrefType″>
    <complexContent>
     <extension base=″cid:UserSensorPreferenceBaseType″/>
    </complexContent>
    </complexType>
  • A humidity sensor type is a tool for describing the sensor adaptation preference related to the humidity sensor.
  • A humidity sensor capability type may include sensor adaptation preference base attributes related to the humidity sensor.
  • The sensor adaptation preference base attributes related to the humidity sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference related to humidity information measured by the humidity sensor.
  • The minimum value denotes a minimum value of the user preference related to the humidity information measured by the humidity sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the humidity information measured by the humidity sensor.
  • Source 26 denotes a sensor adaptation preference related to a distance sensor using the XML. However, a program source shown in Source 26 is only an example embodiment but does not limit the present invention.
  • SOURCE 26
    <!-- ################################################ -->
    <!-- Distance sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=″LengthSensorPrefType″>
    <complexContent>
     <extension base=″cid:UserSensorPreferenceBaseType″/>
    </complexContent>
    </complexType>
  • A distance sensor type is a tool for describing the sensor adaptation preference related to the distance sensor.
  • A distance sensor capability type may include sensor adaptation preference base attributes related to the distance sensor.
  • The sensor adaptation preference base attributes related to the distance sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference related to length information measured by the distance sensor.
  • The minimum value denotes a minimum value of the user preference related to the length information measured by the distance sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the length information measured by the distance sensor.
  • Source 27 denotes a sensor adaptation preference related to a motion sensor using the XML. However, a program source shown in Source 27 is only an example embodiment but does not limit the present invention.
  • SOURCE 27
    <!-- ################################################ -->
    <!—  Motion Sensor Preference type         -->
    <!-- ################################################ -->
       <complexType name=″MotionSensorPrefType″>
         <sequence>
     <element   name=″positionpreference″
    type=″cid:PositionSensorPrefType″ minOccurs=″0″/>
     <element  name=″orientationpreference″
    type=″cid:OrientationSensorPrefType″ minOccurs=″0″/>
     <element   name=″velocitypreference″
    type=″cid:VelocitySensorPrefType″ minOccurs=″0″/>
     <element  name=″angularvelocitypreference″
    type=″cid:AngularVelocitySensorPrefType″ minOccurs=″0″/>
     <element   name=″accelerationpreference″
    type=″cid:AccelerationSensorPrefType″ minOccurs=″0″/>
     <element name=″angularaccelerationpreference″
    type=″cid:AngularAccelerationSensorPrefType″ minOccurs=″0″/>
    </sequence>
    </complexType>
  • A motion sensor capability type is a tool for describing the sensor adaptation preference related to the motion sensor.
  • The motion sensor capability type may include sensor adaptation preference base attributes related to the motion sensor.
  • The sensor adaptation preference base attributes related to the motion sensor may include a position preference, a velocity preference, an acceleration preference, an orientation preference, an angular velocity preference, and an angular acceleration preference.
  • The position preference denotes a user preference with respect to the position.
  • The velocity preference denotes a user preference with respect to the velocity.
  • The acceleration preference denotes a user preference with respect to the acceleration.
  • The orientation preference denotes a user preference with respect to the orientation.
  • The angular velocity preference denotes a user preference with respect to the angular velocity.
  • The angular acceleration preference denotes a user preference with respect to the angular acceleration.
  • Source 28 denotes a sensor adaptation preference related to an intelligent camera sensor using the XML. However, a program source shown in Source 28 is only an example embodiment but does not limit the present invention.
  • SOURCE 28
     <!-- ################################################-->
     <!-- Intelligent Camera Preference Type
    -->
     <!-- ################################################-->
     <complexType name=″IntelligentCameraPreferenceType″>
      <complexContent>
       <extension
    base=″cid:SensorAdaptationPreferenceBaseType″>
        <sequence>
    <element name=″FaceFeatureTrackingOn″
    type=″boolean″ minOccurs=″0″/>
    <element name=″BodyFeatureTrackingOn″
    type=″boolean″ minOccurs=″0″/>
    <element
    name=″FacialExpressionTrackingOn″ type=″boolean″ minOccurs=″0″/>
    <element name=″GestureTrackingOn″
    type=″boolean″ minOccurs=″0″/>
    <element name=″FacialFeatureMask″
    type=″cid:FacialFeatureMaskType″/>
    <element name=″BodyFeatureMask″
    type=″cid:BodyFeatureMaskType″/>
        </sequence>
       </extension>
      </complexContent>
     </complexType>
     <complexType name=″FacialFeatureMaskType″>
      <sequence>
       <element name=″Eyes″ type=″boolean″/>
     <element name=″Mouth″ type=″boolean″/>
     <element name=″Nose″ type=″boolean″/>
     <element name=″Ears″ type=″boolean″/>
      </sequence>
     </complexType>
     <complexType name=″BodyFeatureMaskType″>
      <sequence>
       <element name=″Head″ type=″boolean″/>
       <element name=″Arms″ type=″boolean″/>
     <element name=″Hands″ type=″boolean″/>
       <element name=″Legs″ type=″boolean″/>
       <element name=″Feet″ type=″boolean″/>
     <element name=″MiddleBody″ type=″boolean″/>
      </sequence>
     </complexType>
  • An intelligent camera sensor capability type is a tool for describing the sensor adaptation preference related to the intelligent camera sensor.
  • The intelligent camera sensor capability type may include sensor adaptation preference base attributes related to the intelligent camera sensor.
  • The sensor adaptation preference base attributes related to the intelligent camera sensor may include a face feature tracking on, a body feature tracking on, a facial expression tracking on, a gesture tracking on, a face tracking map, and a body tracking map.
  • The face feature tracking on denotes information on whether to activate a face feature tracking mode in which an intelligent camera sensor tracks features on a face of the user.
  • The body feature tracking on denotes information on whether to activate a body feature tracking mode in which the intelligent camera sensor tracks features on a body of the user.
  • The facial expression tracking on denotes information on user preference with respect to tracking of a facial expression of the user by the intelligent camera sensor.
  • The gesture tracking on denotes information on user preference with respect to tracking of a gesture of the user by the intelligent camera sensor.
  • The face tracking map provides a Boolean map related to a face tracking map type. The Boolean map provides face portions that the user wants to track. Depending on embodiments, the Boolean map according to the face tracking map type may provide eyes, a mouth, a noise, and ears as the face portions.
  • The body tracking map provides a Boolean map related to a body tracking map type. The Boolean map provides body portions that the user wants to track. Depending on embodiments, the Boolean map according to the body tracking map type may provide a head, arms, hands, legs, feet, and a middle body as the body portions.
  • Source 29 denotes a sensor adaptation preference related to an ambient noise sensor using the XML. However, a program source shown in Source 29 is only an example embodiment but does not limit the present invention.
  • [Source 29]
    <!-- ################################################ -->
    <!-- Ambient Noise Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=“AmbientNoiseSensorPrefType”>
    <complexContent>
     <extension base=“cid:UserSensorPreferenceBaseType”/>
    </complexContent>
    </complexType>
  • An ambient noise sensor type is a tool for describing the sensor adaptation preference related to the ambient noise sensor.
  • An ambient noise sensor capability type may include sensor adaptation preference base attributes related to the ambient noise sensor.
  • The sensor adaptation preference base attributes related to the ambient noise sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to ambient noise information measured by the ambient noise sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the ambient noise information measured by the ambient noise sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the ambient noise information measured by the ambient noise sensor.
  • Source 30 denotes a sensor adaptation preference related to an atmospheric pressure sensor using the XML. However, a program source shown in Source 30 is only an example embodiment but does not limit the present invention.
  • [Source 30]
     <!-- ################################################ -->
     <!--  Atmospheric  Pressure  Sensor  Preference  type
    -->
     <!-- ################################################ -->
     <complexType name=“AtmosphericPressureSensorPrefType”>
     <complexContent>
      <extension base=“cid:UserSensorPreferenceBaseType”/>
     </complexContent>
     </complexType>
  • An atmospheric pressure sensor type is a tool for describing the sensor adaptation preference related to the atmospheric pressure sensor.
  • An atmospheric pressure sensor capability type may include sensor adaptation preference base attributes related to the atmospheric pressure sensor.
  • The sensor adaptation preference base attributes related to the atmospheric pressure sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to atmospheric pressure information measured by the atmospheric pressure sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the atmospheric pressure information measured by the atmospheric pressure sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the atmospheric pressure information measured by the atmospheric pressure sensor.
  • Source 31 denotes a sensor adaptation preference related to a velocity sensor using the XML. However, a program source shown in Source 31 is only an example embodiment but does not limit the present invention.
  • [Source 31]
    <!-- ################################################ -->
    <!-- Velocity Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=“VelocitySensorPrefType”>
    <complexContent>
     <extension base=cid:SensorCapabilityBaseType/>
    </complexContent>
    </complexType>
  • A velocity sensor type is a tool for describing the sensor adaptation preference related to the velocity sensor.
  • A velocity sensor capability type may include sensor adaptation preference base attributes related to the velocity sensor.
  • The sensor adaptation preference base attributes related to the velocity sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to velocity information measured by the velocity sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the velocity information measured by the velocity sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the velocity information measured by the velocity sensor.
  • Source 32 denotes a sensor adaptation preference related to an angular velocity sensor using the XML. However, a program source shown in Source 32 is only an example embodiment but does not limit the present invention.
  • [Source 32]
    <!-- ################################################ -->
    <!—Angular Velocity Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=“AngularVelocitySensorPrefType”>
    <complexContent>
     <extension base=cid:SensorCapabilityBaseType/>
    </complexContent>
    </complexType>
  • An angular velocity sensor type is a tool for describing the sensor adaptation preference related to the angular velocity sensor.
  • An angular velocity sensor capability type may include sensor adaptation preference base attributes related to the angular velocity sensor.
  • The sensor adaptation preference base attributes related to the angular velocity sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to angular velocity information measured by the angular velocity sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the angular velocity information measured by the angular velocity sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the angular velocity information measured by the angular velocity sensor.
  • Source 33 denotes a sensor adaptation preference related to an angular acceleration sensor using the XML. However, a program source shown in Source 33 is only an example embodiment but does not limit the present invention.
  • [Source 33]
    <!-- ################################################ -->
    <!-- Angular Acceleration Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=“AngularAccelerationSensorPrefType”>
    <complexContent>
     <extension base=cid:SensorCapabilityBaseType/>
    </complexContent>
    </complexType>
  • An angular acceleration sensor type is a tool for describing the sensor adaptation preference related to the angular acceleration sensor.
  • An angular acceleration sensor capability type may include sensor adaptation preference base attributes related to the angular acceleration sensor.
  • The sensor adaptation preference base attributes related to the angular acceleration sensor may include a maximum value, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to angular acceleration information measured by the angular acceleration sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the angular acceleration information measured by the angular acceleration sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the angular acceleration information measured by the angular acceleration sensor.
  • Source 34 denotes a sensor adaptation preference related to a force sensor using the XML. However, a program source shown in Source 34 is only an example embodiment but does not limit the present invention.
  • [Source 34]
    <!-- ################################################ -->
    <!-- Force Sensor Preference type Preference type -->
    <!-- ################################################ -->
    <complexType name=“ForceSensorPrefType”>
    <complexContent>
     <extension base=“cid:SensorCapabilityBaseType”>
     </extension>
    </complexContent>
    </complexType>
  • A force sensor type is a tool for describing the sensor adaptation preference related to the force sensor.
  • A force sensor capability type may include sensor adaptation preference base attributes related to the force sensor.
  • The sensor adaptation preference base attributes related to the force sensor may include a maximum value and, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to force information measured by the force sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the force information measured by the force sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the force information measured by the force sensor.
  • Source 35 denotes a sensor adaptation preference related to a torque sensor using the XML. However, a program source shown in Source 35 is only an example embodiment but does not limit the present invention.
  • [Source 35]
    <!-- ################################################ -->
    <!-- Torque Sensor Preference type Preference type -->
    <!-- ################################################ -->
    <complexType name=“ForceSensorPrefType”>
    <complexContent>
     <extension base=“cid:SensorCapabilityBaseType”>
     </extension>
    </complexContent>
    </complexType>
  • A torque sensor type is a tool for describing the sensor adaptation preference related to the torque sensor.
  • A torque sensor capability type may include sensor adaptation preference base attributes related to the torque sensor.
  • The sensor adaptation preference base attributes related to the torque sensor may include a maximum value and, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to torque information measured by the torque sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the torque information measured by the torque sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the torque information measured by the torque sensor.
  • Source 36 denotes a sensor adaptation preference related to a pressure sensor using the XML. However, a program source shown in Source 36 is only an example embodiment but does not limit the present invention.
  • [Source 36]
    <!-- ################################################ -->
    <!-- Pressure Sensor Preference type -->
    <!-- ################################################ -->
    <complexType name=“PressureSensorPrefType”>
    <complexContent>
     <extension base=“cid:SensorCapabilityBaseType”>
     </extension>
    </complexContent>
    </complexType>
  • A pressure sensor type is a tool for describing the sensor adaptation preference related to the pressure sensor.
  • A pressure sensor capability type may include sensor adaptation preference base attributes related to the pressure sensor.
  • The sensor adaptation preference base attributes related to the pressure sensor may include a maximum value and, a minimum value, and a number of levels.
  • The maximum value denotes a maximum value of a user preference with respect to pressure information measured by the pressure sensor.
  • The minimum value denotes a minimum value of the user preference with respect to the pressure information measured by the pressure sensor.
  • The number of levels denotes a number of levels of the user preference with respect to the pressure information measured by the pressure sensor.
  • The virtual world processing apparatus according to the example embodiments may include sensed information.
  • The sensed information denotes a command for controlling the sensor. Depending on embodiments, the sensed information may be a command for controlling the sensor so as to reflect the information on the motion, state, intention, shape, and the like of the user of the real world, measured through the sensor, to the virtual world.
  • Depending on embodiments, the sensed information may serve as a root element for sensed information metadata.
  • Hereinafter, the sensed information will be described in detail with reference to FIG. 18.
  • FIG. 18 illustrates a sensed information base type 1000 according to example embodiments.
  • Referring to FIG. 18, the sensed information base type 1000 may include sensed information base attributes 1010 and any attributes 1020.
  • The sensed information base type 1000 may be a topmost type of a base type that may inherit an individual piece of the sensed information.
  • The sensed information base attributes 1010 denote a group of attributes for the commands.
  • The any attributes 1020 denote a group of additional sensed information. The any attributes 1020 may be unique additional sensed information applicable to an arbitrary sensor. The any attributes 1020 may allow scalability for inclusion of attributes other than the base attributes.
  • Source 37 may be a program denoting a sensed information base type using the XML. However, Source 37 is only an example embodiment but does not limit the present invention.
  • [Source 37]
    <!-- ################################################ -->
     <!-- Sensed information base type -->
     <!-- ################################################ -->
     <complexType name=“SensedInfoBaseType” abstract=“true”>
      <sequence>
       <element name=“TimeStamp” type=“mpegvct:TimeStampType”
    use=“optional” />
      </sequence>
      <attributeGroup ref=“iidl:SensedInfoBaseAttributes”/>
     </complexType>
  • The sensed information base type 101 may include an ID 1011, a sensor reference ID 1012, a group ID 1013, a priority 1014, an activation state, and a linked list 1016.
  • The ID 1011 denotes ID information for recognizing individual identity of the sensor.
  • The sensor reference ID 1012 may be information referencing an ID of the sensor that generates information included in particular sensed information.
  • The group ID 1013 denotes ID information for recognizing individual identity of a multi-sensor structure to which a particular sensor belongs.
  • The priority 1014 denotes information on a priority related to sensed information of sensors in the same group, the sensors sharing the same point at time when the sensed information is applied. Depending on embodiments, a value 1 may represent a highest priority and larger values may indicate lower priorities.
  • The activation state 1015 denotes information for determining whether the sensor is in operation.
  • The linked list 1016 denotes information on link data for grouping a plurality of sensors. Depending on embodiments, the linked list 1016 may be information on a multi-sensor structure for grouping the sensors by a method including reference information on IDs of neighboring sensors.
  • Depending on embodiments, the sensed information base attributes 1010 may further include a value, a timestamp, and a life span.
  • The value denotes a measured value of the sensor. The value may be received from the sensor.
  • The timestamp denotes information on a sensing time of the sensor.
  • The life span denotes information on a valid period of a sensor command. Depending on embodiments, the life span may be a second unit.
  • The sensed information base attributes including the ID, the sensor ID reference, the activation state, the value, the linked list, the timestamp, and the lifespan may be rearranged as shown in Table 5.
  • TABLE 5
    Name Definition
    id
    1011 Individual identity of sensor
    sensorIdRef References a sensor that has generated the information
    1012 included in this specific sensed information.
    groupID Identifier for a group multi-sensor structure to which
    1013 this specific sensor belongs.
    Priority Describes the priority for sensed information with
    1014 respect to other sensed information in the same group
    of sensors sharing the same point in time when the
    sensed information become adapted. A value of one
    indicates the highest priority and larger values
    indicate lower priorities.
    Activate whether the effect shall be activated. a value of true
    1015 means the effect shall be activated and false means the
    effect shall be deactivated.
    Value the value of the effect in percentage according to the
    max scale defined within the semantics definition of
    the individual effects.
    Linked grouping sensor structure that consists of a group of
    list sensors such that in each record there is a field that
    contains a reference (id) to the next sensor.
    Time stamp information on a sensing time of the sensor
    Life span information on a valid period of a sensor command
    (expressed with reference to the timestamp in units
    of second)
  • Source 37-2 may be a program representing the sensed information base attributes using the XML.
  • [Source 37-2]
    <!-- ################################################ -->
     <!-- Definition of Sensed information Base Attributes -->
     <!-- ################################################ -->
     <attributeGroup name=“SensedInfoBaseAttributes”>
      <attribute name=“id” type=“ID” use=“optional”/>
      <attribute name=“sensorIdRef” type=“anyURI” use=“optional”/>
      <attribute name=“linkedlist” type=“anyURI” use=“optional”/>
      <attribute name=“groupID” type=“anyURI” use=“optional”/>
      <attribute name=“priority” type=“PositiveInteger” use=“optional”/>
      <attribute name=“activate” type=“boolean” use=“optional”/>
     </attributeGroup>
  • Hereinafter, the sensed information of the sensor will be described in relation to specific embodiments.
  • Source 38 denotes sensed information related to a position sensor using the XML. However, a program source shown in Source 38 is only an example embodiment but does not limit the present invention.
  • [Source 38]
     <!--#################################### -->
     <!--Definition of Position Sensor type -->
     <!--#################################### -->
     <complexType name=“PositionSensorType”>
     <complexContent>
      <extension base=“iidl:SensedInfoBaseType”>
       <sequence>
        <element name=“position”
        type=“mpegvct:Float3DVectorType”
    minOccurs=“0”/>
       </sequence>
       <attribute name=“timestamp” type=“float” use=“optional”/>
       <attribute name=“unit” type=“mpegvct:unitType” use=“optional”/>
      </extension>
     </complexContent>
     </complexType>
  • A position sensor type is a tool for describing sensed information related to the position sensor.
  • The position sensor type may include attributes such as a timestamp, a life span, a position, a position value type, Px, Py, and Pz.
  • The timestamp denotes information on a sensing time of the position sensor.
  • The life span denotes information on a valid period of a command of the position sensor. For example, the life span may be a second unit.
  • The position denotes information on a 3-dimensional (3D) value of the position sensor, expressed by a unit of distance, for example, meter.
  • The position value type denotes a tool for indicating a 3D position vector.
  • The Px denotes information on an x-axis value of the position sensor.
  • The Py denotes information on a y-axis value of the position sensor.
  • The Pz denotes information on a z-axis value of the position sensor.
  • Source 38-2 and Source 38-3 denote binary representation syntax corresponding to the sensed information related to the position sensor according to the example embodiments.
  • [Source 38-2]
    Number of bits Mnemonic
    PositionSensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    positionFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStamp
    Type
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(positionFlag) {
    position Float3DVectorType
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • [Source 38-3]
    Number of bits Mnemonic
    Float3DVectorType {
    X 32 fsbf
    Y 32 fsbf
    Z 32 fsbf
    }
  • Source 38-2 and Source 38-3 may include a binary encoding syntax the sensed information of the position sensor, the number of bits occupied by attributes of the binary encoding syntax, and a mnemonic of the attributes.
  • For example, the binary encoding syntax corresponding to the sensed information of the position sensor may include a priority denoting a priority of the position sensor. Here, the number of bits occupied by the priority may be 8 bits. A priority mnemonic may be uimsbf which stands for ‘unsigned integer most significant bit first.’
  • Depending on embodiments, the mnemonic may further include bslbf which stands for ‘bit string left bit first’ and Unicode transformation format-8 (UTF-8) which stands for a variable length letter encoding method for Unicode. UTF-8 may use 1 to 4 bytes to express one Unicode letter.
  • Table 1-0 shows a mnemonic of data.
  • TABLE 1-0
    bslbf Bit string, left bit first, where “left” is the order in which bits are written
    in ISO/IEC 15938-3. Bit strings are generally written as a string of 1s
    and 0s within single quote marks, e.g. ‘1000 0001’. Blanks within a
    bit string are for ease of reading and have no significance. For
    convenience, large strings are occasionally written in hexadecimal, in
    which case conversion to a binary in the conventional manner will yield
    the value of the bit string. Thus, the left-most hexadecimal digit is first
    and in each hexadecimal digit the most significant of the four digits is
    first.
    UTF-8 Binary string encoding defined in ISO 10646/IETF RFC 2279.
    vluimsbf5 Variable length unsigned integer most significant bit first
    representation consisting of two parts. The first part defines the
    number n of 4-bit bit fields used for the value representation, encoded
    by a sequence of n-1 “1” bits, followed by a “0” bit signaling its end.
    The second part contains the value of the integer encoded using the
    number of bit fields specified in the first part.
    uimsbf Unsigned integer, most significant bit first.
    fsbf Float (32 bit), sign bit first. The semantics of the bits within a float are
    specified in the IEEE Standard for Binary Floating Point Arithmetic
    (ANSI/IEEE Std 754-1985).
  • Table 1-1 shows descriptor components semantics of the position sensor according to example embodiments.
  • TABLE 1-1
    Names Description
    PositionSensorType Tool for describing sensed information with
    respect to a position sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor
    ID reference attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    positionFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time
    stamp type shall be used, “01” means that the clock
    tick time stamp type shall be used, and “10” means
    that the clock tick time delta stamp type shall be
    used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time, is
    defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to a position sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the ID
    of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this position sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is zero.
    If there is more than one sensed information with
    the same priority, the order of process can be
    determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    position Describes the sensed value of the position sensor
    in 3D with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in
    the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference to
    a classification scheme term provided by UnitCS
    defined in xxx of ISO/IEC 23005-6 and use the
    binary representation defined above.
  • Source 39 denotes sensed information related to an orientation sensor using the XML. However, a program source shown in Source 39 is only an example embodiment but does not limit the present invention.
  • [Source 39]
    <!--#################################### -->
     <!--Definition of Orientation Sensor type -->
     <!--#################################### -->
     <complexType name=“OrientationSensorType”>
      <complexContent>
       <extension base=“iidl:SensedInfoBaseType”>
        <sequence>
         <element name=“orientation”
         type=“mpegvct:Float3DVectorType”
    minOccurs=“0”/>
        </sequence>
        <attribute name=“timestamp” type=“float” use=“optional”/>
        <attribute name=“unit” type=“mpegvct:unitType”
        use=“optional”/>
        </extension>
       </complexContent>
      </complexType>
  • An orientation sensor type is a tool for describing sensed information related to the orientation sensor.
  • The orientation sensor type may include attributes such as a timestamp, a life span, an orientation, an orientation value type, Ox, Oy, and Oz.
  • The timestamp denotes information on a sensing time of the orientation sensor.
  • The life span denotes information on a valid period of a command of the orientation sensor. For example, the life span may be a second unit.
  • The orientation denotes information on a 3D value of the orientation sensor, expressed by a unit of orientation, for example, radian.
  • The orientation value type denotes a tool for indicating a 3D orientation vector.
  • The Ox denotes information on a value of an x-axis rotation angle of the orientation sensor.
  • The Oy denotes information on a value of a y-axis rotation angle of the orientation sensor.
  • The Oz denotes information on a value of a z-axis rotation angle of the orientation sensor.
  • Source 39-2 denotes a binary representation syntax corresponding to the sensed information related to the orientation sensor according the example embodiments.
  • [Source 39-2]
    Number of bits Mnemonic
    OrientationSensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    orientationFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(orientationFlag) {
    Orientation Float3DVectorType
    }
    if(unitFlag) {
    Unit unitType
    }
    }
  • Table 1-2 shows descriptor components semantics of the orientation sensor according to the example embodiments.
  • TABLE 1-2
    Names Description
    OrientationSensorType Tool for describing sensed information with
    respect to an orientation sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor
    ID reference attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    orientationFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time
    stamp type shall be used, “01” means that the clock
    tick time stamp type shall be used, and “10” means
    that the clock tick time delta stamp type shall be
    used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time, is
    defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to an orientation sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the ID
    of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this orientation sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is zero.
    If there is more than one sensed information with
    the same priority, the order of process can be
    determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    orientation Describes the sensed value of the orientation
    sensor in 3D with respect to the default unit if the
    unit is not defined. Otherwise, use the unit type
    defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference to
    a classification scheme term provided by UnitCS
    defined in xxx of ISO/IEC 23005-6 and use the
    binary representation defined above.
  • Source 40 denotes sensed information related to an acceleration sensor using the XML. However, a program source shown in Source 40 is only an example embodiment but does not limit the present invention.
  • [Source 40]
    <!--#################################### -->
     <!--Definition of Acceleration Sensor type -->
     <!--####################################-->
     <complexType name=“AccelerationSensorType”>
      <complexContent>
       <extension base=“iidl:SensedInfoBaseType”>
         <sequence>
          <element name=“acceleration”
          type=“mpegvct:Float3DVectorType”
    minOccurs=“0”/>
         </sequence>
         <attribute name=“timestamp” type=“float” use=“optional”/>
         <attribute name=“unit” type=“mpegvct:unitType”
         use=“optional”/>
       </extension>
     </complexContent>
        </complexType>
  • An acceleration sensor type is a tool for describing sensed information related to the acceleration sensor.
  • The acceleration sensor type may include attributes such as a timestamp, a life span, an acceleration, an acceleration value type, Ax, Ay, and Az.
  • The timestamp denotes information on a sensing time of the acceleration sensor.
  • The life span denotes information on a valid period of a command of the acceleration sensor. For example, the life span may be a second unit.
  • The acceleration denotes information on a value of the acceleration sensor, expressed by a unit of acceleration, for example, m/s2.
  • The acceleration value type denotes a tool for indicating a 3D acceleration vector.
  • The Ax denotes information on an x-axis value of the acceleration sensor.
  • The Ay denotes information on a y-axis value of the acceleration sensor.
  • The Az denotes information on a z-axis value of the acceleration sensor.
  • Source 40-2 denotes a binary representation syntax corresponding to the sensed information related to the acceleration sensor according the example embodiments.
  • [Source 40-2]
    Number of bits Mnemonic
    AccelerationSensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    acceleration Flag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStamp
    Type
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(accelerationFlag) {
    acceleration Float3DVectorType
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • Table 1-3 shows descriptor components semantics of the acceleration sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    AccelerationSensorType Tool for describing sensed information with respect
    to an acceleration sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the timestamp shall
    not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor
    ID reference attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    accelerationFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the user-
    defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time
    stamp type shall be used, “01” means that the clock
    tick time stamp type shall be used, and “10” means
    that the clock tick time delta stamp type shall be
    used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time, is
    defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to
    an acceleration sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the ID
    of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which
    this acceleration sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is zero.
    If there is more than one sensed information with
    the same priority, the order of process can be
    determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    acceleration Describes the sensed value of the acceleration
    sensor in 3D with respect to the default unit if the
    unit is not defined. Otherwise, use the unit type
    defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit other
    than the default unit is used, as a reference to a
    classification scheme term provided by UnitCS
    defined in xxx of ISO/IEC 23005-6 and use the
    binary representation defined above.
  • Source 41 denotes sensed information related to a light sensor using the XML. However, a program source shown in Source 41 is only an example embodiment but does not limit the present invention.
  • [Source 41]
    <!--#################################### -->
     <!--Definition of Light Sensor type -->
     <!--#################################### -->
     <complexType name=“LightSensorType”>
      <complexContent>
       <extension base=“iidl:SensedInfoBaseType”>
       <attribute name=“value” type=“float” use=“optional”/>
       <attribute name=“unit” type=“iidl:unitType” use=“optional”/>
       <attribute name=“color” type=“iidl:colorType”
    use=“optional”/>
       </extension>
      </complexContent>
     </complexType>
  • A light sensor type is a tool for describing sensed information related to the light sensor.
  • The light sensor type may include attributes such as a timestamp, a life span, a value, and a color.
  • The timestamp denotes information on a sensing time of the light sensor.
  • The life span denotes information on a valid period of a command of the light sensor. For example, the life span may be a second unit.
  • The value denotes information on a light sensor value expressed by a unit of light intensity, for example, LUX.
  • The color denotes a color that may be provided by the light sensor. For example, the color may be an RGB color value.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the light sensor according to the example embodiments.
  • [Source 41-2]
    LightSensorType{ Number of bits Mnemonic
       TimeStampFlag 1 bslbf
       IDFlag 1 bslbf
       sensorIdRefFlag 1 bslbf
       linkedlistFlag 1 bslbf
       groupIDFlag 1 bslbf
       priorityFlag 1 bslbf
       activateFlag 1 bslbf
       valueFlag 1 bslbf
       unitFlag 1 bslbf
       colorFlag 1 bslbf
       if(TimeStampFlag) {
        TimestampType 2 bslbf
        if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
        } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
        } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
        }
       }
       If(IDFlag) {
       ID See ISO 10646 UTF-8
       }
       if(sensorIdRefFlag) {
       sensorIdRefLength vluimsbf5
       sensorIdRef  8* bslbf
        sensorIdRefLength
       }
       if(linkedlistFlag) {
       linkedlistLength vluimsbf5
       linkedlist  8* bslbf
        linkedlistLength
       }
       if(groupIDFlag) {
       groupIDLength vluimsbf5
       groupID  8* bslbf
        groupIDLength
       }
       If(priorityFlag) {
         priority 8 uimsbf
       }
       if(activateFlag) {
       Activate 1 bslbf
       }
       if(valueFlag) {
       Value 32  fsbf
       }
       if(unitFlag) {
       Unit unitType
       }
       if(colorFlag) {
       Color colorType
       }
    }
  • Table 1-1-1 shows binary encoding of a unit CS corresponding to the light sensor according to the example embodiments. Table 1-1-1 may include a unit type and a term ID of unit.
  • TABLE 1-1-1
    unitType Term ID of unit
    00000000 micrometer
    00000001 mm
    00000010 cm
    00000011 meter
    00000100 km
    00000101 inch
    00000110 yard
    00000111 mile
    00001000 mg
    00001001 gram
    00001010 kg
    00001011 ton
    00001100 micrometerpersec
    00001101 mmpersec
    00001110 cmpersec
    00001111 meterpersec
    00010000 Kmpersec
    00010001 inchpersec
    00010010 yardpersec
    00010011 milepersec
    00010100 micrometerpermin
    00010101 mmpermin
    00010110 cmpermin
    00010111 meterpermin
    00011000 kmpermin
    00011001 inchpermin
    00011010 yardpermin
    00011011 milepermin
    00011100 micrometerperhour
    00011101 mmperhour
    00011110 cmperhour
    00011111 meterperhour
    00100000 kmperhour
    00100001 inchperhour
    00100010 yardperhour
    00100011 mileperhour
    00100100 micrometerpersecsquare
    00100101 mmpersecsquare
    00100110 cmpersecsquare
    00100111 meterpersecsquare
    00101000 kmpersecsquare
    00101001 inchpersecsquare
    00101010 yardpersecsquare
    00101011 milepersecsquare
    00101100 micormeterperminsquare
    00101101 mmperminsquare
    00101110 cmperminsquare
    00101111 meterperminsquare
    00110000 kmpersminsquare
    00110001 inchperminsquare
    00110010 yardperminsquare
    00110011 mileperminsquare
    00110100 micormeterperhoursquare
    00110101 mmperhoursquare
    00110110 cmperhoursquare
    00110111 meterperhoursquare
    00111000 kmperhoursquare
    00111001 inchperhoursquare
    00111010 yardperhoursquare
    00111011 mileperhoursquare
    00111100 Newton
    00111101 Nmm
    00111110 Npmm
    00111111 Hz
    01000000 KHz
    01000001 MHz
    01000010 GHz
    01000011 volt
    01000100 millivolt
    01000101 ampere
    01000110 milliampere
    01000111 milliwatt
    01001000 watt
    01001001 kilowatt
    01001010 lux
    01001011 celsius
    01001100 fahrenheit
    01001101 radian
    01001110 degree
    01001111 radpersec
    01010000 degpersec
    01010001 radpersecsquare
    01010010 degpersecsquare
    01010011 Npermmsquare
    01011100-11111111 Reserved
  • Table 1-1-1 shows binary encoding of a unit CS corresponding to the light sensor according to the example embodiments. Table 1-1-1 may include a unit type and a term ID of unit.
  • TABLE 1-1-1
    colorType Term ID of color
    000000000 alice_blue
    000000001 alizarin
    000000010 amaranth
    000000011 amaranth_pink
    000000100 amber
    000000101 amethyst
    000000110 apricot
    000000111 aqua
    000001000 aquamarine
    000001001 army_green
    000001010 asparagus
    000001011 atomic_tangerine
    000001100 auburn
    000001101 azure_color_wheel
    000001110 azure_web
    000001111 baby_blue
    000010000 beige
    000010001 bistre
    000010010 black
    000010011 blue
    000010100 blue_pigment
    000010101 blue_ryb
    000010110 blue_green
    000010111 blue-green
    000011000 blue-violet
    000011001 bondi_blue
    000011010 brass
    000011011 bright_green
    000011100 bright_pink
    000011101 bright_turquoise
    000011110 brilliant_rose
    000011111 brink_pink
    000100000 bronze
    000100001 brown
    000100010 buff
    000100011 burgundy
    000100100 burnt_orange
    000100101 burnt_sienna
    000100110 burnt_umber
    000100111 camouflage_green
    000101000 caput_mortuum
    000101001 cardinal
    000101010 carmine
    000101011 carmine_pink
    000101100 carnation_pink
    000101101 Carolina_blue
    000101110 carrot_orange
    000101111 celadon
    000110000 cerise
    000110001 cerise_pink
    000110010 cerulean
    000110011 cerulean_blue
    000110100 champagne
    000110101 charcoal
    000110110 chartreuse_traditional
    000110111 chartreuse_web
    000111000 cherry_blossom_pink
    000111001 chestnut
    000111010 chocolate
    000111011 cinnabar
    000111100 cinnamon
    000111101 cobalt
    000111110 Columbia_blue
    000111111 copper
    001000000 copper_rose
    001000001 coral
    001000010 coral_pink
    001000011 coral_red
    001000100 corn
    001000101 cornflower_blue
    001000110 cosmic_latte
    001000111 cream
    001001000 crimson
    001001001 cyan
    001001010 cyan_process
    001001011 dark_blue
    001001100 dark_brown
    001001101 dark_cerulean
    001001110 dark_chestnut
    001001111 dark_coral
    001010000 dark_goldenrod
    001010001 dark_green
    001010010 dark_khaki
    001010011 dark_magenta
    001010100 dark_pastel_green
    001010101 dark_pink
    001010110 dark_scarlet
    001010111 dark_salmon
    001011000 dark_slate_gray
    001011001 dark_spring_green
    001011010 dark_tan
    001011011 dark_turquoise
    001011100 dark_violet
    001011101 deep_carmine_pink
    001011110 deep_cerise
    001011111 deep_chestnut
    001100000 deep_fuchsia
    001100001 deep_lilac
    001100010 deep_magenta
    001100011 deep_magenta
    001100100 deep_peach
    001100101 deep_pink
    001100110 denim
    001100111 dodger_blue
    001101000 ecru
    001101001 egyptian_blue
    001101010 electric_blue
    001101011 electric_green
    001101100 elctric_indigo
    001101101 electric_lime
    001101110 electric_purple
    001101111 emerald
    001110000 eggplant
    001110001 falu_red
    001110010 fern_green
    001110011 firebrick
    001110100 flax
    001110101 forest_green
    001110110 french_rose
    001110111 fuchsia
    001111000 fuchsia_pink
    001111001 gamboge
    001111010 gold_metallic
    001111011 gold_web_golden
    001111100 golden_brown
    001111101 golden_yellow
    001111110 goldenrod
    001111111 grey-asparagus
    010000000 green_color_wheel_x11_green
    010000001 green_html/css_green
    010000010 green_pigment
    010000011 green_ryb
    010000100 green_yellow
    010000101 grey
    010000110 han_purple
    010000111 harlequin
    010001000 heliotrope
    010001001 Hollywood_cerise
    010001010 hot_magenta
    010001011 hot_pink
    010001100 indigo_dye
    010001101 international_klein_blue
    010001110 international_orange
    010001111 Islamic_green
    010010000 ivory
    010010001 jade
    010010010 kelly_green
    010010011 khaki
    010010100 khaki_x11_light_khaki
    010010101 lavender_floral
    010010110 lavender_web
    010010111 lavender_blue
    010011000 lavender_blush
    010011001 lavender_grey
    010011010 lavender_magenta
    010011011 lavender_pink
    010011100 lavender_purple
    010011101 lavender_rose
    010011110 lawn_green
    010011111 lemon
    010100000 lemon_chiffon
    010100001 light_blue
    010100010 light_pink
    010100011 lilac
    010100100 lime_color_wheel
    010100101 lime_web_x11_green
    010100110 lime_green
    010100111 linen
    010101000 magenta
    010101001 magenta_dye
    010101010 magenta_process
    010101011 magic_mint
    010101100 magnolia
    010101101 malachite
    010101110 maroon_html/css
    010101111 marron_x11
    010110000 maya_blue
    010110001 mauve
    010110010 mauve_taupe
    010110011 medium_blue
    010110100 medium_carmine
    010110101 medium_lavender_magenta
    010110110 medum_purple
    010110111 medium_spring_green
    010111000 midnight_blue
    010111001 midnight_green_eagle_green
    010111010 mint_green
    010111011 misty_rose
    010111100 moss_green
    010111101 mountbatten_pink
    010111110 mustard
    010111111 myrtle
    011000000 navajo_white
    011000001 navy_blue
    011000010 ochre
    011000011 office_green
    011000100 old_gold
    011000101 old_lace
    011000110 old_lavender
    011000111 old_rose
    011001000 olive
    011001001 olive_drab
    011001010 olivine
    011001011 orange_color_wheel
    011001100 orange_ryb
    011001101 orange_web
    011001110 orange_peel
    011001111 orange-red
    011010000 orchid
    011010001 pale_blue
    011010010 pale_brown
    011010011 pale_carmine
    011010100 pale_chestnut
    011010101 pale_cornflower_blue
    011010110 pale_magenta
    011010111 pale_pink
    011011000 pale_red-violet
    011011001 papaya_whip
    011011010 pastel_green
    011011011 pastel_pink
    011011100 peach
    011011101 peach-orange
    011011110 peach-yellow
    011011111 pear
    011100000 periwinkle
    011100001 persian_blue
    011100010 persian_green
    011100011 persian_indigo
    011100100 persian_orange
    011100101 persian_red
    011100110 persian_pink
    011100111 persian_rose
    011101000 persimmon
    011101001 pine_green
    011101010 pink
    011101011 pink-orange
    011101100 platinum
    011101101 plum_web
    011101110 powder_blue_web
    011101111 puce
    011110000 prussian_blue
    011110001 psychedelic_purple
    011110010 pumpkin
    011110011 purple_html/css
    011110100 purple_x11
    011110101 purple_taupe
    011110110 raw_umber
    011110111 razzmatazz
    011111000 red
    011111001 red_pigment
    011111010 red_ryb
    011111011 red-violet
    011111100 rich_carmine
    011111101 robin_egg_blue
    011111110 rose
    011111111 rose_madder
    100000000 rose_taupe
    100000001 royal_blue
    100000010 royal_purple
    100000011 ruby
    100000100 russet
    100000101 rust
    100000110 safety_orange_blaze_orange
    100000111 saffron
    100001000 salmon
    100001001 sandy_brown
    100001010 sangria
    100001011 sapphire
    100001100 scarlet
    100001101 school_bus_yellow
    100001110 sea_green
    100001111 seashell
    100010000 selective_yellow
    100010001 sepia
    100010010 shamrock_green
    100010011 shocking_pink
    100010100 silver
    100010101 sky_blue
    100010110 slate_grey
    100010111 smalt_dark_powder_blue
    100011000 spring_bud
    100011001 spring_green
    100011010 steel_blue
    100011011 tan
    100011100 tangerine
    100011101 tangerine_yellow
    100011110 taupe
    100011111 tea_green
    100100000 tea_rose_orange
    100100001 tea_rose_rose
    100100010 teal
    100100011 tenné_tawny
    100100100 terra_cotta
    100100101 thistle
    100100110 tomato
    100100111 turquoise
    100101000 tyrian_purple
    100101001 ultramarine
    100101010 ultra_pink
    100101011 united_nation_blue
    100101100 vegas_gold
    100101101 vermilion
    100101110 violet
    100101111 violet_web
    100110000 violet_ryb
    100110001 viridian
    100110010 wheat
    100110011 white
    100110100 wisteria
    100110101 yellow
    100110110 yellow_process
    100110111 yellow_ryb
    100111000 yellow_green
    100111001-111111111 Reserved
  • Table 1-1 shows descriptor components semantics of the light sensor according to the example embodiments.
  • TABLE 1-1
    Names Description
    LightSensorType Tool for describing sensed information with respect to a
    light sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be used
    and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    colorFlag This field, which is only present in the binary
    representation, signals the presence of color attribute. A
    value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme shall
    be used. “00” means that the absolute time stamp type
    shall be used, “01” means that the clock tick time stamp
    type shall be used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined in
    A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a light
    sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the information
    included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor structure
    that consists of a group of sensors in a way that each
    record contains a reference to the ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which this
    light sensor belongs.
    priority Describes a priority for sensed information with respect to
    other sensed information sharing the same point in time
    when the sensed information becomes adapted. A value
    of zero indicates the highest priority and larger values
    indicate lower priorities. The default value of the priority
    is zero. If there is more than one sensed information with
    the same priority, the order of process can be determined
    by the adaptation engine itself.
    activate Describes whether the sensor is activated. A value of “1”
    means the sensor is activated and “0” means the sensor is
    deactivated.
    value Describes the sensed value of the light sensor with
    respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other than
    the default unit is used, as a reference to a classification
    scheme term provided by UnitCS defined in xxx of
    ISO/IEC 23005-6 and use the binary representation
    defined above.
    color Describes the list of colors which the lighting device can
    sense as a reference to a classification scheme term or as
    RGB value. A CS that may be used for this purpose is
    the ColorCS defined in A.2.3 of ISO/IEC 23005-6 and use
    the binary representation defined above.
  • Source 42 denotes sensed information related to a sound sensor using the XML. However, a program source shown in Source 42 is only an example embodiment but does not limit the present invention.
  • [Source 42]
    <!--######################## -->
    <!-- SCmd Sound Sensor type -->
    <!--######################## -->
    <complexType name=“SoundSensorType”>
     <complexContent>
      <extension base=“cid: SCmdBaseType”/>
     </complexContent>
    </complexType>
  • A sound sensor command type is a tool for describing sensed information related to the sound sensor.
  • Source 43 denotes sensed information related to a temperature sensor using the XML. However, a program source shown in Source 43 is only an example embodiment but does not limit the present invention.
  • [Source 43]
    <!--#################################### -->
    <!--Definition of Temperature Sensor type -->
    <!--#################################### -->
    <complexType name=″TemperatureSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″iidl:unitType″ use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A temperature sensor type is a tool for describing sensed information related to the temperature sensor.
  • The temperature sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the temperature sensor.
  • The life span denotes information on a valid period of a command of the temperature sensor. For example, the life span may be a second unit.
  • The value denotes information on a temperature sensor value expressed by a unit of temperature, for example, ° C. and ° F.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the temperature sensor according to the example embodiments.
  • [Source 41-2]
    TemperatureSensorType{ Number of bits Mnemonic
       TimeStampFlag 1 bslbf
       IDFlag 1 bslbf
       sensorIdRefFlag 1 bslbf
       linkedlistFlag 1 bslbf
       groupIDFlag 1 bslbf
       priorityFlag 1 bslbf
       activateFlag 1 bslbf
       valueFlag 1 bslbf
       unitFlag 1 bslbf
       if(TimeStampFlag) {
        TimestampType 2 bslbf
        if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
        } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
        } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
        }
       }
       If(IDFlag) {
       ID See ISO 10646 UTF-8
       }
       if(sensorIdRefFlag) {
       sensorIdRefLength vluimsbf5
       sensorIdRef  8* bslbf
        sensorIdRefLength
       }
       if(linkedlistFlag) {
       linkedlistLength vluimsbf5
       linkedlist  8* bslbf
        linkedlistLength
       }
       if(groupIDFlag) {
       groupIDLength vluimsbf5
       groupID  8* bslbf
        groupIDLength
       }
       If(priorityFlag) {
         priority 8 uimsbf
       }
       if(activateFlag) {
         activate 1 bslbf
       }
       if(lifespanFlag) {
         lifespan 32  fsbf
       }
       if(valueFlag) {
         value 32  fsbf
       }
       if(unitFlag) {
         unit unitType
       }
    }
  • Table 1-3 shows descriptor components semantics of the temperature sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    TemperatureSensorType Tool for describing sensed information with respect to a
    temperature sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be used
    and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme shall
    be used. “00” means that the absolute time stamp type
    shall be used, “01” means that the clock tick time stamp
    type shall be used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined in
    A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    temperature sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the information
    included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor structure
    that consists of a group of sensors in a way that each
    record contains a reference to the ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which this
    temperature sensor belongs.
    priority Describes a priority for sensed information with respect to
    other sensed information sharing the same point in time
    when the sensed information becomes adapted. A value
    of zero indicates the highest priority and larger values
    indicate lower priorities. The default value of the priority
    is zero. If there is more than one sensed information with
    the same priority, the order of process can be determined
    by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    value Describes the sensed value of the temperature sensor
    with respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other than
    the default unit is used, as a reference to a classification
    scheme term provided by UnitCS defined in xxx of
    ISO/IEC 23005-6 and use the binary representation
    defined above.
  • Source 44 denotes sensed information related to a humidity sensor using the XML. However, a program source shown in Source 44 is only an example embodiment but does not limit the present invention.
  • [Source 44]
    <!--#################################### -->
    <!--Definition of Humidity Sensor type -->
    <!--#################################### -->
    <complexType name=″HumiditySensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″iidl:unitType″ use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A humidity sensor type is a tool for describing sensed information related to the humidity sensor.
  • The humidity sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the humidity sensor.
  • The life span denotes information on a valid period of a command of the humidity sensor. For example, the life span may be a second unit.
  • The value denotes information on a humidity sensor value expressed by a unit of humidity, for example, %.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the humidity sensor according to the example embodiments.
  • [Source 41-2]
    HumiditySensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      valueFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if (sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRefLength
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
       activate 1 bslbf
      }
      if(valueFlag) {
       value 32  fsbf
      }
      if(unitFlag) {
       Unit unitType
      }
    }
  • Table 1-3 shows descriptor components semantics of the humidity sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    HumiditySensorType Tool for describing sensed information with respect to a
    humidity sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme shall
    be used. “00” means that the absolute time stamp type
    shall be used, “01” means that the clock tick time stamp
    type shall be used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined in
    A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    humidity sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the information
    included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor structure
    that consists of a group of sensors in a way that each
    record contains a reference to the ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which this
    humidity sensor belongs.
    priority Describes a priority for sensed information with respect to
    other sensed information sharing the same point in time
    when the sensed information becomes adapted. A value
    of zero indicates the highest priority and larger values
    indicate lower priorities. The default value of the priority
    is zero. If there is more than one sensed information with
    the same priority, the order of process can be determined
    by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    value Describes the sensed value of the humidity sensor with
    respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other than
    the default unit is used, as a reference to a classification
    scheme term provided by UnitCS defined in xxx of
    ISO/IEC 23005-6 and use the binary representation
    defined above.
  • Source 45 denotes sensed information related to a distance sensor using the XML. However, a program source shown in Source 45 is only an example embodiment but does not limit the present invention.
  • [Source 45]
    <!--#################################### -->
    <!--Definition of Distance Sensor type -->
    <!--#################################### -->
    <complexType name=″DistanceSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″iidl:unitType″ use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A distance sensor type is a tool for describing sensed information related to the distance sensor.
  • The distance sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the distance sensor.
  • The life span denotes information on a valid period of a command of the distance sensor. For example, the life span may be a second unit.
  • The value denotes information on a distance sensor value expressed by a unit of distance, for example, meter.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the distance sensor according to the example embodiments.
  • [Source 41-2]
    DistanceSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      valueFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       }     else     if
     (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       }     else     if
     (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRefLength
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
       linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(valueFlag) {
        value 32  fsbf
      }
      if(unitFlag) {
        unit unitType
      }
      }
  • Table 1-3 shows descriptor components semantics of the distance sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    DistanceSensorType Tool for describing sensed information with respect to a
    distance sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be used
    and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme shall
    be used. “00” means that the absolute time stamp type
    shall be used, “01” means that the clock tick time stamp
    type shall be used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined in
    A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    distance sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the information
    included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor structure
    that consists of a group of sensors in a way that each
    record contains a reference to the ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which this
    distance sensor belongs.
    priority Describes a priority for sensed information with respect
    to other sensed information sharing the same point in time
    when the sensed information becomes adapted. A value
    of zero indicates the highest priority and larger values
    indicate lower priorities. The default value of the priority
    is zero. If there is more than one sensed information with
    the same priority, the order of process can be determined
    by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    value Describes the sensed value of the distance sensor with
    respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other than
    the default unit is used, as a reference to a classification
    scheme term provided by UnitCS defined in xxx of
    ISO/IEC 23005-6 and use the binary representation
    defined above.
  • Source 46 denotes sensed information related to a motion sensor using the XML. However, a program source shown in Source 46 is only an example embodiment but does not limit the present invention.
  • [Source 46]
    <!-- ################################################ -->
    <!-- Definition of Motion Sensor Type      -->
    <!-- ################################################ -->
    <complexType name=″MotionSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″position″
    type=″siv:PositionSensorType″ minOccurs=″0″/>
    <element name=″orientation″
    type=″siv:OrientationSensorType″ minOccurs=″0″/>
    <element name=″velocity″
    type=″siv:VelocitySensorType″ minOccurs=″0″/>
    <element name=″angularvelocity″
    type=″siv:AngularVelocitySensorType″ minOccurs=″0″/>
    <element name=″acceleration″
    type=″siv:AccelerationSensorType″ minOccurs=″0″/>
    <element name=″angularacceleration″
    type=″siv:AngularAccelerationSensorType″ minOccurs=″0″/>
    </sequence>
    </extension>
    </complexContent>
    </complexType>
  • A motion sensor type is a tool for describing sensed information related to the length sensor.
  • The motion sensor type may include attributes such as an ID, an ID reference, a position, a velocity, an acceleration, an orientation, an angular velocity, and an angular acceleration.
  • The ID denotes ID information for recognizing individual identity of the motion sensor.
  • The ID reference denotes additional information related to the ID, the additional information for recognizing individual identity of the motion sensor.
  • The position denotes information on a position vector value of a unit of position, for example, meter.
  • The velocity denotes information on a velocity vector value of a unit of velocity, for example, m/s.
  • The acceleration denotes information on an acceleration vector value of a unit of acceleration, for example, m/s2.
  • The orientation denotes information on an orientation vector value of a unit of orientation, for example, radian.
  • The angular velocity denotes information on an angular velocity vector value of a unit of angular velocity, for example, radian/s.
  • The angular acceleration denotes information on an angular acceleration vector value of a unit of angular acceleration, for example, radian/s2.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the motion sensor according to the example embodiments.
  • [Source 41-2]
    MotionSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      positionFlag 1 bslbf
      orientationFlag 1 bslbf
      velocityFlag 1 bslbf
      angularvelocityFlag 1 bslbf
      accelerationFlag 1 bslbf
      angularaccelerationFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       }     else     if
     (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       }     else     if
     (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRefLength
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(positionFlag) {
       position PositionSensorType
      }
      if(orientationFlag) {
       orientation OrientationSensorType
      }
      if(velocityFlag) {
       velocity VelocitySensorType
      }
      if(angularvelocityFlag) {
       angularvelocity AngularVelocitySensorType
      }
      if(accelerationFlag) {
       acceleration AccelerationSensorType
      }
      if(angularaccelerationFlag) {
       angularacceleration AngularAccelerationSensorType
      }
  • Table 1-3 shows descriptor components semantics of the motion sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    MotionSensorType Tool for describing sensed information with respect to a
    motion sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    positionFlag This field, which is only present in the binary
    representation, signals the presence of position value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    orientationFlag This field, which is only present in the binary
    representation, signals the presence of orientation value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    velocityFlag This field, which is only present in the binary
    representation, signals the presence of velocity value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    angularvelocityFlag This field, which is only present in the binary
    representation, signals the presence of angular velocity
    value attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be used.
    accelerationFlag This field, which is only present in the binary
    representation, signals the presence of acceleration
    value attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be used.
    angularaccelerationFlag This field, which is only present in the binary
    representation, signals the presence of angular
    acceleration value attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute shall
    not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time stamp
    type shall be used, “01” means that the clock tick time
    stamp type shall be used, and “10” means that the clock
    tick time delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined
    in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    motion sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a way
    that each record contains a reference to the ID of the
    next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which
    this motion sensor belongs.
    priority Describes a priority for sensed information with respect
    to other sensed information sharing the same point in
    time when the sensed information becomes adapted.
    A value of zero indicates the highest priority and larger
    values indicate lower priorities. The default value of
    the priority is zero. If there is more than one sensed
    information with the same priority, the order of process
    can be determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    position Describes the sensed position value of the motion
    sensor with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the
    sensor capability.
    orientation Describes the sensed orientation value of the motion
    sensor with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the
    sensor capability.
    velocity Describes the sensed velocity value of the motion
    sensor with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the
    sensor capability.
    angularvelocity Describes the sensed angular velocity value of the
    motion sensor with respect to the default unit if the unit
    is not defined. Otherwise, use the unit type defined in
    the sensor capability.
    acceleration Describes the sensed acceleration value of the motion
    sensor with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the
    sensor capability.
    angularacceleration Describes the sensed angular acceleration value of the
    motion sensor with respect to the default unit if the unit
    is not defined. Otherwise, use the unit type defined in
    the sensor capability.
  • Source 47 denotes sensed information related to an intelligent camera sensor using the XML. However, a program source shown in Source 47 is only an example embodiment but does not limit the present invention.
  • [Source 47]
    <!-- ################################################ -->
    <!-- Definition of Intelligent Camera Type     -->
    <!-- ################################################ -->
    <complexType name=″IntelligentCameraType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″FacialAnimationID″
    type=″anyURI″ minOccurs=″0″/>
    <element name=″BodyAnimationID″
    type=″anyURI″ minOccurs=″0″/>
    <element name=″FaceFeature″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″ maxOccurs=″260″>
    <element name=″BodyFeature″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″ maxOccurs=″260″>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • An intelligent camera sensor type is a tool for describing sensed information related to the intelligent camera sensor.
  • The intelligent camera sensor type may include a facial animation ID, a body animation ID, a face feature, and a body feature.
  • The facial animation ID denotes an ID referencing an animation clip with respect to a facial expression.
  • The body animation ID denotes an ID referencing an animation clip with respect to a body.
  • The face feature denotes information on a 3D position of each face feature sensed by the intelligent camera sensor.
  • The body feature denotes information on a 3D position of each body feature sensed by the intelligent camera sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the intelligent camera sensor according to the example embodiments.
  • [Source 41-2]
    IntelligentCameraType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      FacialIDFlag 1 bslbf
      BodyIDFlag 1 bslbf
      FaceFeatureFlag 1 bslbf
      BodyFeatureFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       }     else     if
     (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       }     else     if
     (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRefLength
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
       activate 1 bslbf
      }
      if( FacialIDFlag ) {
       FacialAnimationIDLength vluimsbf5
       FacialAnimationID 8 * bslbf
    FacialAnimationIDLength
      }
      if( BodyIDFlag ) {
       BodyAnimationIDLength vluimsbf5
       BodyAnimationID 8 * bslbf
    BodyAnimationIDLength
      }
      if( FaceFeatureFlag ) {
       NumOfFaceFeature 8 uimsbf
       for( k=0;
         k<NumOfFaceFeature;
     k++ ) {
        FaceFeature[k] Float3DVectorType
       }
      }
      if( BodyFeatureFlag ) {
       NumOfBodyFeature 8 uimsbf
       for( k=0;
     k<NumOfBodyFeature;
         k++ ) {
        BodyFeature[k] Float3DVectorType
       }
      }
    }
  • Table 1-3 shows descriptor components semantics of the intelligent camera sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    IntelligentCameraType Tool for describing sensed information with respect to
    an intelligent camera sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    FacialIDFlag This field, which is only present in the binary
    representation, signals the presence of the facial
    animation ID. A value of “1” means the facial animation
    ID mode shall be used and “0” means the facial
    animation ID mode shall not be used.
    BodyIDFlag This field, which is only present in the binary
    representation, signals the presence of the body
    animation ID. A value of “1” means the body animation
    ID mode shall be used and “0” means the body
    animation ID mode shall not be used.
    FaceFeatureFlag This field, which is only present in the binary
    representation, signals the presence of the face
    features. A value of “1” means the face feature
    tracking mode shall be used and “0” means the face
    feature tracking mode shall not be used.
    BodyFeatureFlag This field, which is only present in the binary
    representation, signals the presence of the body
    features. A value of “1” means the body feature
    tracking mode shall be used and “0” means the body
    feature tracking mode shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time stamp
    type shall be used, “01” means that the clock tick time
    stamp type shall be used, and “10” means that the clock
    tick time delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined
    in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    intelligent camera sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a way
    that each record contains a reference to the ID of the
    next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which
    this intelligent camera sensor belongs.
    priority Describes a priority for sensed information with respect
    to other sensed information sharing the same point in
    time when the sensed information becomes adapted.
    A value of zero indicates the highest priority and larger
    values indicate lower priorities. The default value of
    the priority is zero. If there is more than one sensed
    information with the same priority, the order of process
    can be determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    FacialAnimationIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    FacialAnimationID attribute.
    FacialAnimationID Describes the ID referencing the facial expression
    animation clip.
    BodyAnimationIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    BodyAnimationID attribute.
    BodyAnimationID Describes the ID referencing the body animation clip.
    NumOfFaceFeature This field, which is only present in the binary
    representation, specifies the number of face feature
    points.
    FaceFeature Describes the 3D position of each of the face feature
    points detected by the camera.
    Note: The order of the elements corresponds to the
    order of the face feature points defined at the
    featureControl for face in 2.2.15 of ISO/IEC_23005-4
    NumOfBodyFeature This field, which is only present in the binary
    representation, specifies the number of body feature
    points.
    BodyFeature Describes the 3D position of each of the body feature
    points detected by the camera.
    Note: The order of the elements corresponds to the
    order of the body feature points defined at the
    featureControl for body in 2.2.14 of ISO/IEC_23005-4.
  • Source 48 denotes sensed information related to an ambient noise sensor using the XML. However, a program source shown in Source 48 is only an example embodiment but does not limit the present invention.
  • [Source 48]
    <!--#################################### -->
    <!--Definition of Ambient Noise Sensor type -->
    <!--#################################### -->
    <complexType name=″AmbientNoiseSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″lifespan″ type=″float″ use=″optional″/>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″iidl:unitType″ use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • An ambient noise sensor type is a tool for describing sensed information related to the ambient noise sensor.
  • The ambient noise sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the ambient noise sensor.
  • The life span denotes information on a valid period of a command of the ambient noise sensor. For example, the life span may be a second unit.
  • The value denotes information on an ambient noise sensor value expressed by a unit of sound intensity, for example, dB.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the ambient noise sensor according to the example embodiments.
  • [Source 41-2]
    Number of bits Mnemonic
    AmbientNoiseSensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    lifespanFlag 1 bslbf
    valueFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(lifespanFlag) {
    lifespan 32  fsbf
    }
    if(valueFlag) {
    value 32  fsbf
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • Table 1-3 shows descriptor components semantics of the ambient noise sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    AmbientNoiseSensorType Tool for describing sensed information with respect to an
    ambient noise sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be
    used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    lifespanFlag This field, which is only present in the binary
    representation, signals the presence of the life span
    attribute. A value of “1” means the lifespan shall be
    used and “0” means the lifespan shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute. A
    value of “1” means the user-defined unit shall be used
    and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time stamp
    type shall be used, “01” means that the clock tick time
    stamp type shall be used, and “10” means that the clock
    tick time delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the time
    delta between the present and the past time, is defined
    in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    temperature sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the information
    included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a way
    that each record contains a reference to the ID of the
    next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which this
    temperature sensor belongs.
    priority Describes a priority for sensed information with respect
    to other sensed information sharing the same point in
    time when the sensed information becomes adapted.
    A value of zero indicates the highest priority and larger
    values indicate lower priorities. The default value of
    the priority is zero. If there is more than one sensed
    information with the same priority, the order of process
    can be determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    lifespan Describes the duration taken to measure the information
    based on the timestamp.
    value Describes the sensed value of the ambient noise sensor
    with respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other
    than the default unit is used, as a reference to a
    classification scheme term provided by UnitCS defined
    in xxx of ISO/IEC 23005-6 and use the binary
    representation defined above.
  • Source 49 denotes sensed information related to an atmospheric pressure sensor using the XML. However, a program source shown in Source 49 is only an example embodiment but does not limit the present invention.
  • [Source 49]
    <!--#################################### -->
    <!--Definition of Atmospheric pressure Sensor type -->
    <!--#################################### -->
    <complexType name=″AtmosphericPressureSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″iidl:unitType″ use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • An atmospheric pressure sensor type is a tool for describing sensed information related to the atmospheric pressure sensor.
  • The atmospheric pressure sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the atmospheric pressure sensor.
  • The life span denotes information on a valid period of a command of the atmospheric pressure sensor. For example, the life span may be a second unit.
  • The value denotes information on an atmospheric pressure sensor value expressed by a unit of atmospheric pressure, for example, hPa.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the atmospheric pressure sensor according to the example embodiments.
  • [Source 41-2]
    Number of bits Mnemonic
    AtmosphericPressureSensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    valueFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(valueFlag) {
    value 32  fsbf
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • Table 1-3 shows descriptor components semantics of the atmospheric pressure sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    AtmosphericPressureSensorType Tool for describing sensed information with
    respect to an atmospheric pressure sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the
    sensor ID reference attribute. A value of “1”
    means the attribute shall be used and “0” means
    the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the
    group ID attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the
    priority attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall
    be used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is
    the time delta between the present and the past
    time, is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with
    respect to an atmospheric pressure sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the
    following sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the
    following linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the
    following groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this atmospheric pressure sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is
    zero. If there is more than one sensed
    information with the same priority, the order of
    process can be determined by the Adaptation
    engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated
    and “0” means the sensor is deactivated.
    value Describes the sensed value of the atmospheric
    pressure sensor with respect to the default unit if
    the unit is not defined. Otherwise, use the unit
    type defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and
    use the binary representation defined above.
  • Source 50 denotes sensed information related to a velocity sensor using the XML. However, a program source shown in Source 50 is only an example embodiment but does not limit the present invention.
  • [Source 50]
    <!--#################################### -->
    <!--Definition of Velocity Sensor type -->
    <!--#################################### -->
    <complexType name=″VelocitySensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″velocity″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″/>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A velocity sensor type is a tool for describing sensed information related to the velocity sensor.
  • The velocity sensor type may include attributes such as a timestamp, a life span, a velocity, a velocity value type, Vx, Vy, and Vz.
  • The timestamp denotes information on a sensing time of the velocity sensor.
  • The life span denotes information on a valid period of a command of the velocity sensor. For example, the life span may be a second unit.
  • The velocity denotes information on a velocity sensor value expressed by a unit of velocity, for example, m/s.
  • The velocity value type denotes a tool for indicating a 3D velocity vector.
  • The Vx denotes information on an x-axis value of the velocity sensor.
  • The Vy denotes information on a y-axis value of the velocity sensor.
  • The Vz denotes information on a z-axis value of the velocity sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the velocity sensor according to the example embodiments.
  • [Source 41-2]
    Number of bits Mnemonic
    VelocitySensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    velocityFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(velocityFlag) {
    velocity Float3DVectorType
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • Table 1-3 shows descriptor components semantics of the velocity sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    VelocitySensorType Tool for describing sensed information with
    respect to a velocity sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the
    sensor ID reference attribute. A value of “1”
    means the attribute shall be used and “0” means
    the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group
    ID attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the
    priority attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    velocityFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall be
    used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time,
    is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to a velocity sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the
    following sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the
    following linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the
    following groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this velocity sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is
    zero. If there is more than one sensed
    information with the same priority, the order of
    process can be determined by the Adaptation
    engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    velocity Describes the sensed value of the velocity sensor
    in 3D with respect to the default unit if the unit is
    not defined. Otherwise, use the unit type
    defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and
    use the binary representation defined above.
  • Source 51 denotes sensed information related to an angular velocity sensor using the XML. However, a program source shown in Source 51 is only an example embodiment but does not limit the present invention.
  • [Source 51]
    <!--#################################### -->
    <!--Definition of Angular Velocity Sensor type -->
    <!--#################################### -->
    <complexType name=″AngularVelocitySensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″AngularVelocity″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″/>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • An angular velocity sensor type is a tool for describing sensed information related to the angular velocity sensor.
  • The angular velocity sensor type may include attributes such as a timestamp, a life span, an angular velocity, an angular velocity value type, AVx, AVy, and AVz.
  • The timestamp denotes information on a sensing time of the angular velocity sensor.
  • The life span denotes information on a valid period of a command of the angular velocity sensor. For example, the life span may be a second unit.
  • The angular velocity denotes information on an angular velocity sensor value expressed by a unit of angular velocity, for example, radian.
  • The angular velocity value type denotes a tool for indicating a 3D angular velocity vector.
  • The AVx denotes information on a value of an x-axis rotation angular velocity of the angular velocity sensor.
  • The AVy denotes information on a value of a y-axis rotation angular velocity of the angular velocity sensor.
  • The AVz denotes information on a value of a z-axis rotation angular velocity of the angular velocity sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the angular velocity sensor according to the example embodiments.
  • [Source 41-2]
    Number of bits Mnemonic
    AngularVelocitySensorType{
    TimeStampFlag 1 bslbf
    IDFlag 1 bslbf
    sensorIdRefFlag 1 bslbf
    linkedlistFlag 1 bslbf
    groupIDFlag 1 bslbf
    priorityFlag 1 bslbf
    activateFlag 1 bslbf
    angularvelocityFlag 1 bslbf
    unitFlag 1 bslbf
    if(TimeStampFlag) {
    TimestampType 2 bslbf
    if(TimeStampType==00){
    AbsoluteTimeStamp AbsoluteTimeStampType
    } else if (TimeStampType==01){
    ClockTickTimeStamp ClockTickTimeStampType
    } else if (TimeStampType==10){
    ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
    }
    }
    If(IDFlag) {
    ID See ISO 10646 UTF-8
    }
    if(sensorIdRefFlag) {
    sensorIdRefLength vluimsbf5
    sensorIdRef 8* bslbf
    sensorIdRef
    Length
    }
    if(linkedlistFlag) {
    linkedlistLength vluimsbf5
    linkedlist 8* bslbf
    linkedlistLength
    }
    if(groupIDFlag) {
    groupIDLength vluimsbf5
    groupID 8* bslbf
    groupIDLength
    }
    If(priorityFlag) {
    priority 8 uimsbf
    }
    if(activateFlag) {
    activate 1 bslbf
    }
    if(angularvelocityFlag) {
    angularvelocity Float3DVectorType
    }
    if(unitFlag) {
    unit unitType
    }
    }
  • Table 1-3 shows descriptor components semantics of the angular velocity sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    AngularVelocitySensorType Tool for describing sensed information with
    respect to an angular velocity sensor
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the
    sensor ID reference attribute. A value of “1”
    means the attribute shall be used and “0” means
    the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group
    ID attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the
    priority attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    angularvelocityFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall be
    used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time,
    is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to an angular velocity sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the
    following sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the
    following linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the
    following groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this angular velocity sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is
    zero. If there is more than one sensed
    information with the same priority, the order of
    process can be determined by the Adaptation
    engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    angularvelocity Describes the sensed value of the angular
    velocity sensor in 3D with respect to the default
    unit if the unit is not defined. Otherwise, use the
    unit type defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and
    use the binary representation defined above.
  • Source 52 denotes sensed information related to an angular acceleration sensor using the XML. However, a program source shown in Source 52 is only an example embodiment but does not limit the present invention.
  • [Source 52]
    <!--############################################### -->
    <!--Definition of Angular Acceleration Sensor type -->
    <!--############################################### -->
    <complexType name=″AngularAccelerationSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″AngularAcceleration″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″/>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • An angular acceleration sensor type is a tool for describing sensed information related to the angular acceleration sensor.
  • The angular acceleration sensor type may include attributes such as a timestamp, a life span, an angular acceleration, an angular acceleration value type, an AAx, an AAy, and an AAz.
  • The timestamp denotes information on a sensing time of the angular acceleration sensor.
  • The life span denotes information on a valid period of a command of the angular acceleration sensor. For example, the life span may be a second unit.
  • The angular acceleration denotes information on an angular acceleration sensor value expressed by a unit of angular acceleration, for example, radian/s2.
  • The angular acceleration value type denotes a tool for indicating a 3D angular acceleration vector.
  • The AAx denotes information on an x-axis value of the angular acceleration sensor.
  • The AAy denotes information on a y-axis value of the angular acceleration sensor.
  • The AAz denotes information on a z-axis value of the angular acceleration sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the angular acceleration sensor according to the example embodiments.
  • [Source 41-2]
    AngularAccelerationSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      angularaccelerationFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
      }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRef
    Length
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(angularaccelerationFlag) {
       angularacceleration Float3DVectorType
      }
      if(unitFlag) {
        unit unitType
      }
    }
  • Table 1-3 shows descriptor components semantics of the angular acceleration sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    AngularAccelerationSensorType Tool for describing sensed information with
    respect to an angular acceleration sensor
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the timestamp
    shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the
    sensor ID reference attribute. A value of “1”
    means the attribute shall be used and “0” means
    the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not
    be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group
    ID attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall
    not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the
    priority attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    angularaccelerationFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall be
    used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time,
    is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to an angular acceleration sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the
    following sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the
    following linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the
    following groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this angular acceleration sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is
    zero. If there is more than one sensed
    information with the same priority, the order of
    process can be determined by the Adaptation
    engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    angularacceleration Describes the sensed value of the angular
    acceleration sensor in 3D with respect to the
    default unit if the unit is not defined. Otherwise,
    use the unit type defined in the sensor capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and
    use the binary representation defined above.
  • Source 53 denotes sensed information related to a force sensor using the XML. However, a program source shown in Source 53 is only an example embodiment but does not limit the present invention.
  • [Source 53]
    <!--#################################### -->
    <!--Definition of Force Sensor type  -->
    <!--#################################### -->
    <complexType name=″ForceSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″force″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″/>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A force sensor type is a tool for describing sensed information related to the force sensor.
  • The force sensor type may include attributes such as a timestamp, a life span, a force, a force value type, FSx, FSy, and FSz.
  • The timestamp denotes information on a sensing time of the force sensor.
  • The life span denotes information on a valid period of a command of the force sensor. For example, the life span may be a second unit.
  • The force denotes information on a force sensor value expressed by a unit of force, for example, N.
  • The force value type denotes a tool for indicating a 3D force vector.
  • The FSx denotes information on an x-axis force value of the force sensor.
  • The FSy denotes information on a y-axis force value of the force sensor.
  • The FSz denotes information on a z-axis force value of the force sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the force sensor according to the example embodiments.
  • [Source 41-2]
    ForceSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      forceFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRef
    Length
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(forceFlag) {
       Force Float3DVectorType
      }
      if(unitFlag) {
        Unit unitType
      }
    }
  • Table 1-3 shows descriptor components semantics of the force sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    ForceSensorType Tool for describing sensed information with
    respect to a force sensor
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the
    timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the
    sensor ID reference attribute. A value of “1”
    means the attribute shall be used and “0” means
    the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group
    ID attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the
    priority attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    forceFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the
    attribute shall be used and “0” means the
    attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall be
    used, and “10” means that the clock tick time
    delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time,
    is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to a force sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the
    following sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the
    following linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the
    following groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this force sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is
    zero. If there is more than one sensed
    information with the same priority, the order of
    process can be determined by the Adaptation
    engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0”
    means the sensor is deactivated.
    force Describes the sensed value of the force sensor in
    3D with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and use the
    binary representation defined above.
  • Source 54 denotes sensed information related to a torque sensor using the XML. However, a program source shown in Source 54 is only an example embodiment but does not limit the present invention.
  • [Source 54]
    <!--#################################### -->
    <!--Definition of Torque Sensor type  -->
    <!--#################################### -->
    <complexType name=″TorqueSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <sequence>
    <element name=″Torque″
    type=″mpegvct:Float3DVectorType″ minOccurs=″0″/>
    </sequence>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A torque sensor type is a tool for describing sensed information related to the torque sensor.
  • The torque sensor type may include attributes such as a timestamp, a life span, a torque, a torque value type, TSx, TSy, and TSz.
  • The timestamp denotes information on a sensing time of the torque sensor.
  • The life span denotes information on a valid period of a command of the torque sensor. For example, the life span may be a second unit.
  • The torque denotes information on a torque sensor value expressed by a unit of torque, for example, N-mm.
  • The torque value type denotes a tool for indicating a 3D torque vector.
  • The TSx denotes information on an x-axis torque value of the torque sensor.
  • The TSy denotes information on a y-axis torque value of the torque sensor.
  • The TSz denotes information on a z-axis torque value of the force sensor.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the torque sensor according to the example embodiments.
  • [Source 41-2]
    TorqueSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      TorqueFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRef
    Length
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(torqueFlag) {
       torque Float3DVectorType
      }
      if(unitFlag) {
        unit unitType
      }
    }
  • Table 1-3 shows descriptor components semantics of the torque sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    ForceSensorType Tool for describing sensed information with respect to a
    force sensor
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the
    timestamp element. A value of “1” means the
    timestamp shall be used and “0” means the timestamp
    shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be
    used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor
    ID reference attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute
    shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked
    list attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group
    ID attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall
    be used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the
    activation attribute. A value of “1” means the
    attribute shall be used and “0” means the attribute shall not
    be used.
    torqueFlag This field, which is only present in the binary
    representation, signals the presence of sensor
    value attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit
    attribute. A value of “1” means the user-
    defined unit shall be used and “0” means the
    user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp
    scheme shall be used. “00” means that the
    absolute time stamp type shall be used, “01”
    means that the clock tick time stamp type shall be
    used, and “10” means that the clock tick time delta
    stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time,
    is defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect
    to a torque sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed
    information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a
    way that each record contains a reference to the
    ID of the next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to
    which this torque sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the
    same point in time when the sensed information
    becomes adapted. A value of zero indicates the
    highest priority and larger values indicate lower
    priorities. The default value of the priority is zero.
    If there is more than one sensed information with
    the same priority, the order of process can be
    determined by the Adaptation engine itself.
    activate Describes whether the sensor is activated. A
    value of “1” means the sensor is activated and “0” means
    the sensor is deactivated.
    torque Describes the sensed value of the torque sensor in
    3D with respect to the default unit if the unit is not
    defined. Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit
    other than the default unit is used, as a reference
    to a classification scheme term provided by
    UnitCS defined in xxx of ISO/IEC 23005-6 and use the
    binary representation defined above.
  • Source 55 denotes sensed information related to a pressure sensor using the XML. However, a program source shown in Source 55 is only an example embodiment but does not limit the present invention.
  • [Source 55]
    <!--#################################### -->
    <!--Definition of Pressure Sensor type -->
    <!--#################################### -->
    <complexType name=″PressureSensorType″>
    <complexContent>
    <extension base=″iidl:SensedInfoBaseType″>
    <attribute name=″timestamp″ type=″float″
    use=″optional″/>
    <attribute name=″value″ type=″float″ use=″optional″/>
    <attribute name=″unit″ type=″mpegvct:unitType″
    use=″optional″/>
    </extension>
    </complexContent>
    </complexType>
  • A pressure sensor type is a tool for describing sensed information related to the pressure sensor.
  • The pressure sensor type may include attributes such as a timestamp, a life span, and a value.
  • The timestamp denotes information on a sensing time of the pressure sensor.
  • The life span denotes information on a valid period of a command of the pressure sensor. For example, the life span may be a second unit.
  • The value denotes information on a pressure sensor value expressed by a unit of pressure, for example, N/mm2.
  • Source 41-2 denotes a binary representation syntax corresponding to the sensed information related to the torque pressure according to the example embodiments.
  • [Source 41-2]
    PressureSensorType{ Number of bits Mnemonic
      TimeStampFlag 1 bslbf
      IDFlag 1 bslbf
      sensorIdRefFlag 1 bslbf
      linkedlistFlag 1 bslbf
      groupIDFlag 1 bslbf
      priorityFlag 1 bslbf
      activateFlag 1 bslbf
      valueFlag 1 bslbf
      unitFlag 1 bslbf
      if(TimeStampFlag) {
       TimestampType 2 bslbf
       if(TimeStampType==00){
        AbsoluteTimeStamp AbsoluteTimeStampType
       } else if (TimeStampType==01){
        ClockTickTimeStamp ClockTickTimeStampType
       } else if (TimeStampType==10){
        ClockTickTimeDeltaStamp ClockTickTimeDeltaStampType
       }
      }
      If(IDFlag) {
      ID See ISO 10646 UTF-8
      }
      if(sensorIdRefFlag) {
      sensorIdRefLength vluimsbf5
      sensorIdRef 8 * bslbf
    sensorIdRef
    Length
      }
      if(linkedlistFlag) {
      linkedlistLength vluimsbf5
      linkedlist 8 * bslbf
    linkedlistLength
      }
      if(groupIDFlag) {
      groupIDLength vluimsbf5
      groupID 8 * bslbf
    groupIDLength
      }
      If(priorityFlag) {
        priority 8 uimsbf
      }
      if(activateFlag) {
        activate 1 bslbf
      }
      if(valueFlag) {
        value 32  fsbf
      }
      if(unitFlag) {
        unit unitType
      }
    }
  • Table 1-3 shows descriptor components semantics of the pressure sensor according to the example embodiments.
  • TABLE 1-3
    Names Description
    PressureSensorType Tool for describing sensed information with respect to
    a pressure sensor.
    TimeStampFlag This field, which is only present in the binary
    representation, signals the presence of the timestamp
    element. A value of “1” means the timestamp shall be
    used and “0” means the timestamp shall not be used.
    IDFlag This field, which is only present in the binary
    representation, signals the presence of the ID attribute.
    A value of “1” means the attribute shall be used and “0”
    means the attribute shall not be used.
    sensorIdRefFlag This field, which is only present in the binary
    representation, signals the presence of the sensor ID
    reference attribute. A value of “1” means the attribute
    shall be used and “0” means the attribute shall not be used.
    linkedlistFlag This field, which is only present in the binary
    representation, signals the presence of the linked list
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    groupIDFlag This field, which is only present in the binary
    representation, signals the presence of the group ID
    attribute. A value of “1” means the attribute shall be used
    and “0” means the attribute shall not be used.
    priorityFlag This field, which is only present in the binary
    representation, signals the presence of the priority
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    activateFlag This field, which is only present in the binary
    representation, signals the presence of the activation
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    valueFlag This field, which is only present in the binary
    representation, signals the presence of sensor value
    attribute. A value of “1” means the attribute shall be
    used and “0” means the attribute shall not be used.
    unitFlag This field, which is only present in the binary
    representation, signals the presence of unit attribute.
    A value of “1” means the user-defined unit shall be
    used and “0” means the user-defined unit shall not be used.
    TimeStampType This field, which is only present in the binary
    representation, describes which time stamp scheme
    shall be used. “00” means that the absolute time
    stamp type shall be used, “01” means that the clock tick
    time stamp type shall be used, and “10” means that the
    clock tick time delta stamp type shall be used.
    AbsoluteTimeStamp The absolute time stamp is defined in A.2.3 of ISO/IEC
    23005-6.
    ClockTickTimeStamp The clock tick time stamp is defined in A.2.3 of
    ISO/IEC 23005-6.
    ClockTickTimeDeltaStamp The clock tick time delta stamp, which value is the
    time delta between the present and the past time, is
    defined in A.2.3 of ISO/IEC 23005-6.
    ID ID to identify the sensed information with respect to a
    pressure sensor.
    sensorIdRefLength This field, which is only present in the binary
    representation, specifies the length of the following
    sensorIdRef attribute.
    sensorIdRef References a sensor that has generated the
    information included in this specific sensed information.
    linkedlistLength This field, which is only present in the binary
    representation, specifies the length of the following
    linkedlist attribute.
    linkedlist Identifier for the next sensor of the multi-sensor
    structure that consists of a group of sensors in a way
    that each record contains a reference to the ID of the
    next sensor.
    groupIDLength This field, which is only present in the binary
    representation, specifies the length of the following
    groupID attribute.
    groupID Identifier for a group multi-sensor structure to which
    this pressure sensor belongs.
    priority Describes a priority for sensed information with
    respect to other sensed information sharing the same
    point in time when the sensed information becomes
    adapted. A value of zero indicates the highest priority
    and larger values indicate lower priorities. The default
    value of the priority is zero. If there is more than one
    sensed information with the same priority, the order of
    process can be determined by the Adaptation engine
    itself.
    activate Describes whether the sensor is activated. A value
    of “1” means the sensor is activated and “0” means the
    sensor is deactivated.
    value Describes the sensed value of the pressure sensor
    with respect to the default unit if the unit is not defined.
    Otherwise, use the unit type defined in the sensor
    capability.
    unit Specifies the unit of the sensed value, if a unit other
    than the default unit is used, as a reference to a
    classification scheme term provided by UnitCS defined
    in xxx of ISO/IEC 23005-6 and use the binary
    representation defined above.
  • Source 0-0 may be a program representing a timestamp type using the XML. However, the program source shown in Source 0-0 is only an example embodiment but does not limit the present invention.
  • <complexType name=″TimeStampType″ abstract=″true″/>
    <complexType name=″AbsoluteTimeType″>
    <complexContent>
    <extension base=″ct:TimeStampType″>
    <attribute name=″absTimeScheme″ type=″string″ use=″optional″/>
    <attribute name=″absTime″ type=″string″/>
    </extension>
    </complexContent>
    </complexType>
    <complexType name=″ClockTickTimeType″>
    <complexContent>
    <extension base=″ct:TimeStampType″>
    <attribute name=″timeScale″ type=″unsignedInt″ use=″optional″/>
    <attribute name=″pts″ type=″nonNegativeInteger″/>
    </extension>
    </complexContent>
    </complexType>
    <complexType name=″ClockTickTimeDeltaType″>
    <complexContent>
    <extension base=″ct:TimeStampType″>
    <attribute name=″timeScale″ type=″unsignedInt″ use=″optional″/>
    <attribute name=″ptsDelta″ type=″unsignedInt>
    </extension>
    </complexContent>
    </complexType>
  • Source 41-2 denotes a binary representation syntax related to the timestamp type according to the example embodiments.
  • [Source 41-2]
    Number of bits Mnemonic
    AbsoluteTimeStampType {
      absTimeSchemeFlag 1 bslbf
     if(absTimeSchemeFlag){
      AbsTimeSchemeLength vluimsbf5
     absTimeScheme 8 * bslbf
    AbsTimeSchemeLength
     }
     AbsTimeLength vluimsbf5
     absTime 8 * AbsTimeLength bslbf
    }
    ClockTickTimeType {
     timeScaleFlag 1 bslbf
     if(timeScaleFlag){
      timeScale 32  uimsbf
     }
     pts vluimsbf5
    }
    ClockTickTimeDeltaType {
     timeScaleFlag 1 bslbf
     if(timeScaleFlag){
      timeScale 32  uimsbf
     }
     ptsDelta 32  uimsbf
    }
  • FIG. 19 is a flowchart illustrating a virtual world processing method according to example embodiments.
  • Referring to FIG. 19, the virtual world processing method may store a sensor capability related to capability of a sensor in operation 1110.
  • In operation 1120, a first value received from the sensor may be determined based on the sensor capability and a second value corresponding to the first value may be transmitted to the virtual world.
  • Depending on embodiments, the sensor capability may include a maximum value and a minimum value measurable by the sensor. When the first value is less than or equal to the maximum value and greater than or equal to the minimum value, the virtual world processing method may transmit the second value corresponding to the first value to the virtual world.
  • Depending on embodiments, the sensor capability may include a unit of the first value measured by the sensor. In addition, the sensor capability may include an offset value added to the first value measured by the sensor to obtain an absolute value. The sensor capability may further include a number of values measurable by the sensor. The sensor capability may further include a minimum input value required for the sensor to measure an output value. The sensor capability may further include an SNR of the sensor. The sensor capability may further include an error of the sensor. Additionally, the sensor capability may further include a position of the sensor.
  • The virtual world processing method may further include an operation (not shown) of storing a sensor adaptation preference for manipulation of the first value received from the sensor. The operation of transmitting the first value may include generating a third value from the first value based on the sensor capability and generating the second value from the third value based on the sensor adaptation preference.
  • Depending on embodiments, the sensor adaptation preference may include information on a method of applying the sensor adaptation preference to the first value. The sensor adaptation preference may further include information on whether to activate the sensor in the virtual world. The sensor adaptation preference may further include a unit of the second value used in the virtual world. The sensor adaptation preference may further include a maximum value and a minimum value of the second value used in the virtual world. In addition, the sensor adaptation preference may further include a number of the second values used in the virtual world.
  • FIG. 20 illustrates a flowchart of a virtual world processing method according to other example embodiments.
  • Referring to FIG. 20, the virtual world processing method may perform initial setting to be input with information of a real world from a sensor in operation 1210. Depending on embodiments, the initial setting may be an operation of activating the sensor.
  • The virtual world processing method may store a sensor capability as information on capability of the sensor and a sensor adaptation preference as information for manipulation of a value received from the sensor, in operation 1220.
  • The virtual world processing method may measure information on a motion, state, intention, shape, and the like of a user of the real world through eth sensor, in operation 1230. When the sensor is incapable of measuring the information, operation 1230 may be repeated until the information is measured.
  • When the information is measured through the sensor, preprocessing with respect to the information may be performed in operation 1240.
  • Also, the virtual world processing method may control the sensor using sensed information which is a command for controlling the sensor in operation 1250.
  • An adaptation RV may determine a first value received from the sensor based on the sensor capability and transmit a second value corresponding to the first value to a virtual world, in operation 1260. Depending on embodiments, a third value may be generated from the first value based on the sensor capability, the second value may be generated from the third value based on the sensor adaptation preference, and the second value may be transmitted to the virtual world.
  • FIG. 21 illustrates an operation of using a virtual world processing apparatus according to example embodiments.
  • Referring to FIG. 21, a user 1310 of a real world may input his or her intention through a sensor 1301. Depending on embodiments, the sensor 1301 may include a motion sensor configured to measure a motion of the user 1310 of the real world, and remote pointers attached to ends of arms and legs of the user 1310 and configured to measure directions and positions indicated by the ends of the arms and legs.
  • A sensor signal may be transmitted to the virtual world processing apparatus, the sensor signal which includes CI 1302 related to an arm opening motion, a still standing state, positions of hands and feet, an open angle of a hand, and the like of the user 1310.
  • Depending on embodiments, the CI 1302 may include a sensor capability, a sensor adaptation preference, and sensed information.
  • Depending on embodiments, the CI 1302 may include position information of the arms and the legs of the user 1310, expressed by Xreal, Yreal, and Zreal denoting values on an x-axis, y-axis, and z-axis and θXreal, θYreal, and θZreal denoting angles with respect to the x-axis, y-axis, and z-axis.
  • The virtual world processing apparatus may include an RV engine 1320. The RV engine 1320 may convert information of the real world to information applicable to a virtual world, using the CI 1302 included in the sensor signal.
  • Depending on embodiments, the RV engine 1320 may convert VWI 1303 using the CI 1302.
  • The VWI 1303 denotes information on the virtual world. For example, the VWI 1303 may include information on an object of the virtual world or elements constituting the object.
  • The VWI 1303 may include virtual world object information 1304 and avatar information 1305.
  • The virtual world object information 1304 denotes information on the object of the virtual world. The virtual world object information 1304 may include an object ID denoting ID information for recognizing identity of the object of the virtual world, and an object control and scale denoting information for controlling a state, size, and the like of the object of the virtual world.
  • Depending on embodiments, the virtual world processing apparatus may control the virtual world object information 1304 and the avatar information 1305 by a control command. The control command may include commands such as generation, disappearance, copy, and the like. The virtual world processing apparatus may generate the commands by selecting information to be manipulated from the virtual world object information 1304 and the avatar information 1305, along with the control command, and designating an ID corresponding to the selected information.
  • Source 56 denotes a method of constructing the control command using an XML. However, a program source shown in Source 56 is only an example embodiment but does not limit the present invention.
  • [Source 56]
    <!-- ################################################ -->
    <!-- Definition of Control command for Avatar and virtual object  -->
    <!-- ################################################ -->
    <complexType name=″ControlCommand″>
    <SimpleContent>
    <attribute name=″command″ type=″cid:commandType″
    use=″required″/>
    <attribute name=″Object″ type=″cid:ObjectType″
    use=″required″/>
    <attribute name=″ObjectID″ type=″ID″ use=″optional″/>
    </SimpleContent>
    </complexType>
    <simpleType name=″commandType″>
    <restriction base=″string″>
    <enumeration value=″Create″/>
    <enumeration value=″Remove″/>
    <enumeration value=″Copy″/>
    </restriction>
    </simpleType>
    <simpleType name=″ObjectType″>
    <restriction base=″string″>
    <enumeration value=″Avatar″/>
    <enumeration value=″VirtualObject>
    </restriction>
    </simpleType>
  • The RV engine 1320 may convert the VWI 1303 by applying information on the arm opening motion, the still standing state, the positions of hands and feet, the open angle of a hand, and the like, using the CI 1302.
  • The RV engine 1320 may transmit information 1306 on the converted VWI to the virtual world. The information 1306 on the converted VWI may include position information of arms and legs of an avatar of the virtual world, expressed by Xvirtual, Yvirtual, and Zvirtual denoting values on the x-axis, y-axis, and z-axis and θXvirtual, θYvirtual, and θZvirtual denoting angles with respect to the x-axis, y-axis, and z-axis. In addition, the information 1306 may include information on a size of the object of the virtual world, expressed by a scale(s,d,h)virtual denoting a width value, a height value, and a depth value of the object.
  • Depending on embodiments, in a virtual world 1330 of before transmission of the information 1306, the avatar is holding the object. In a virtual world 1340 of after transmission of the information 1306, since the arm opening motion, the still standing state, the positions of hands and feet, the open angle of a hand, and the like are reflected, the avatar of the virtual world may scale up the object.
  • That is, when the user 1310 of the real world makes a motion of holding and enlarging the object, the CI 1302 related to the arm opening motion, the still standing state, the positions of hands and feet, the open angle of a hand, and the like may be generated through the sensor 1301. Also, the RV engine 1320 may convert the CI 1302 related to the user 1310 of the virtual world, which is data measured in the real world, to the information applicable to the virtual world. The converted information may be applied to a structure of information related to the avatar and the object of the virtual world. Therefore, the motion of holding and enlarging the object may be reflected to the avatar, and the object may be enlarged.
  • Example embodiments include computer-readable media including program instructions to implement various operations embodied by a computer. The media may also include, alone or in combination with the program instructions, data files, data structures, tables, and the like. The media and program instructions may be those specially designed and constructed for the purposes of example embodiments, or they may be of the kind well known and available to those having skill in the computer software arts. Examples of computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD ROM disks; magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM). Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter. The described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described example embodiments, or vice versa.
  • Although a few example embodiments have been shown and described, the present disclosure is not limited to the described example embodiments. Instead, it would be appreciated by those skilled in the art that changes may be made to these example embodiments without departing from the principles and spirit of the disclosure, the scope of which is defined by the claims and their equivalents.

Claims (20)

What is claimed:
1. A sensor comprising:
an encoder configured to generate sensed information based on a predefined representation syntax,
wherein the predefined representation syntax defines elements, and flags corresponding to the elements, and
wherein the sensed information includes the flags corresponding to the elements, and at least one element corresponding to at least one of the flags having a predefined logic value.
2. The sensor of claim 1, wherein the elements comprise at least one of an ID element, a sensor ID reference element, a linked list element, a group ID element, a priority element, and an activation element.
3. The sensor of claim 2, wherein the elements further comprise a time stamp element.
4. The sensor of claim 1, when the sensor is a light sensor, wherein the elements comprise at least one of a value element, a unit element, and a color element.
5. The sensor of claim 1, when the sensor is an ambient noise sensor, wherein the elements comprise at least one of a life span element, a value element, and a unit element.
6. The sensor of claim 1, when the sensor is one of a temperature sensor, a humidity sensor, a distance sensor, an atmospheric pressure sensor, and a pressure sensor, wherein the elements comprise at least one of a value element, and a unit element.
7. The sensor of claim 1, when the sensor is a position sensor, wherein the elements comprise at least one of a position element, and a unit element.
8. The sensor of claim 1, when the sensor is a velocity sensor, wherein the elements comprise at least one of a velocity element, and a unit element.
9. The sensor of claim 1, when the sensor is an acceleration sensor, wherein the elements comprise at least one of an acceleration element, and a unit element.
10. The sensor of claim 1, when the sensor is an orientation sensor, wherein the elements comprise at least one of an orientation element, and a unit element.
11. The sensor of claim 1, when the sensor is an angular velocity sensor, wherein the elements comprise at least one of an angular velocity element, and a unit element.
12. The sensor of claim 1, when the sensor is an angular acceleration sensor, wherein the elements comprise at least one of an angular acceleration element, and a unit element.
13. The sensor of claim 1, when the sensor is a force sensor, wherein the elements comprise at least one of a force element, and a unit element.
14. The sensor of claim 1, when the sensor is a torque sensor, wherein the elements comprise at least one of a torque element, and a unit element.
15. The sensor of claim 1, when the sensor is a motion sensor, wherein the elements comprise at least one of a position element, an orientation element, a velocity element, an angular velocity element, an acceleration element, and an angular acceleration element.
16. The sensor of claim 1, when the sensor is an intelligent camera, wherein the elements comprise at least one of a facial animation ID element, a body animation ID element, a facial feature element, and a body feature element.
17. An apparatus for enabling interoperability between virtual world and real world, the apparatus comprising:
a processor comprising:
an adaptation unit configured to acquire sensed information of a predefined representation syntax from a sensor,
wherein the predefined representation syntax defines elements, and flags corresponding to the elements, and
wherein the sensed information includes the flags corresponding to the elements, and at least one element corresponding to at least one flag having a predefined logic value.
18. The apparatus of claim 17, wherein the elements comprise at least one of an ID element, a sensor ID reference element, a linked list element, a group ID element, a priority element, and an activation element.
19. The apparatus of claim 17, when the sensor is an intelligent camera, wherein the elements comprise at least one of a facial animation ID element, a body animation ID element, a facial feature element, and a body feature element.
20. The apparatus of claim 17, when the sensor is a motion sensor, wherein the elements comprise at least one of a position element, an orientation element, a velocity element, an angular velocity element, an acceleration element, and an angular acceleration element.
US14/825,784 2009-06-25 2015-08-13 Virtual world processing device and method Abandoned US20160030851A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/825,784 US20160030851A1 (en) 2009-06-25 2015-08-13 Virtual world processing device and method

Applications Claiming Priority (14)

Application Number Priority Date Filing Date Title
KR20090057313 2009-06-25
KR10-2009-0057313 2009-06-25
KR10-2009-0101434 2009-10-23
KR1020090101434A KR20100138702A (en) 2009-06-25 2009-10-23 Method and apparatus for processing virtual world
US25563609P 2009-10-28 2009-10-28
KR1020090104474A KR20100138706A (en) 2009-06-25 2009-10-30 Method and apparatus for processing virtual world
KR10-2009-0104474 2009-10-30
KR1020100003607A KR20100138722A (en) 2009-06-25 2010-01-14 Method and apparatus for processing virtual world
KR10-2010-0003607 2010-01-14
KR1020100006119A KR20100138725A (en) 2009-06-25 2010-01-22 Method and apparatus for processing virtual world
KR10-2010-0006119 2010-01-22
PCT/KR2010/004087 WO2010151051A2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method
US201213380743A 2012-03-09 2012-03-09
US14/825,784 US20160030851A1 (en) 2009-06-25 2015-08-13 Virtual world processing device and method

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/KR2010/004087 Continuation WO2010151051A2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method
US13/380,743 Continuation US9108106B2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method

Publications (1)

Publication Number Publication Date
US20160030851A1 true US20160030851A1 (en) 2016-02-04

Family

ID=43387048

Family Applications (4)

Application Number Title Priority Date Filing Date
US13/380,743 Active 2031-03-17 US9108106B2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method
US13/380,318 Active 2031-11-02 US9566506B2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method
US14/825,784 Abandoned US20160030851A1 (en) 2009-06-25 2015-08-13 Virtual world processing device and method
US15/429,811 Active US9987553B2 (en) 2009-06-25 2017-02-10 Virtual world processing device and method

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US13/380,743 Active 2031-03-17 US9108106B2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method
US13/380,318 Active 2031-11-02 US9566506B2 (en) 2009-06-25 2010-06-23 Virtual world processing device and method

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/429,811 Active US9987553B2 (en) 2009-06-25 2017-02-10 Virtual world processing device and method

Country Status (6)

Country Link
US (4) US9108106B2 (en)
EP (2) EP2446941A4 (en)
JP (3) JP6216512B2 (en)
KR (1) KR20100138725A (en)
CN (2) CN102481487B (en)
WO (2) WO2010151051A2 (en)

Families Citing this family (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007136745A2 (en) 2006-05-19 2007-11-29 University Of Hawaii Motion tracking system for real time adaptive imaging and spectroscopy
KR20100138725A (en) * 2009-06-25 2010-12-31 삼성전자주식회사 Method and apparatus for processing virtual world
KR20100138700A (en) * 2009-06-25 2010-12-31 삼성전자주식회사 Method and apparatus for processing virtual world
KR101685922B1 (en) * 2010-04-05 2016-12-13 삼성전자주식회사 Method and Apparatus for Processing Virtual World
KR101746453B1 (en) * 2010-04-12 2017-06-13 삼성전자주식회사 System and Method for Processing Sensory Effect
US9821224B2 (en) * 2010-12-21 2017-11-21 Microsoft Technology Licensing, Llc Driving simulator control with virtual skeleton
US20120192088A1 (en) * 2011-01-20 2012-07-26 Avaya Inc. Method and system for physical mapping in a virtual world
US10152116B2 (en) * 2011-04-26 2018-12-11 The Regents Of The University Of California Systems and devices for recording and reproducing senses
CN102707804B (en) * 2011-05-23 2015-03-25 中国科学院软件研究所 Acceleration transducer based character action control method
EP2747641A4 (en) 2011-08-26 2015-04-01 Kineticor Inc Methods, systems, and devices for intra-scan motion correction
US9717461B2 (en) 2013-01-24 2017-08-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9305365B2 (en) 2013-01-24 2016-04-05 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US10327708B2 (en) 2013-01-24 2019-06-25 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
EP2950714A4 (en) 2013-02-01 2017-08-16 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
EP3157422A4 (en) 2014-03-24 2018-01-24 The University of Hawaii Systems, methods, and devices for removing prospective motion correction from medical imaging scans
US10222953B2 (en) * 2014-04-30 2019-03-05 Disney Enterprises, Inc. Systems and methods for editing virtual content of a virtual space
CN106714681A (en) 2014-07-23 2017-05-24 凯内蒂科尔股份有限公司 Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
EP3189400A4 (en) * 2014-09-05 2018-07-04 Ballcraft, LLC Motion detection for portable devices
EP3099030A1 (en) * 2015-05-26 2016-11-30 Thomson Licensing Method and device for encoding/decoding a packet comprising data representative of a haptic effect
CN105118044B (en) * 2015-06-16 2017-11-07 华南理工大学 A kind of wheel shape cast article defect automatic testing method
WO2016203282A1 (en) 2015-06-18 2016-12-22 The Nielsen Company (Us), Llc Methods and apparatus to capture photographs using mobile devices
US9943247B2 (en) 2015-07-28 2018-04-17 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
WO2017091479A1 (en) 2015-11-23 2017-06-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
JP6311695B2 (en) 2015-12-16 2018-04-18 カシオ計算機株式会社 Autonomous mobile device, autonomous mobile method and program
JP6187623B1 (en) * 2016-03-14 2017-08-30 カシオ計算機株式会社 Autonomous mobile device, autonomous mobile method and program
US10092827B2 (en) * 2016-06-16 2018-10-09 Disney Enterprises, Inc. Active trigger poses
WO2018064965A1 (en) 2016-10-07 2018-04-12 Mediatek Inc. Method and apparatus for generating projection-based frame with 360-degree image content represented by triangular projection faces assembled in octahedron projection layout
WO2018064967A1 (en) * 2016-10-07 2018-04-12 Mediatek Inc. Video encoding method and apparatus with syntax element signaling of employed projection layout and associated video decoding method and apparatus
US10380715B2 (en) 2016-12-07 2019-08-13 Mediatek Inc. Method and apparatus for generating and encoding projection-based frame with 360-degree content represented by triangular projection faces packed in octahedron projection layout
US20180285644A1 (en) * 2017-03-29 2018-10-04 Electronics And Telecommunications Research Institute Sensor information processing method and system between virtual world and real world
US10154360B2 (en) * 2017-05-08 2018-12-11 Microsoft Technology Licensing, Llc Method and system of improving detection of environmental sounds in an immersive environment
US10948450B2 (en) 2018-07-23 2021-03-16 Genmark Diagnostics, Inc. Electrochemical measurements of components in coatings
CN112805750A (en) 2018-08-13 2021-05-14 奇跃公司 Cross-reality system
KR20200019486A (en) 2018-08-14 2020-02-24 삼성전자주식회사 Electronic apparatus, contorl method thereof and electronic system
WO2020072985A1 (en) 2018-10-05 2020-04-09 Magic Leap, Inc. Rendering location specific virtual content in any location
GB2586059B (en) * 2019-08-01 2023-06-07 Sony Interactive Entertainment Inc System and method for generating user inputs for a video game
EP4046401A4 (en) 2019-10-15 2023-11-01 Magic Leap, Inc. Cross reality system with wireless fingerprints
CN114600064A (en) 2019-10-15 2022-06-07 奇跃公司 Cross reality system with location services
CN110633013B (en) * 2019-10-17 2022-12-20 于生明 Control method of all-directional universal mobile platform
WO2021096931A1 (en) 2019-11-12 2021-05-20 Magic Leap, Inc. Cross reality system with localization service and shared location-based content
CN114762008A (en) 2019-12-09 2022-07-15 奇跃公司 Simplified virtual content programmed cross reality system
CN115398314A (en) 2020-02-13 2022-11-25 奇跃公司 Cross reality system for map processing using multi-resolution frame descriptors
EP4103910A4 (en) 2020-02-13 2024-03-06 Magic Leap Inc Cross reality system with accurate shared maps
WO2021163295A1 (en) 2020-02-13 2021-08-19 Magic Leap, Inc. Cross reality system with prioritization of geolocation information for localization
WO2021173779A1 (en) 2020-02-26 2021-09-02 Magic Leap, Inc. Cross reality system with fast localization
JP2023524446A (en) 2020-04-29 2023-06-12 マジック リープ, インコーポレイテッド Cross-reality system for large-scale environments
CN112842670B (en) * 2021-01-28 2022-07-08 吉林大学第一医院 Skin nursing device for nephrotic patient with obvious edema

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5442147A (en) * 1991-04-03 1995-08-15 Hewlett-Packard Company Position-sensing apparatus
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
US5597373A (en) * 1991-11-08 1997-01-28 Cedaron Medical, Inc. Physiological evaluation and exercise system
US6120461A (en) * 1999-08-09 2000-09-19 The United States Of America As Represented By The Secretary Of The Army Apparatus for tracking the human eye with a retinal scanning display, and method thereof
US6137403A (en) * 1998-12-10 2000-10-24 Phoenix Controls Corporation Sash sensor and method of sensing a sash using an array of multiplexed elements
US20020111773A1 (en) * 2000-07-13 2002-08-15 Feola Christopher J. System and method for associating historical information with sensory data and distribution thereof
US6542081B2 (en) * 1996-08-19 2003-04-01 William C. Torch System and method for monitoring eye movement
US6792165B1 (en) * 1999-05-25 2004-09-14 Silverbrook Research Pty Ltd Sensing device
US20070210916A1 (en) * 2006-03-07 2007-09-13 Munoru Ogushi Sensor Network System, Base Station, and Method to Forward Measured Data
US20070263699A1 (en) * 2006-05-09 2007-11-15 Thermal Solutions, Inc. Magnetic element temperature sensors
US20100045816A1 (en) * 1999-05-19 2010-02-25 Rhoads Geoffrey B User Feedback in Connection with Object Recognition
USRE41376E1 (en) * 1996-08-19 2010-06-15 Torch William C System and method for monitoring eye movement
US9317110B2 (en) * 2007-05-29 2016-04-19 Cfph, Llc Game with hand motion control

Family Cites Families (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2026082A (en) * 1935-08-31 1935-12-31 Parker Brothers Inc Board game apparatus
US5330980A (en) * 1992-08-25 1994-07-19 Shin Jae S Medicament for the topical treatment of skin
US5577981A (en) * 1994-01-19 1996-11-26 Jarvik; Robert Virtual reality exercise machine and computer controlled video system
RU2109336C1 (en) * 1995-07-14 1998-04-20 Нурахмед Нурисламович Латыпов Method and device for immersing user into virtual world
AU2978397A (en) 1996-06-05 1998-01-05 Sega Enterprises, Ltd. Image processor for games
KR100221335B1 (en) * 1996-12-26 1999-09-15 전주범 Transmission system of glove sensor value in a virtual reality system and method thereof
KR100289860B1 (en) 1997-07-17 2001-05-15 나카무라 하사오 Croman's manufacturing method
EP1852823B1 (en) * 1999-12-28 2011-07-27 Sony Corporation Signal processing method and apparatus and recording medium
JP2002159004A (en) * 2000-04-21 2002-05-31 Sony Corp Device and method for encoding, recording medium, and program
JP2002305758A (en) * 2001-04-04 2002-10-18 Nippon Telegr & Teleph Corp <Ntt> Device and system for presence communication
US7149691B2 (en) * 2001-07-27 2006-12-12 Siemens Corporate Research, Inc. System and method for remotely experiencing a virtual environment
KR100452917B1 (en) * 2001-10-09 2004-10-14 주식회사 제노프릭스 Method and System for sensing Three-Dimensional body motion using color marker
US7225343B1 (en) * 2002-01-25 2007-05-29 The Trustees Of Columbia University In The City Of New York System and methods for adaptive model generation for detecting intrusions in computer systems
US20040001734A1 (en) * 2002-02-07 2004-01-01 Burrell James W. Virtual keyboard and control means
JP4447823B2 (en) * 2002-06-14 2010-04-07 ソニー株式会社 Portable information equipment
CN1206007C (en) * 2002-10-31 2005-06-15 威盛电子股份有限公司 Pressure-sensitive virtual running machine
JP4144374B2 (en) * 2003-02-25 2008-09-03 ソニー株式会社 Image processing apparatus and method, recording medium, and program
US8566292B2 (en) * 2003-04-14 2013-10-22 Schneider Electric It Corporation Method and system for journaling and accessing sensor and configuration data
US7217224B2 (en) * 2003-08-14 2007-05-15 Tom Thomas Virtual exercise system and method
US7752544B2 (en) 2003-11-17 2010-07-06 International Business Machines Corporation Method, system, and apparatus for remote interactions
US20050168485A1 (en) * 2004-01-29 2005-08-04 Nattress Thomas G. System for combining a sequence of images with computer-generated 3D graphics
DE102004012072A1 (en) * 2004-03-12 2005-09-29 Robert Bosch Gmbh tracking device
CN1973508B (en) * 2004-06-30 2010-09-08 诺基亚公司 Transfer of data objects
JP2006065518A (en) 2004-08-25 2006-03-09 Asahi Kasei Electronics Co Ltd Pointing device and signal processing method
US20060262120A1 (en) * 2005-05-19 2006-11-23 Outland Research, Llc Ambulatory based human-computer interface
US20070044539A1 (en) * 2005-03-01 2007-03-01 Bryan Sabol System and method for visual representation of a catastrophic event and coordination of response
US8239162B2 (en) * 2006-04-13 2012-08-07 Tanenhaus & Associates, Inc. Miniaturized inertial measurement unit and associated methods
US8156232B2 (en) * 2005-09-12 2012-04-10 Rockwell Automation Technologies, Inc. Network communications in an industrial automation environment
JP4794957B2 (en) * 2005-09-14 2011-10-19 任天堂株式会社 GAME PROGRAM, GAME DEVICE, GAME SYSTEM, AND GAME PROCESSING METHOD
JP2007156779A (en) * 2005-12-05 2007-06-21 Hitachi Ltd Sensor network system, base station and relay method for sensing data
US20070132722A1 (en) * 2005-12-08 2007-06-14 Electronics And Telecommunications Research Institute Hand interface glove using miniaturized absolute position sensors and hand interface system using the same
KR100838181B1 (en) * 2005-12-08 2008-06-16 한국전자통신연구원 Hand interface device using miniaturized absolute position sensors, system for hand interface using it
JP2007206746A (en) * 2006-01-30 2007-08-16 Canon Inc Information processor, and control method, program and storage medium therefor
JP5196515B2 (en) 2006-04-25 2013-05-15 任天堂株式会社 GAME PROGRAM, GAME DEVICE, GAME SYSTEM, AND GAME PROCESSING METHOD
EP2013865A4 (en) * 2006-05-04 2010-11-03 Sony Comp Entertainment Us Methods and apparatus for applying gearing effects to input based on one or more of visual, acoustic, inertial, and mixed data
CN102580314B (en) * 2006-05-04 2015-05-20 美国索尼电脑娱乐公司 Obtaining input for controlling execution of a game program
JP5089079B2 (en) * 2006-05-08 2012-12-05 任天堂株式会社 Program, information storage medium, and image generation system
KR100770890B1 (en) * 2006-08-18 2007-10-26 삼성전자주식회사 Device and method for detecting file type in digital video broadcasting terminal
US7937349B2 (en) * 2006-11-09 2011-05-03 Pucher Max J Method for training a system to specifically react on a specific input
US20080310707A1 (en) * 2007-06-15 2008-12-18 Microsoft Corporation Virtual reality enhancement using real world data
JP4916390B2 (en) * 2007-06-20 2012-04-11 任天堂株式会社 Information processing program, information processing apparatus, information processing system, and information processing method
KR20090002292A (en) * 2007-06-26 2009-01-09 삼성전자주식회사 Apparatus and method for synchronizing virtual character with music and sharing it through xml based language
US8812261B2 (en) * 2007-08-23 2014-08-19 Applied Materials, Inc. Method and apparatus to automatically create virtual sensors with templates
US9387402B2 (en) * 2007-09-18 2016-07-12 Disney Enterprises, Inc. Method and system for converting a computer virtual environment into a real-life simulation environment
JP2009140107A (en) * 2007-12-04 2009-06-25 Sony Corp Input device and control system
US8386918B2 (en) * 2007-12-06 2013-02-26 International Business Machines Corporation Rendering of real world objects and interactions into a virtual universe
CA2714534C (en) * 2008-02-28 2018-03-20 Kenneth Perlin Method and apparatus for providing input to a processor, and a sensor pad
JP4571198B2 (en) * 2008-03-07 2010-10-27 京セラ株式会社 Mobile communication terminal
EP2291829A1 (en) * 2008-05-09 2011-03-09 Koninklijke Philips Electronics N.V. System and method for processing application logic of a virtual and a real-world ambient intelligence environment
US20110132089A1 (en) * 2008-09-08 2011-06-09 Hitachi, Ltd. Inertial Sensor
JP5334595B2 (en) * 2009-01-09 2013-11-06 任天堂株式会社 GAME DEVICE AND GAME PROGRAM
KR20100138725A (en) * 2009-06-25 2010-12-31 삼성전자주식회사 Method and apparatus for processing virtual world

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5442147A (en) * 1991-04-03 1995-08-15 Hewlett-Packard Company Position-sensing apparatus
US5597373A (en) * 1991-11-08 1997-01-28 Cedaron Medical, Inc. Physiological evaluation and exercise system
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
USRE41376E1 (en) * 1996-08-19 2010-06-15 Torch William C System and method for monitoring eye movement
US6542081B2 (en) * 1996-08-19 2003-04-01 William C. Torch System and method for monitoring eye movement
US6137403A (en) * 1998-12-10 2000-10-24 Phoenix Controls Corporation Sash sensor and method of sensing a sash using an array of multiplexed elements
US20100045816A1 (en) * 1999-05-19 2010-02-25 Rhoads Geoffrey B User Feedback in Connection with Object Recognition
US6792165B1 (en) * 1999-05-25 2004-09-14 Silverbrook Research Pty Ltd Sensing device
US6120461A (en) * 1999-08-09 2000-09-19 The United States Of America As Represented By The Secretary Of The Army Apparatus for tracking the human eye with a retinal scanning display, and method thereof
US20020111773A1 (en) * 2000-07-13 2002-08-15 Feola Christopher J. System and method for associating historical information with sensory data and distribution thereof
US20070210916A1 (en) * 2006-03-07 2007-09-13 Munoru Ogushi Sensor Network System, Base Station, and Method to Forward Measured Data
US20070263699A1 (en) * 2006-05-09 2007-11-15 Thermal Solutions, Inc. Magnetic element temperature sensors
US9317110B2 (en) * 2007-05-29 2016-04-19 Cfph, Llc Game with hand motion control

Also Published As

Publication number Publication date
CN102481486A (en) 2012-05-30
JP2012531658A (en) 2012-12-10
JP6216512B2 (en) 2017-10-18
WO2010151054A3 (en) 2011-04-21
US9566506B2 (en) 2017-02-14
US20120188237A1 (en) 2012-07-26
JP2012531657A (en) 2012-12-10
KR20100138725A (en) 2010-12-31
US9987553B2 (en) 2018-06-05
WO2010151051A3 (en) 2011-04-21
CN102481486B (en) 2015-05-20
EP2446942A4 (en) 2017-06-21
CN102481487B (en) 2016-05-11
CN102481487A (en) 2012-05-30
EP2446942A2 (en) 2012-05-02
JP2016137248A (en) 2016-08-04
EP2446941A2 (en) 2012-05-02
WO2010151051A2 (en) 2010-12-29
US9108106B2 (en) 2015-08-18
EP2446941A4 (en) 2014-03-19
US20120191737A1 (en) 2012-07-26
WO2010151054A2 (en) 2010-12-29
US20170246532A1 (en) 2017-08-31

Similar Documents

Publication Publication Date Title
US9108106B2 (en) Virtual world processing device and method
US9374087B2 (en) Apparatus and method for processing virtual world
US9612737B2 (en) Device and method for processing virtual worlds
KR101686139B1 (en) Method and Apparatus for Processing Virtual World
Gajjar Mobile sensors and context-aware computing
Jølstad et al. Validity of the AdMos, advanced sport instruments, GNSS sensor for use in alpine skiing
CN113282171B (en) Oracle text augmented reality content interaction system, method, equipment and terminal
KR20110108060A (en) Method and apparatus for processing virtual world
Tehrani et al. Interactive kinect using mobile phone for education
Yusof et al. Mathematics Lesson using Accelerometer Sensor Interaction in Handheld Augemented Reality Application for Kindergarten
CN110944085A (en) Method for generating music by shaking smart phone
CN116899199A (en) Dance training method, device, equipment and readable storage medium
Willim Walking through the screen: digital media on the go
Wisman et al. Mobile science

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION