US20130038548A1 - Touch system - Google Patents
Touch system Download PDFInfo
- Publication number
- US20130038548A1 US20130038548A1 US13/566,151 US201213566151A US2013038548A1 US 20130038548 A1 US20130038548 A1 US 20130038548A1 US 201213566151 A US201213566151 A US 201213566151A US 2013038548 A1 US2013038548 A1 US 2013038548A1
- Authority
- US
- United States
- Prior art keywords
- touch
- area
- operation area
- touch position
- coordinate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000001514 detection method Methods 0.000 claims abstract description 27
- 230000010365 information processing Effects 0.000 claims description 16
- 230000004044 response Effects 0.000 claims description 16
- 230000008859 change Effects 0.000 claims description 6
- 230000007246 mechanism Effects 0.000 claims description 2
- 238000000034 method Methods 0.000 description 13
- 238000006243 chemical reaction Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- TVEXGJYMHHTVKP-UHFFFAOYSA-N 6-oxabicyclo[3.2.1]oct-3-en-7-one Chemical compound C1C2C(=O)OC1C=CC2 TVEXGJYMHHTVKP-UHFFFAOYSA-N 0.000 description 1
- 229920000877 Melamine resin Polymers 0.000 description 1
- 239000004640 Melamine resin Substances 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 230000005057 finger movement Effects 0.000 description 1
- 238000009413 insulation Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 229920003002 synthetic resin Polymers 0.000 description 1
- 239000000057 synthetic resin Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/044—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
- G06F3/0446—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means using a grid-like structure of electrodes in at least two directions, e.g. using row and column electrodes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
Abstract
In a touch table system having a touch table apparatus provided with a touch panel main body in a tabletop and a PC connected to the touch table apparatus, the touch table apparatus has a touch position detector detecting a touch position within a touch detection area. A touch position converter converts a coordinate of a touch position into a coordinate of a screen area of the PC, the touch position being obtained in an operation is set for each user within the touch detection area.
Description
- The present application claims priority under 35 U.S.C. §119 of Japanese Application No. 2011-176536 filed on Aug. 12, 2011, the disclosure of which is expressly incorporated by reference herein in its entirety.
- 1. Field of the Invention
- The present invention relates to a touch system having a touch support member apparatus provided with a touch screen.
- 2. Description of Related Art
- In a meeting where a screen of a PC is displayed on a large screen, an attendee uses a position input device, such as a mouse or a tablet, to operate the screen of the PC. In a case where one position input device is shared by a plurality of attendees, the attendees cannot readily operate the screen of the PC. Preparing a position input device for exclusive use for each of a plurality of attendees allows them to readily operate the screen of the PC. It is cumbersome, however, to prepare a large number of position input devices.
- Thus, there is demand for a system that allows all attendees to readily operate a PC without providing exclusive position input devices to all the attendees. In connection with such a demand, a known technology is directed to a touch table apparatus having a touch screen in a tabletop (refer to Related Art 1). With such a touch table apparatus, users around the touch table apparatus can readily operate a screen of a PC.
- To use a conventional touch table apparatus in a meeting, the touch table apparatus should have a size similar to a regular meeting table. With such a size of the touch table apparatus, however, it is sometimes difficult to reach a desired position on a touch surface of a tabletop while seated. In this case, a user needs to stand up and move from the user's seat to operate the screen, causing inconvenience.
- [Related Art 1] Japanese Patent Laid-open Publication No. 2007-108678
- In view of the above circumstances, an advantage of the present invention is to provide a touch system configured to enhance convenience of use by a plurality of users.
- A touch system comprising: a touch support member apparatus having a touch surface on which a touch operation is performed by a user and on which electrodes are arranged in a grid shape; and an information processing apparatus connected to the touch support member apparatus. The touch support member apparatus comprises: a touch position detector configured to detect a touch position on an operation area of the touch surface based on a change of output signals from the electrodes associated with a change in capacitance in response to the touch operation; and a touch position converter configured to convert a coordinate of the touch position, in the operation area, obtained by the touch position detector, into a coordinate of a screen area of the information processing apparatus
- According to the present invention, the user has an operation area on the touch surface, thus enhancing convenience.
- The present invention is further described in the detailed description which follows, in reference to the noted plurality of drawings by way of non-limiting examples of exemplary embodiments of the present invention, in which like reference numerals represent similar parts throughout the several views of the drawings, and wherein:
-
FIG. 1 illustrates an overall configuration of a touch table system according to an embodiment of the present invention; -
FIG. 2 is a perspective view illustrating an example of use of the touch table system; -
FIG. 3 is a cross-sectional view of a panel main body incorporated in a tabletop of a touch table apparatus; -
FIGS. 4A and 4B each illustrate a state in which an operation area is set for each user to operate a screen; -
FIGS. 5A and 5B each illustrate a state in which an operation area is set for each user to operate a screen in another example; -
FIG. 6 illustrates two-finger operation in which two fingers are used for position input operation; -
FIG. 7 illustrates a state in which an operation area is designated on the touch table apparatus; -
FIG. 8 is a perspective view of an area designation tool; -
FIGS. 9A and 9B each illustrate a state in which an area is designated using the area designation tool; -
FIG. 10 is a functional block diagram of the touch table apparatus and a PC; -
FIG. 11 is a flowchart illustrating processing procedures in the touch table apparatus and the PC; -
FIG. 12 is a flowchart illustrating processing procedures for operation area designation shown in a portion A ofFIG. 11 ; -
FIGS. 13A to 13D each illustrate a screen displayed on a display during operation area designation; -
FIG. 14 is a flowchart illustrating processing procedures for screen operation shown in a portion B ofFIG. 11 ; -
FIGS. 15A and 15B each illustrate a state of coordinate conversion during screen operation; -
FIG. 16 is a perspective view illustrating another example of use of the touch table system; and -
FIG. 17 is a perspective view illustrating yet another example of use of the touch table system. - The particulars shown herein are by way of example and for purposes of illustrative discussion of the embodiments of the present invention only and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects of the present invention. In this regard, no attempt is made to show structural details of the present invention in more detail than is necessary for the fundamental understanding of the present invention, the description taken with the drawings making apparent to those skilled in the art how the forms of the present invention may be embodied in practice.
- Embodiments of the present invention are described below with reference to the drawings.
-
FIG. 1 illustrates an overall configuration of a touch table system according to an embodiment.FIG. 2 is a perspective view illustrating an example of use of the touch table system.FIG. 3 is a cross-sectional view of a panelmain body 5 incorporated in a tabletop of atouch table apparatus 1. - With reference to
FIG. 1 , the touch table system includes thetouch table apparatus 1, a PC (information processing apparatus) 2, and a display (display apparatus) 3. - The touch panel
main body 5 of thetouch table apparatus 1 has atouch surface 6 on which a touch operation is performed by a pointing object (conductive body, such as a user's finger or a stylus). The touch panelmain body 5 includes a plurality of transmittingelectrodes 7 in parallel to one another and a plurality of receivingelectrodes 8 in parallel to one another, which are disposed in a grid pattern. With reference toFIG. 2 , the touch panelmain body 5 is disposed in atabletop 12 of thetouch table apparatus 1. An upper surface of thetabletop 12 serves as thetouch surface 6 on which users A to D perform touch operations. - In the example of
FIG. 2 , thedisplay 3 and thePC 2 are mounted on astand 13 disposed beside thetouch table apparatus 1. The users A to D seated around thetouch table apparatus 1 each perform a touch operation on thetouch table apparatus 1 while watching a screen of thedisplay 3, and thereby operate a screen of the PC 2. A small footprint PC integrated with a display may be mounted on thetabletop 12 of thetouch table apparatus 1. - With reference to
FIG. 3 , the touch panelmain body 5 has anelectrode sheet 15 including the transmittingelectrodes 7 and the receivingelectrodes 8, afront protection member 16 disposed on a front surface of theelectrode sheet 15, and arear projection member 17 disposed on a rear surface of theelectrode sheet 15. In theelectrode sheet 15, the transmittingelectrodes 7 and the receivingelectrodes 8 are disposed on front and rear surfaces, respectively, of asupport sheet 18 that provides insulation between the transmittingelectrodes 7 and the receivingelectrodes 8. Thefront protection member 16 has thetouch surface 6 on which a touch operation is performed by a pointing object, such as a finger. In order to increase detection sensitivity to touch operation by a pointing object, thefront protection member 16 is composed of a synthetic resin material having high permittivity, such as, for example, a melamine resin. - As shown in
FIG. 1 , thetouch table apparatus 1 has atransmitter 9, areceiver 10, and acontroller 11. Thetransmitter 9 applies a drive signal to the transmittingelectrode 7. Thereceiver 10 receives a response signal from the receivingsignal 8 that responds to the drive signal applied to the transmittingelectrode 7 and outputs a level signal at each electrode intersection where the transmittingelectrode 7 and the receivingelectrode 8 intersect with each other. Thecontroller 11 detects a touch position based on the level signal output from thereceiver 10 and controls operations of thetransmitter 9 and thereceiver 10. - The transmitting
electrode 7 and the receivingelectrode 8 intersect in a stacked state with an insulating layer therebetween. A capacitor is formed at the electrode intersection where the transmittingelectrode 7 and the receivingelectrode 7 intersect. A pointing object, such as a finger, approaches or comes into contact with thetouch surface 6 as a user performs a touch operation with the pointing object. Then, the capacitance at the electrode intersection is substantially reduced, thus allowing detection of the touch operation. - A mutual capacitance system is employed herein. A drive signal is applied to the transmitting
electrode 7, and then a charge-discharge current flows to the receivingelectrode 8 in response. The charge-discharge current is output from the receivingelectrode 8 as a response signal. A variation in the capacitance at the electrode intersection at this time in response to a user's touch operation varies the response signal of the receivingelectrode 8. A touch position is calculated based on the variation amount. In this mutual capacitance system, a level signal obtained from signal processing of the response signal in thereceiver 10 is output for each electrode intersection of the transmittingelectrode 7 and the receivingelectrode 8, thus enabling what is commonly-called multi-touch (multiple point detection), which simultaneously detects a plurality of touch positions. Of course, other systems can be utilized, and are within the scope of the instant disclosure. - The
transmitter 9 selects the transmittingelectrodes 7 one by one and applies drive signals. Thereceiver 10 selects the receivingelectrodes 8 one by one and converts response signals of the receivingelectrodes 8 into analog signals and then into digital signals for output. Thetransmitter 9 and thereceiver 10 operate in response to a synchronization signal output from thecontroller 11. During a time when thetransmitter 9 applies a drive signal to one transmittingelectrode 7, thereceiver 10 selects the receivingelectrodes 8 one by one and sequentially processes response signals from the receivingelectrodes 8. Sequentially repeating this scanning of one line for all transmittingelectrodes 7 provides a level signal at every electrode intersection. - The
controller 11 obtains a touch position (center coordinate of a touch area) based on predetermined calculation of a level signal at each electrode intersection output from thereceiver 10. In this touch position calculation, a touch position is calculated by a predetermined interpolating method (e.g., centroid method) from a level signal of each of a plurality of adjacent electrode intersections (e.g., 4×4) in the X direction (array direction of the receiving electrodes 8) and the Y direction (array direction of the transmitting electrodes 7). Thereby, the touch position can be detected at a higher resolution (e.g., 1 mm or less) than the placement pitch (e.g., 10 mm) of the transmittingelectrodes 7 and the receivingelectrodes 8. - The
controller 11 also obtains a touch position every frame period in which reception of a level signal at each electrode intersection is completed across thetouch surface 6 and outputs the touch position information to thePC 2 in units of frames. Based on the touch position information of a plurality of temporally continuing frames, thePC 2 generates and outputs to thedisplay 3, display screen data of touch positions connected in time series. In a case where touch operations are simultaneously performed at a plurality of positions, the touch position information including the plurality of touch positions is output in units of frames. -
FIGS. 4A and 4B each illustrate a state in whichoperation areas 22 a to 22 d are set for the users A to D, respectively, for screen operation.FIG. 4A illustrates thetouch table apparatus 1 on which the users A to D perform screen operations.FIG. 4B illustrates a screen displayed on thedisplay 3. - In the present embodiment, the
operation areas 22 a to 22 d for the users A to D, respectively, are individually set within atouch detection area 21 of the touch panelmain body 5. Thus, a position input device is virtually assigned exclusively for each of the users A to D. With theoperation areas 22 a to 22 d set for the users A to D, respectively, within reach, the users A to D each can perform a position input operation on the entire screen without moving from their seats, thus enhancing convenience. - In the
operation areas 22 a to 22 d, the users perform touch operations to operate the screen, specifically, to move a pointer (cursor) on the screen, to select a button on the screen, and to draw a line.FIGS. 4A and 4B each illustrate an example in which a line is drawn in a hand-writing mode. The users A to D move their fingers in theoperation areas 22 a to 22 d, respectively, as shown inFIG. 4A . Then, lines associated with the finger movements of the respective users A to D are displayed together on the screen ofdisplay 3, as shown inFIG. 4B . - In the present embodiment, in a case where a touch position is not included in any of the
operation areas 22 a to 22 d, specifically, a touch position is out of theoperation areas 22 a to 22 d, the touch position is processed as invalid. Thus, a position input operation cannot be performed outside theoperation areas 22 a to 22 d. Furthermore, even when the users A to D place their hands or an object outside theoperation areas 22 a to 22 d, erroneous detection as a touch position can be prevented, thus improving usability. -
FIGS. 5A and 5B each illustrate a state in which theoperation areas 22 a to 22 d are set for the users A to D, respectively, for screen operation in another example.FIG. 5A illustrates thetouch table apparatus 1 on which the users A to D perform screen operations.FIG. 5B illustrates a screen displayed on thedisplay 3. - In the present embodiment, each of the operation areas can be set to an absolute coordinate mode or a relative coordinate mode according to a coordinate mode selected by each of the users A to D, the absolute coordinate mode outputting a coordinate of a touch position with an absolute coordinate, the relative coordinate mode outputting a coordinate of a touch position with a relative coordinate. In the example of
FIGS. 5A and 5B , theoperation areas 22 a to 22 c of the users A to C, respectively, are set to the absolute coordinate mode and theoperation area 22 d of the user D is set to the relative coordinate mode. - In the absolute coordinate mode, the
operation areas 22 a to 22 c each correspond to the entire screen area, similar to a tablet, and a coordinate value indicating an absolute position on each of theoperation areas 22 a to 22 c is output. In the relative coordinate mode, a coordinate value indicating a position relative to a position pointed immediately prior thereto is output, similar to a mouse. - Since the absolute coordinate mode or the relative coordinate mode can be set separately for each of the operation areas, the absolute coordinate mode or the relative coordinate mode can be selected depending on user's needs, thus improving convenience.
- It is basically unnecessary to set an operation area in particular in the relative coordinate mode. Without a boundary of an operation area, however, erroneous detection of a user's hand or an object placed on the
touch surface 6 cannot be prevented, causing inconvenience. Thus, it is preferable to set an operation area even in the relative coordinate mode. -
FIG. 6 illustrates two-finger operation mode in which two fingers are used for position input operation. In the present embodiment, a user keeps a first finger F1 still (or stationary) in contact with thetouch surface 6 and moves a second finger F2 to enter a position. Based on a relative position of the second finger F2 to the still first finger F1, a coordinate value of the touch position is output with a relative coordinate. - In the example of
FIG. 6 , the two fingers of one hand are used. Alternatively, one finger of each of the hands may be used. -
FIG. 7 illustrates a state in which anoperation area 22 is designated on thetouch table apparatus 1. To designate theoperation area 22, two diagonal vertexes (upper left and lower right herein) that define therectangular operation area 22 are designated by touch operations. Thus, therectangular operation area 22 is defined such that the two vertexes are passed or intersected and four sides are provided in parallel to each side of thetouch detection area 21. - The operation area is designated by touch operations by a user as above. Alternatively, an area designation tool may be used to designate an operation area as described below.
FIG. 8 is a perspective view of anarea designation tool 31.FIGS. 9A and 9B each illustrate a state in which an area is designated using thearea designation tool 31.FIG. 9A illustrates a state in which thearea designation tool 31 is placed on thetouch table apparatus 1.FIG. 9B illustrates a touch area that appears within a touch detection area. - With reference to
FIG. 8 , thearea designation tool 31, which has a rectangular shape to define an operation area thereinside, is extendable and contractable on each side with a telescopic mechanism so as to change the size. Specifically, thearea designation tool 31 has anangular member 32,side members side members angular member 32 having an L shape and a large diameter or cross-section is positioned at a corner portion. Theside members angular member 32. Theside member side members angular members 32 of thearea designation tool 31, at least two diagonally positioned members are formed of conductive bodies. - The
area designation tool 31 is placed on thetouch surface 6 of thetouch table apparatus 1 as shown inFIG. 9A . Then, an L-shapedtouch area 37 is detected based on the position of theangular member 32 formed of a conductive body as shown inFIG. 9B . Thus, it is detected that thearea designation tool 31 is placed or positioned on the touch surface. Then, anangular point 38 of the L-shapedtouch area 37 is set as each of two diagonal vertexes that define therectangular operation area 22, and thus theoperation area 22 is determined. - A user can perform a touch operation on the
touch surface 6 inside thearea designation tool 31 as shown inFIG. 9A . Since theoperation area 22 is partitioned by thearea designation tool 31, the user can visually confirm a range of theoperation area 22. The user can thus prevent the inconvenience of being unsure of a range of theoperation area 22 after designating theoperation area 22 by touch operations, as in the case ofFIG. 7 , thus improving convenience. - A configuration associated with the operation area of the
touch table apparatus 1 and thePC 2 is explained below. Operation procedures of thetouch table apparatus 1 and thePC 2 are also explained. -
FIG. 10 is a functional block diagram of thetouch table apparatus 1 and thePC 2. Thecontroller 11 of thetouch table apparatus 1 has atouch position detector 41, atouch position converter 42, and a transmitter/receiver 48. Thetouch position detector 41 detects a touch position within thetouch detection area 21 of the touch panelmain body 5, based on a level signal output from thereceiver 10. In a case where users perform touch operations simultaneously, a plurality of touch positions are detected simultaneously. Thetouch position detector 41 outputs a coordinate value of a touch position in a coordinate system of the touch table. A touch position obtained by thetouch position detector 41 during operation area designation is directly transmitted from the transmitter/receiver 48 to thePC 2. - The
touch position converter 42 converts a touch position obtained by thetouch position detector 41 into a touch position of each operation area and outputs the converted touch position. In particular, thetouch position converter 42 converts a coordinate of a touch position obtained in the operation area for each user set within the touch detection area of thetouch table apparatus 1 into a coordinate in the screen area of thePC 2. Thetouch position converter 42 has anoperation area memory 43, anoperation area determinator 44, and a coordinateconverter 45. - The
operation area memory 43 stores information (coordinate value) on the position of the operation area set within thetouch detection area 21, the information being transmitted from thePC 2 and being received by the transmitter/receiver 48. Based on the information on the operation area stored in theoperation area memory 43, theoperation area determinator 44 determines in which operation area a touch position obtained by thetouch position detector 41 is included. When the touch position is not included in any operation area, specifically, when the touch position is located outside the operation area, theoperation area determinator 44 invalidates the touch position. Based on the information on the operation area stored in theoperation area memory 43, the coordinateconverter 45 converts a coordinate value of the touch position obtained by thetouch position detector 41 from a coordinate system of the touch table to a coordinate system of an output screen (e.g., display 3) of thePC 2. The converted coordinate value of the touch position by the coordinateconverter 45 is transmitted from the transmitter/receiver 48 to thePC 2 along with an ID (identification information) of the operation area obtained by theoperation area determinator 44. - When the
touch position converter 42 detects that two fingers F1 and F2 touch simultaneously as shown inFIG. 6 , thetouch position converter 42 switches to a two-finger operation mode to output a coordinate value of a touch position with a relative coordinate, based on a relative position of the second finger F2 to the still or stationary first finger F1. - The
PC 2 has anoperation area setter 46, ascreen operation processor 47, and a transmitter/receiver 49. Theoperation area setter 46 sets an operation area within the touch detection area individually for each user, based on a touch position obtained by thetouch position detector 41 of thetouch table apparatus 1 during operation area designation and received by the transmitter/receiver 49. Information on the position of the operation area obtained herein is transmitted from the transmitter/receiver 49 to thetouch table apparatus 1 and is stored in theoperation area memory 43 of thetouch table apparatus 1. - The
screen operation processor 47 reflects an operation performed in the operation area of each user in the same screen area, based on a coordinate of the screen area obtained by thetouch position converter 42 during screen operation and received by the transmitter/receiver 49. Thescreen operation processor 47 perfoi ins processing corresponding to touch operations to operate the screen by a user, specifically, to move a pointer (cursor) on the screen, to select a button on the screen, and to draw a line, based on a coordinate value of a touch position and an ID (identification information) of an operation area received from thetouch table apparatus 1. -
FIG. 11 is a flowchart illustrating processing procedures in thetouch table apparatus 1 and thePC 2. First, thetouch table apparatus 1 is turned on, and then is initialized (ST 210). In the initialization, a level signal is obtained in an untouched state in which no touch operation is performed. This allows thetouch position detector 41 to detect a touch position based on a change amount of the level signal associated with a touch operation. - The
PC 2 starts an application for screen operation using thetouch table apparatus 1 and performs, in theoperation area setter 46, operation area setting processing that allows a user to designate an operation area. At this time, thetouch table apparatus 1 enters an area designation mode. The user performs a touch operation to designate an operation area (ST 110), and then thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 220) and transmits touch position information to thePC 2. ThePC 2 sets an operation area based on the touch position (ST 310). - After the operation area is set as above, the
touch table apparatus 1 enters a screen operation mode to allow a position input operation in the operation area. The user performs a touch operation for screen operation (ST 120). Then, thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 230) and transmits touch position information to thePC 2. ThePC 2 performs screen operation processing in thescreen operation processor 47 based on the touch position (ST 320). - Processing during operation area designation shown in a portion A of
FIG. 11 is described in detail below.FIG. 12 is a flowchart illustrating processing procedures for operation area designation shown in the portion A ofFIG. 11 .FIGS. 13A to 13D each illustrate a screen displayed on thedisplay 3 during operation area designation. Specifically,FIGS. 13A and 13B each illustrate a screen prompting a user to designate an operation area;FIG. 13C illustrates a screen prompting the user to select a coordinate mode;FIG. 13D illustrates a screen prompting the user to select whether or not to add an operation area. - With reference to
FIG. 12 , thePC 2 first performs in theoperation area setter 46 processing for displaying on thedisplay 3 an operation area designation screen (refer toFIG. 13A ) that prompts a user to designate one vertex (upper left herein) to define an operation area (ST 311). In response, the user touches a predetermined position on thetouch surface 6 or places thearea designation tool 31 in a predetermined position on the touch surface 6 (ST 111). Then, thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 221) and transmits a detected touch position to thePC 2. - The
PC 2 performs, in theoperation area setter 46, processing for detecting thearea designation tool 31 based on the touch position received from the touch table apparatus 1 (ST 312). When thePC 2 does not detect the area designation tool 31 (ST 312: No), thePC 2 performs, in theoperation area setter 46, processing for displaying on thedisplay 3 the operation area designation screen (refer toFIG. 13B ) that prompts the user to designate the other vertex (lower right herein) to define the operation area (ST 313). In response, the user touches a predetermined position on the touch surface 6 (ST 112). Then, thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 222) and transmits a detected touch position to thePC 2. ThePC 2 performs operation area setting in theoperation area setter 46 based on the two obtained vertexes (upper left and lower right) (ST 314). - When the
PC 2 detects the area designation tool 31 (ST 312: Yes), it is unnecessary to designate the other vertex to define the operation area. Thus, thePC 2 eliminates display of the operation area designation screen that prompts the user to designate the vertex (ST 313), and performs operation area setting in theoperation area setter 46 based on the placement position of the area designation tool 31 (ST 314). - Subsequently, the
PC 2 performs, in theoperation area setter 46, processing for displaying on the display 3 a coordinate mode selection screen (refer toFIG. 13C ) that prompts a user to select an absolute coordinate mode or a relative coordinate mode (ST 315). In response, the user touches thetouch surface 6 to select a predetermined coordinate mode (ST 113). When the user touches the right half area according to the indication on thedisplay 3, the “relative coordinate” is selected, whereas when the user touches the left half area, the “absolute coordinate” is selected. At this time, thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 223) and transmits a detected touch position to thePC 2. ThePC 2 determines the coordinate mode selected by the user based on the obtained touch position and performs coordinate mode setting processing in the operation area setter 46 (ST 316). - Subsequently, the
PC 2 performs in theoperation area setter 46 processing for displaying on thedisplay 3 an additional area selection screen (refer toFIG. 13D ) that prompts a user to select whether or not to add an operation area (ST 317). In response, the user touches thetouch surface 6 so as to select whether or not to add an operation area (ST 114). When the user touches the right half area according to the indication on thedisplay 3, “Yes” is selected, whereas when the user touches the left half area, “No” is selected. At this time, thetouch table apparatus 1 performs touch position detection processing in the touch position detector 41 (ST 224) and transmits a detected touch position to thePC 2. ThePC 2 determines whether or not to add an operation area based on the obtained touch position (ST 318). When there is an operation area to be added (ST 318: Yes), thePC 2 returns to the operation area designation screen (ST 311) to allow the user to designate a new operation area. - After setting the position and the coordinate mode of the operation area in the
operation area setter 46, thePC 2 transmits the information on the position and the coordinate mode of the operation area to thetouch table apparatus 1 to be stored in theoperation area memory 43. - Processing during screen operation shown in a portion B of
FIG. 11 is described in detail below.FIG. 14 is a flowchart illustrating processing procedures for screen operation shown in the portion B ofFIG. 11 .FIGS. 15A and 15B each illustrate a state of coordinate conversion during screen operation. Specifically,FIG. 15A illustrates a coordinate system of a touch table;FIG. 15B illustrates a coordinate system of an output screen. - With reference to
FIG. 14 , the user performs a touch operation for screen operation (ST 121), thetouch table apparatus 1 detects the touch operation (ST 231: Yes) and performs touch position detection processing in the touch position detector 41 (ST 232). In the touch position detection processing, a touch position is obtained in the coordinate system of the touch table. - Subsequently, operation area determination processing is performed in the operation area determinator 44 (ST 233). In the operation area determination processing, an operation area is determined in which the touch position obtained in the touch position detection processing (ST 232) is included, based on the operation area information in the
operation area memory 43. When the touch position is not included in any operation area (ST 233: No), the touch position is invalidated (ST 234). - When the touch position is included in any operation area (ST 233: Yes), coordinate conversion processing is performed in the coordinate converter 45 (ST 235). In the coordinate conversion processing, a coordinate value of the touch position obtained in the touch position detection processing (ST 232) is converted from the coordinate system of the touch table shown in
FIG. 15A into the coordinate system of the output screen shown inFIG. 15B . - In the example shown in
FIG. 15A , both operation areas A and B are set in the absolute coordinate mode. In the operation area A, coordinate values (Xa1, Ya1) to (Xa4, Ya4) in the coordinate system of the touch table are converted into coordinate values (0, 0) to (100, 50) in the coordinate system of the output screen. In the operation area B, coordinate values (Xb1, Yb1) to (Xb4, Yb4) in the coordinate system of the touch table are converted into coordinate values (0, 0) to (100, 50) in the coordinate system of the screen of thePC 2. - The operation areas A and B are provided for two users opposite to each other with the
touch table apparatus 1 therebetween. The operation area relative to the user has a positional relationship of 180 degrees. The operation area may have a positional relationship of 90 degrees depending on the position of the user, and the positional relationship is not constant. Thus, during operation area setting, the user is asked to enter the positional relationship of the operation area. Based on the entered information, coordinate conversion is performed so as to match the up, down, left, and right of the operation area as viewed from the user and the up, down, left, and right of the screen area. The coordinate conversion associated with the positional relationship of the operation area relative to the user is also required for the relative coordinate mode and the two-finger operation mode in addition to the absolute coordinate mode. - Then, as shown in
FIG. 14 , thetouch table apparatus 1 notifies thePC 2 of the touch position information (ST 236). Specifically, thetouch table apparatus 1 transmits to thePC 2, the ID (identification information) of the operation area obtained in the operation area determination process (ST 233) and the coordinate value in the coordinate system of the output screen obtained in the coordinate conversion processing (ST 235). Upon receiving the touch position information from the touch table apparatus 1 (ST 321: Yes), thePC 2 determines the content of the screen operation based on the touch position and performs predetermined processing associated with the content of the screen operation (ST 322). -
FIGS. 16 and 17 are each a perspective view illustrating an alternative example of use of the touch table system. - In the example shown in
FIG. 16 , a laptop PC (information processing apparatus) 61, instead of thedesktop PC 2 above, is placed on thetabletop 12 of thetouch table apparatus 1. For enlarged display of a screen of thelaptop PC 61, a projector (display apparatus) 62 is used to project the screen on a screen or a wall surface in a room as aprojection surface 63. - In this case, normally only the user D in front of the
laptop PC 61 can operate the screen. The remaining users A to C can also operate the screen by moving thelaptop PC 61. However, setting theoperation areas 22 a to 22 c for the users A to C, respectively, on thetouch table apparatus 1 allows the users A to C to each operate the screen of thelaptop PC 61 without moving thelaptop PC 61. - In the example shown in
FIG. 17 , a projector (display apparatus) 71 is used similar to the example above. Theprojector 71, which is of a short focus type, is placed on thetabletop 12 of thetouch table apparatus 1. Thetouch surface 6 of the upper surface of thetabletop 12 is used as a projection surface to project a screen of theprojector 71 so as to display the screen of thePC 2. - In this case, a
screen display area 72 is set as an operation area on thetouch surface 6 of thetouch table apparatus 1, allowing a user to operate the screen as if directly operating the screen displayed in thescreen display area 72. In particular, in this example, the screen is displayed proximate to the users A and B, who thus can operate the screen with a touch operation on thescreen display area 72. Theoperation areas screen display area 72, to allow them to operate the screen without moving from their seats. - In the present embodiment, a standalone display apparatus (
display 3 andprojectors 62 and 71) that displays a screen is used. Alternatively, the touch table apparatus may be integrally provided with a display apparatus. Specifically, a display apparatus may be disposed on the rear of the touch panel main body in the tabletop so as to display an image on the touch surface. In this case, the screen may be displayed in a portion of the touch detection area and the operation area may be set in the remaining space. - In the present embodiment, the
touch position converter 42 is provided in thetouch table apparatus 1, but may be provided in the information processing apparatus (PC 2). In the present embodiment, theoperation area setter 46 is provided in the information processing apparatus (PC 2), but may be provided in thetouch table apparatus 1. - In the present embodiment, the area designation tool having a frame shape is placed on the touch surface to allow touch operation on the touch surface inside the tool. An area designation tool is not limited to the configuration above in the present invention, and may be a chip-shaped member or an L-shaped member to designate two vertexes that define a rectangular operation area.
- In the present embodiment, a mutual capacitance system of an electrostatic capacitance system is employed as a method of detecting a touch position. Alternatively, a self-capacitance system may be employed. The self-capacitance system, however, does not support multi-touch which allows detection of a plurality of touch positions simultaneously, causing inconvenience in use. Thus, it is preferred to employ the mutual capacitance system.
- The touch system according to the present invention enhances convenience in use by a plurality of users. The touch system is useful as a touch system having a touch support member apparatus provided with a touch screen.
- It is noted that the foregoing examples have been provided merely for the purpose of explanation and are in no way to be construed as limiting of the present invention. While the present invention has been described with reference to exemplary embodiments, it is understood that the words which have been used herein are words of description and illustration, rather than words of limitation. Changes may be made, within the purview of the appended claims, as presently stated and as amended, without departing from the scope and spirit of the present invention in its aspects. Although the present invention has been described herein with reference to particular structures, materials and embodiments, the present invention is not intended to be limited to the particulars disclosed herein; rather, the present invention extends to all functionally equivalent structures, methods and uses, such as are within the scope of the appended claims.
- The present invention is not limited to the above described embodiments, and various variations and modifications may be possible without departing from the scope of the present invention.
Claims (20)
1. A touch system comprising:
a touch support member apparatus having a touch surface on which a touch operation is performed by a user and on which electrodes are arranged in a grid shape; and
an information processing apparatus connected to the touch support member apparatus,
the touch support member apparatus comprising:
a touch position detector configured to detect a touch position on an operation area of the touch surface based on a change of output signals from the electrodes associated with a change in capacitance in response to the touch operation; and
a touch position converter configured to convert a coordinate of the touch position, in the operation area, obtained by the touch position detector, into a coordinate of a screen area of the information processing apparatus.
2. The touch system according to claim 1 , wherein the information processing apparatus comprises an operation area setter setting the operation area.
3. The touch system according to claim 2 , wherein the operation area setter sets the operation area for the user based on touch positions obtained by the touch position detector.
4. The touch system according to claim 1 , wherein the information processing apparatus comprises a screen operation processor reflecting the touch operation performed in the operation area into the screen area based on the coordinate of the screen area converted by the touch position converter.
5. The touch system according to claim 3 , wherein the operation area is rectangular, and two diagonal vertexes of the operation area are designated by touch operations by the user.
6. The touch system according to claim 3 , further comprising an area designation tool at least partially comprising a conductive body, wherein
the operation area setter sets the operation area based on a placement position of the area designation tool upon detecting the area designation tool based on a detection result of the touch position detector.
7. The touch system according to claim 6 , wherein each side of the area designation tool is extendable and contractable.
8. The touch system according to claim 6 , wherein each side of the area designation tool has a telescopic mechanism.
9. The touch system according to claim 6 , wherein
the area designation tool is rectangular to define the operation area inside the area designation tool,
two diagonally positioned members of the area designation tool are formed of conductive bodies.
10. The touch system according to claim 1 , wherein
the operation area setter sets one of an absolute coordinate mode and a relative coordinate mode for the operation area according to a coordinate mode selection operation by a user, the absolute coordinate mode outputting a coordinate value of a touch position with an absolute coordinate, the relative coordinate mode outputting a coordinate value of a touch position with a relative coordinate, and
the touch position converter outputs a coordinate indicating a touch position relative to an immediately precedingly designated touch position, for the operation area set in the relative coordinate mode.
11. The touch system according to claim 1 , wherein the touch position converter comprises an operation area memory that stores information on the operation area set by the operation area setter, and an operation area determinator that determines whether or not the touch position detected by the touch position detector is in the operation area.
12. The touch system according to claim 11 , wherein the operation area determinator invalidates the touch position when the operation area determinator determines that the touch position is not in the operation area.
13. The touch system according to claim 1 , wherein the touch position converter switches to a two-finger operation mode to output a coordinate value of a touch position with a relative coordinate, based on a relative position of one finger to another finger, when the touch position converter detects that the two fingers touch the touch surface simultaneously.
14. A touch system comprising:
a touch support member apparatus having a touch surface on which touch operations are performed by a plurality of users and on which electrodes are arranged in a grid shape; and
an information processing apparatus connected to the touch support member apparatus,
the touch support member apparatus comprising:
a touch position detector configured to detect touch positions on a plurality of operation areas of the touch surface based on changes of output signals from the electrodes associated with changes in capacitance in response to the touch operations; and
a touch position converter configured to convert coordinates of the touch positions, in the operation areas, obtained by the touch position detector, into coordinates of a screen area of the information processing apparatus, wherein
each of the operation areas comprises a position input device assigned to one of the users.
15. The touch system according to claim 14 , each of the plurality of operation areas being configured to input a touch operation over the entire screen area of the information processing apparatus.
16. The touch system according to claim 14 , wherein
the operation area setter sets one of an absolute coordinate mode and a relative coordinate mode for each operation area according to a coordinate mode selection operation by a user, the absolute coordinate mode outputting a coordinate value of a touch position with an absolute coordinate, the relative coordinate mode outputting a coordinate value of a touch position with a relative coordinate, and
the touch position converter outputs a coordinate indicating a touch position relative to an immediately precedingly designated touch position, for an operation area set in the relative coordinate mode.
17. The touch system according to claim 16 , the touch position converter being configured to convert coordinates of a plurality of touch positions in a plurality of operation areas into coordinates of the screen area of the information processing apparatus, the operation area setter being configured to concurrently set at least one of the plurality of operation areas to the absolute coordinate mode and at least one of the plurality of operation areas to the relative coordinate mode.
18. The touch system according to claim 16 , wherein, in the relative coordinate mode, a position input operation comprises moving a second contact member with respect to a fixedly positioned contact member.
19. The touch system according to claim 14 , the information processing apparatus comprises a laptop with a display on the touch surface and a projector containing the screen area, each of the plurality of operation areas and the laptop being configured to control the display of the laptop.
20. The touch system according to claim 14 , the information processing apparatus comprising a projector, a projector area of the projector being projected onto the touch surface and comprising one of the plurality of operation areas.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011-176536 | 2011-08-12 | ||
JP2011176536A JP2013041350A (en) | 2011-08-12 | 2011-08-12 | Touch table system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130038548A1 true US20130038548A1 (en) | 2013-02-14 |
Family
ID=47677229
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/566,151 Abandoned US20130038548A1 (en) | 2011-08-12 | 2012-08-03 | Touch system |
Country Status (2)
Country | Link |
---|---|
US (1) | US20130038548A1 (en) |
JP (1) | JP2013041350A (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120192084A1 (en) * | 2010-10-25 | 2012-07-26 | Dedo Interactive, Inc. | Synchronized panel technology |
US20140075330A1 (en) * | 2012-09-12 | 2014-03-13 | Samsung Electronics Co., Ltd. | Display apparatus for multiuser and method thereof |
US20140104190A1 (en) * | 2012-10-17 | 2014-04-17 | Perceptive Pixel, Inc. | Selective Reporting of Touch Data |
US20140160054A1 (en) * | 2012-12-06 | 2014-06-12 | Qualcomm Incorporated | Anchor-drag touch symbol recognition |
WO2014158488A1 (en) * | 2013-03-14 | 2014-10-02 | Motorola Mobility Llc | Off-center sensor target region |
WO2015057496A1 (en) * | 2013-10-14 | 2015-04-23 | Microsoft Corporation | Shared digital workspace |
US9128548B2 (en) | 2012-10-17 | 2015-09-08 | Perceptive Pixel, Inc. | Selective reporting of touch data |
US20160216789A1 (en) * | 2015-01-28 | 2016-07-28 | Coretronic Corporation | Touch projection screen and manufacturing method thereof |
JP2016186734A (en) * | 2015-03-27 | 2016-10-27 | 富士通株式会社 | Window setting method, program and display controller |
US20170192511A1 (en) * | 2015-09-29 | 2017-07-06 | Telefonaktiebolaget Lm Ericsson (Publ) | Touchscreen Device and Method Thereof |
US10331277B2 (en) | 2015-03-13 | 2019-06-25 | Coretronic Corporation | Touch projection screen and touch projection system |
US10733959B2 (en) | 2017-11-21 | 2020-08-04 | Samsung Electronics Co., Ltd. | Method for configuring input interface and electronic device using same |
US11347367B2 (en) * | 2019-01-18 | 2022-05-31 | Dell Products L.P. | Information handling system see do user interface management |
US20220398008A1 (en) * | 2020-01-24 | 2022-12-15 | Ming Li | Volume Adjusting Gesture and Mistouch Prevention on Rolling Devices |
US11656654B2 (en) | 2019-01-18 | 2023-05-23 | Dell Products L.P. | Portable information handling system user interface selection based on keyboard configuration |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6327834B2 (en) * | 2013-11-01 | 2018-05-23 | シャープ株式会社 | Operation display device, operation display method and program |
WO2019003304A1 (en) * | 2017-06-27 | 2019-01-03 | マクセル株式会社 | Projection image display system |
JP6959529B2 (en) * | 2018-02-20 | 2021-11-02 | 富士通株式会社 | Input information management program, input information management method, and information processing device |
JP7129352B2 (en) * | 2019-01-30 | 2022-09-01 | シャープ株式会社 | Operation range setting device, game device, operation range setting method, and program |
JP6654722B2 (en) * | 2019-03-08 | 2020-02-26 | シャープ株式会社 | Image display device and image display method |
CN112984295A (en) * | 2021-02-03 | 2021-06-18 | 衡阳师范学院 | Data processing and receiving terminal for overall human resource planning |
Citations (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5561811A (en) * | 1992-11-10 | 1996-10-01 | Xerox Corporation | Method and apparatus for per-user customization of applications shared by a plurality of users on a single display |
US5825352A (en) * | 1996-01-04 | 1998-10-20 | Logitech, Inc. | Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad |
US6029214A (en) * | 1995-11-03 | 2000-02-22 | Apple Computer, Inc. | Input tablet system with user programmable absolute coordinate mode and relative coordinate mode segments |
US6392675B1 (en) * | 1999-02-24 | 2002-05-21 | International Business Machines Corporation | Variable speed cursor movement |
US20020189113A1 (en) * | 2001-06-19 | 2002-12-19 | Hsien-Chung Chou | Auxiliary drafting instrument combination applicable to digital board |
US20040046784A1 (en) * | 2000-08-29 | 2004-03-11 | Chia Shen | Multi-user collaborative graphical user interfaces |
US20040201628A1 (en) * | 2003-04-08 | 2004-10-14 | Johanson Bradley E. | Pointright: a system to redirect mouse and keyboard control among multiple machines |
US20050183035A1 (en) * | 2003-11-20 | 2005-08-18 | Ringel Meredith J. | Conflict resolution for graphic multi-user interface |
US20050183023A1 (en) * | 2004-02-12 | 2005-08-18 | Yukinobu Maruyama | Displaying and operating methods for a table-shaped information terminal |
US20050251800A1 (en) * | 2004-05-05 | 2005-11-10 | Microsoft Corporation | Invoking applications with virtual objects on an interactive display |
US20050259845A1 (en) * | 2004-05-24 | 2005-11-24 | Microsoft Corporation | Restricting the display of information with a physical object |
US20050285845A1 (en) * | 2004-06-28 | 2005-12-29 | Microsoft Corporation | Orienting information presented to users located at different sides of a display surface |
US20060001645A1 (en) * | 2004-06-30 | 2006-01-05 | Microsoft Corporation | Using a physical object to control an attribute of an interactive display application |
US20060092170A1 (en) * | 2004-10-19 | 2006-05-04 | Microsoft Corporation | Using clear-coded, see-through objects to manipulate virtual objects |
US20060214907A1 (en) * | 2005-03-23 | 2006-09-28 | Devos John A | Token configured to interact |
US20060230192A1 (en) * | 2005-03-29 | 2006-10-12 | Travis Parry | Display of a user interface |
US20070020604A1 (en) * | 2005-07-19 | 2007-01-25 | Pranaya Chulet | A Rich Media System and Method For Learning And Entertainment |
US20070124370A1 (en) * | 2005-11-29 | 2007-05-31 | Microsoft Corporation | Interactive table based platform to facilitate collaborative activities |
US20070171273A1 (en) * | 2006-01-26 | 2007-07-26 | Polycom, Inc. | System and Method for Controlling Videoconference with Touch Screen Interface |
US20070247435A1 (en) * | 2006-04-19 | 2007-10-25 | Microsoft Corporation | Precise selection techniques for multi-touch screens |
US20070262964A1 (en) * | 2006-05-12 | 2007-11-15 | Microsoft Corporation | Multi-touch uses, gestures, and implementation |
US20070294632A1 (en) * | 2006-06-20 | 2007-12-20 | Microsoft Corporation | Mutli-User Multi-Input Desktop Workspaces and Applications |
US20080022328A1 (en) * | 2006-06-30 | 2008-01-24 | Miller Robert R | Method and system for providing interactive virtual tablecloth |
US7397464B1 (en) * | 2004-04-30 | 2008-07-08 | Microsoft Corporation | Associating application states with a physical object |
US20080198138A1 (en) * | 2007-02-20 | 2008-08-21 | Microsoft Corporation | Identification of devices on touch-sensitive surface |
US20090106667A1 (en) * | 2007-10-19 | 2009-04-23 | International Business Machines Corporation | Dividing a surface of a surface-based computing device into private, user-specific areas |
WO2009070619A1 (en) * | 2007-11-26 | 2009-06-04 | Warren Daniel Child | Modular system and method for managing chinese, japanese, and korean linguistic data in electronic form |
US20090177862A1 (en) * | 2008-01-07 | 2009-07-09 | Kuo-Shu Cheng | Input device for executing an instruction code and method and interface for generating the instruction code |
US20090199338A1 (en) * | 2008-02-07 | 2009-08-13 | House Steven C | Furniture Attachment System and Methods of Use |
US20100083109A1 (en) * | 2008-09-29 | 2010-04-01 | Smart Technologies Ulc | Method for handling interactions with multiple users of an interactive input system, and interactive input system executing the method |
US20100079369A1 (en) * | 2008-09-30 | 2010-04-01 | Microsoft Corporation | Using Physical Objects in Conjunction with an Interactive Surface |
US20100095233A1 (en) * | 2006-10-13 | 2010-04-15 | Charlotte Skourup | Device, system and computer implemented method to display and process technical data for a device in an industrial control system |
US20100194703A1 (en) * | 2007-09-19 | 2010-08-05 | Adam Fedor | Multimedia, multiuser system and associated methods |
US20100205190A1 (en) * | 2009-02-09 | 2010-08-12 | Microsoft Corporation | Surface-based collaborative search |
US7849410B2 (en) * | 2007-02-27 | 2010-12-07 | Awind Inc. | Pointing-control system for multipoint conferences |
US20100328227A1 (en) * | 2009-06-29 | 2010-12-30 | Justin Frank Matejka | Multi-finger mouse emulation |
US20110019875A1 (en) * | 2008-08-11 | 2011-01-27 | Konica Minolta Holdings, Inc. | Image display device |
WO2011023225A1 (en) * | 2009-08-25 | 2011-03-03 | Promethean Ltd | Interactive surface with a plurality of input detection technologies |
US7970870B2 (en) * | 2005-06-24 | 2011-06-28 | Microsoft Corporation | Extending digital artifacts through an interactive surface |
US7980858B2 (en) * | 2007-06-29 | 2011-07-19 | Steelcase Development Corporation | Learning environment |
US20110187664A1 (en) * | 2010-02-02 | 2011-08-04 | Mark Rinehart | Table computer systems and methods |
US20110197263A1 (en) * | 2010-02-11 | 2011-08-11 | Verizon Patent And Licensing, Inc. | Systems and methods for providing a spatial-input-based multi-user shared display experience |
US20110304557A1 (en) * | 2010-06-09 | 2011-12-15 | Microsoft Corporation | Indirect User Interaction with Desktop using Touch-Sensitive Control Surface |
US20110304573A1 (en) * | 2010-06-14 | 2011-12-15 | Smith George C | Gesture recognition using neural networks |
US20110307578A1 (en) * | 2009-06-05 | 2011-12-15 | Samsung Electronics Co., Ltd. | Method for providing user interface for each user, method for performing service, and device applying the same |
US20120169623A1 (en) * | 2011-01-05 | 2012-07-05 | Tovi Grossman | Multi-Touch Integrated Desktop Environment |
US20120169618A1 (en) * | 2011-01-04 | 2012-07-05 | Lenovo (Singapore) Pte, Ltd. | Apparatus and method for gesture input in a dynamically zoned environment |
US20130016046A1 (en) * | 2011-07-13 | 2013-01-17 | Compal Electronics, Inc. | Control method and system of touch panel |
-
2011
- 2011-08-12 JP JP2011176536A patent/JP2013041350A/en not_active Withdrawn
-
2012
- 2012-08-03 US US13/566,151 patent/US20130038548A1/en not_active Abandoned
Patent Citations (49)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5561811A (en) * | 1992-11-10 | 1996-10-01 | Xerox Corporation | Method and apparatus for per-user customization of applications shared by a plurality of users on a single display |
US6029214A (en) * | 1995-11-03 | 2000-02-22 | Apple Computer, Inc. | Input tablet system with user programmable absolute coordinate mode and relative coordinate mode segments |
US5825352A (en) * | 1996-01-04 | 1998-10-20 | Logitech, Inc. | Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad |
US6392675B1 (en) * | 1999-02-24 | 2002-05-21 | International Business Machines Corporation | Variable speed cursor movement |
US20040046784A1 (en) * | 2000-08-29 | 2004-03-11 | Chia Shen | Multi-user collaborative graphical user interfaces |
US20020189113A1 (en) * | 2001-06-19 | 2002-12-19 | Hsien-Chung Chou | Auxiliary drafting instrument combination applicable to digital board |
US20040201628A1 (en) * | 2003-04-08 | 2004-10-14 | Johanson Bradley E. | Pointright: a system to redirect mouse and keyboard control among multiple machines |
US20050183035A1 (en) * | 2003-11-20 | 2005-08-18 | Ringel Meredith J. | Conflict resolution for graphic multi-user interface |
US20050183023A1 (en) * | 2004-02-12 | 2005-08-18 | Yukinobu Maruyama | Displaying and operating methods for a table-shaped information terminal |
US7397464B1 (en) * | 2004-04-30 | 2008-07-08 | Microsoft Corporation | Associating application states with a physical object |
US20050251800A1 (en) * | 2004-05-05 | 2005-11-10 | Microsoft Corporation | Invoking applications with virtual objects on an interactive display |
US20050259845A1 (en) * | 2004-05-24 | 2005-11-24 | Microsoft Corporation | Restricting the display of information with a physical object |
US20050285845A1 (en) * | 2004-06-28 | 2005-12-29 | Microsoft Corporation | Orienting information presented to users located at different sides of a display surface |
US20060001645A1 (en) * | 2004-06-30 | 2006-01-05 | Microsoft Corporation | Using a physical object to control an attribute of an interactive display application |
US20060092170A1 (en) * | 2004-10-19 | 2006-05-04 | Microsoft Corporation | Using clear-coded, see-through objects to manipulate virtual objects |
US20060214907A1 (en) * | 2005-03-23 | 2006-09-28 | Devos John A | Token configured to interact |
US20060230192A1 (en) * | 2005-03-29 | 2006-10-12 | Travis Parry | Display of a user interface |
US7970870B2 (en) * | 2005-06-24 | 2011-06-28 | Microsoft Corporation | Extending digital artifacts through an interactive surface |
US20070020604A1 (en) * | 2005-07-19 | 2007-01-25 | Pranaya Chulet | A Rich Media System and Method For Learning And Entertainment |
US20070124370A1 (en) * | 2005-11-29 | 2007-05-31 | Microsoft Corporation | Interactive table based platform to facilitate collaborative activities |
US20070171273A1 (en) * | 2006-01-26 | 2007-07-26 | Polycom, Inc. | System and Method for Controlling Videoconference with Touch Screen Interface |
US20070247435A1 (en) * | 2006-04-19 | 2007-10-25 | Microsoft Corporation | Precise selection techniques for multi-touch screens |
US20070262964A1 (en) * | 2006-05-12 | 2007-11-15 | Microsoft Corporation | Multi-touch uses, gestures, and implementation |
US20070294632A1 (en) * | 2006-06-20 | 2007-12-20 | Microsoft Corporation | Mutli-User Multi-Input Desktop Workspaces and Applications |
US20080022328A1 (en) * | 2006-06-30 | 2008-01-24 | Miller Robert R | Method and system for providing interactive virtual tablecloth |
US20100095233A1 (en) * | 2006-10-13 | 2010-04-15 | Charlotte Skourup | Device, system and computer implemented method to display and process technical data for a device in an industrial control system |
US20080198138A1 (en) * | 2007-02-20 | 2008-08-21 | Microsoft Corporation | Identification of devices on touch-sensitive surface |
US7849410B2 (en) * | 2007-02-27 | 2010-12-07 | Awind Inc. | Pointing-control system for multipoint conferences |
US7980858B2 (en) * | 2007-06-29 | 2011-07-19 | Steelcase Development Corporation | Learning environment |
US20100194703A1 (en) * | 2007-09-19 | 2010-08-05 | Adam Fedor | Multimedia, multiuser system and associated methods |
US20090106667A1 (en) * | 2007-10-19 | 2009-04-23 | International Business Machines Corporation | Dividing a surface of a surface-based computing device into private, user-specific areas |
US20110320468A1 (en) * | 2007-11-26 | 2011-12-29 | Warren Daniel Child | Modular system and method for managing chinese, japanese and korean linguistic data in electronic form |
WO2009070619A1 (en) * | 2007-11-26 | 2009-06-04 | Warren Daniel Child | Modular system and method for managing chinese, japanese, and korean linguistic data in electronic form |
US20090177862A1 (en) * | 2008-01-07 | 2009-07-09 | Kuo-Shu Cheng | Input device for executing an instruction code and method and interface for generating the instruction code |
US20090199338A1 (en) * | 2008-02-07 | 2009-08-13 | House Steven C | Furniture Attachment System and Methods of Use |
US20110019875A1 (en) * | 2008-08-11 | 2011-01-27 | Konica Minolta Holdings, Inc. | Image display device |
US20100083109A1 (en) * | 2008-09-29 | 2010-04-01 | Smart Technologies Ulc | Method for handling interactions with multiple users of an interactive input system, and interactive input system executing the method |
US20100079369A1 (en) * | 2008-09-30 | 2010-04-01 | Microsoft Corporation | Using Physical Objects in Conjunction with an Interactive Surface |
US20100205190A1 (en) * | 2009-02-09 | 2010-08-12 | Microsoft Corporation | Surface-based collaborative search |
US20110307578A1 (en) * | 2009-06-05 | 2011-12-15 | Samsung Electronics Co., Ltd. | Method for providing user interface for each user, method for performing service, and device applying the same |
US20100328227A1 (en) * | 2009-06-29 | 2010-12-30 | Justin Frank Matejka | Multi-finger mouse emulation |
WO2011023225A1 (en) * | 2009-08-25 | 2011-03-03 | Promethean Ltd | Interactive surface with a plurality of input detection technologies |
US20110187664A1 (en) * | 2010-02-02 | 2011-08-04 | Mark Rinehart | Table computer systems and methods |
US20110197263A1 (en) * | 2010-02-11 | 2011-08-11 | Verizon Patent And Licensing, Inc. | Systems and methods for providing a spatial-input-based multi-user shared display experience |
US20110304557A1 (en) * | 2010-06-09 | 2011-12-15 | Microsoft Corporation | Indirect User Interaction with Desktop using Touch-Sensitive Control Surface |
US20110304573A1 (en) * | 2010-06-14 | 2011-12-15 | Smith George C | Gesture recognition using neural networks |
US20120169618A1 (en) * | 2011-01-04 | 2012-07-05 | Lenovo (Singapore) Pte, Ltd. | Apparatus and method for gesture input in a dynamically zoned environment |
US20120169623A1 (en) * | 2011-01-05 | 2012-07-05 | Tovi Grossman | Multi-Touch Integrated Desktop Environment |
US20130016046A1 (en) * | 2011-07-13 | 2013-01-17 | Compal Electronics, Inc. | Control method and system of touch panel |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120192084A1 (en) * | 2010-10-25 | 2012-07-26 | Dedo Interactive, Inc. | Synchronized panel technology |
US9235312B2 (en) * | 2010-10-25 | 2016-01-12 | Dedo Interactive, Inc. | Synchronized panel technology |
US20140075330A1 (en) * | 2012-09-12 | 2014-03-13 | Samsung Electronics Co., Ltd. | Display apparatus for multiuser and method thereof |
US20140104190A1 (en) * | 2012-10-17 | 2014-04-17 | Perceptive Pixel, Inc. | Selective Reporting of Touch Data |
US8954638B2 (en) * | 2012-10-17 | 2015-02-10 | Perceptive Pixel, Inc. | Selective reporting of touch data |
US9128548B2 (en) | 2012-10-17 | 2015-09-08 | Perceptive Pixel, Inc. | Selective reporting of touch data |
US20140160054A1 (en) * | 2012-12-06 | 2014-06-12 | Qualcomm Incorporated | Anchor-drag touch symbol recognition |
WO2014158488A1 (en) * | 2013-03-14 | 2014-10-02 | Motorola Mobility Llc | Off-center sensor target region |
US9506966B2 (en) | 2013-03-14 | 2016-11-29 | Google Technology Holdings LLC | Off-center sensor target region |
WO2015057496A1 (en) * | 2013-10-14 | 2015-04-23 | Microsoft Corporation | Shared digital workspace |
US10754490B2 (en) | 2013-10-14 | 2020-08-25 | Microsoft Technology Licensing, Llc | User interface for collaborative efforts |
US9720559B2 (en) | 2013-10-14 | 2017-08-01 | Microsoft Technology Licensing, Llc | Command authentication |
US9740361B2 (en) | 2013-10-14 | 2017-08-22 | Microsoft Technology Licensing, Llc | Group experience user interface |
US9921671B2 (en) * | 2015-01-28 | 2018-03-20 | Coretronic Corporation | Touch projection screen and manufacturing method thereof |
US20160216789A1 (en) * | 2015-01-28 | 2016-07-28 | Coretronic Corporation | Touch projection screen and manufacturing method thereof |
US10331277B2 (en) | 2015-03-13 | 2019-06-25 | Coretronic Corporation | Touch projection screen and touch projection system |
JP2016186734A (en) * | 2015-03-27 | 2016-10-27 | 富士通株式会社 | Window setting method, program and display controller |
US20170192511A1 (en) * | 2015-09-29 | 2017-07-06 | Telefonaktiebolaget Lm Ericsson (Publ) | Touchscreen Device and Method Thereof |
US10733959B2 (en) | 2017-11-21 | 2020-08-04 | Samsung Electronics Co., Ltd. | Method for configuring input interface and electronic device using same |
US11347367B2 (en) * | 2019-01-18 | 2022-05-31 | Dell Products L.P. | Information handling system see do user interface management |
US11656654B2 (en) | 2019-01-18 | 2023-05-23 | Dell Products L.P. | Portable information handling system user interface selection based on keyboard configuration |
US20220398008A1 (en) * | 2020-01-24 | 2022-12-15 | Ming Li | Volume Adjusting Gesture and Mistouch Prevention on Rolling Devices |
Also Published As
Publication number | Publication date |
---|---|
JP2013041350A (en) | 2013-02-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20130038548A1 (en) | Touch system | |
US5365461A (en) | Position sensing computer input device | |
US9448645B2 (en) | Digitizer using multiple stylus sensing techniques | |
US8115744B2 (en) | Multi-point touch-sensitive system | |
US8466934B2 (en) | Touchscreen interface | |
CN102541365B (en) | Multi-touch command generating apparatus and method | |
US20130038549A1 (en) | Input device for touch screen and touch screen system having the same | |
US8106891B2 (en) | Multi-point touch-sensitive device | |
US20140210748A1 (en) | Information processing apparatus, system and method | |
KR101084438B1 (en) | Method of operating a multi-point touch-sensitive system | |
US8743089B2 (en) | Information processing apparatus and control method thereof | |
US20110298722A1 (en) | Interactive input system and method | |
EP1769328A2 (en) | Zooming in 3-d touch interaction | |
US20120249463A1 (en) | Interactive input system and method | |
KR20080095085A (en) | Apparatus and method for user interface through revolution input device | |
US20080284730A1 (en) | Device, method, and computer readable medium for mapping a graphics tablet to an associated display | |
US20130106745A1 (en) | Touch pad operable with multi-objects and method of operating same | |
US11360642B2 (en) | Method and apparatus for setting parameter | |
US9367228B2 (en) | Fine object positioning | |
JP2016122345A (en) | Image projection device and interactive input/output system | |
US9395828B2 (en) | Electronic information board apparatus, that displays image input from external apparatus | |
US10803836B2 (en) | Switch device and switch system and the methods thereof | |
US20210286460A1 (en) | Pressure activated accurate pointing | |
KR102169236B1 (en) | Touchscreen device and method for controlling the same and display apparatus | |
KR102254091B1 (en) | Touchscreen device and method for controlling the same and display apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: PANASONIC CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KITADA, TAKASHI;MAKI, TADASHI;REEL/FRAME:029116/0432 Effective date: 20120727 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |