US20100265209A1 - Power reduction for touch screens - Google Patents
Power reduction for touch screens Download PDFInfo
- Publication number
- US20100265209A1 US20100265209A1 US12/746,732 US74673210A US2010265209A1 US 20100265209 A1 US20100265209 A1 US 20100265209A1 US 74673210 A US74673210 A US 74673210A US 2010265209 A1 US2010265209 A1 US 2010265209A1
- Authority
- US
- United States
- Prior art keywords
- touch screen
- user input
- engine
- area
- touch
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/325—Power saving in peripheral device
- G06F1/3265—Power saving in display device
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3215—Monitoring of peripheral devices
- G06F1/3218—Monitoring of peripheral devices of display devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/325—Power saving in peripheral device
- G06F1/3262—Power saving in digitizer or tablet
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- This specification relates to activation of touch detection within a touch screen.
- portable information terminals for example PDAs, laptops, tablet PCs, video players, music players, multimedia players, cameras, mobile phone, and the like
- touch screens for receiving user input are emerging into the market.
- Touch screens for receiving user input may be understood as touch sensible input screens, which are arranged to detect a user input from depressing a screen which displays user information.
- Touch screens may be a combination of a display arranged below a touch sensitive sheet, which is capable of sensing the location of contact with a finger or a pen.
- touch screens may be a combination of a display arranged with a touch sensitive switch matrix, e.g. a display integrated touch screen, which is capable of sensing the location of contact with a finger, a pen or any other object.
- a touch screen may receive user inputs, for example pressing a button or an icon, or selecting certain areas, writing memos, selecting programs, and the like within a user interface of a computer program.
- a microprocessor e.g. a drive engine, which is responsible for the operation of the portable information terminal, i.e. the computer program running on the terminal, needs to receive the detected user inputs and convert them into the appropriate program instructions.
- the drive engine responsible for the operation of the information terminal needs to receive the signals from the touch screen, and to convert these signals into the appropriate program logic.
- the drive engine For being able to receive and process the signals received from the touch screen, respectively a microprocessor operating the touch screen, e.g. a touch screen engine, the drive engine needs to dedicate at least parts of its processing power to the touch screen engine.
- the drive engine may operate, besides the touch screen, also loudspeakers, transmission and reception antennas and modules, for example for wireless communication, e.g. GSM, UMTS, WiFi, Near Field Communication (NFC), Bluetooth and the like, keyboards, global positioning devices (GPS), microphones, camera devices, display devices, multimedia processors, and the like. All of these devices may be operated by the drive engine and the interoperation between the devices is controlled by the drive engine.
- the touch screen engine may issue an interrupt for the drive engine.
- the drive engine may dedicate at least a part of its processing power to the touch screen and/or the touch screen engine in order to receive and process the signals received from the touch screen and/or the touch screen engine.
- the drive engine may be activated, and power consumption of the drive engine may thus increase.
- a portable information terminal For example, from U.S. Pat. No. 6,091,031, there is known a portable information terminal.
- the information terminal has a predetermined area of a touch screen panel, which covers a liquid crystal screen. Further, there is provided a program selection screen.
- the terminal may be activated by touching the program selection screen, which acts as a system activation area. The system is only activated after depressing the activation area for a predetermined time.
- the program selection area is a predetermined area, which covers at least 10-20% of the whole area of the touch screen. Depressing the program selection area for a predetermined time inadvertently may cause the panel to be activated. Activating the system upon depressing the program selection area for a predetermined time thus may not necessarily be user initiated. The power consumption of the system thus may increases due to faulty activation.
- a method comprising determining areas within a touch screen, where user input is possible. Activating touch detection of the touch screen for sensing a user input only within the determined areas, where user input is possible is provided.
- Areas within the touch screen may change during changing operational states of the terminal. For example, when the terminal is in a sleep mode, a.k.a. rest state, sleep state, power save mode, or the like, there is only a small area within the whole touch screen, which may be used as activation area. This activation area may change dynamically and may cover only less then 10% of the touch screen. The activation area might be indicated by an appropriate button or icon or any other means within the display.
- the software running on the device with the touch screen may control the display to show the activation icon at a certain position within the touch screen. The position of the activation area may change dynamically.
- Determination of the area, where user input is possible may allow for detecting automatically, and dynamically, where user input is possible at all. In other areas than the activation areas, no user input may be required and possible. User input within other areas may not be detected and may not cause issuing an interrupt. Only those areas, where user input is possible, may be determined as activation areas.
- the determination of the area(s), where user input is possible, may depend on the content actually shown on the display. There may be means which analyse the content shown and which determine, which areas are used as activation areas. Touch detection of the touch screen may only be activated for areas, which are determined to be possible input areas. The touch detection itself may already consume energy, as, for example, the touch state of the display needs to be analysed almost in real time. The display area needs to be activated for touch detection. By determining the areas, where user input is possible, only those areas need to be analysed, providing reduced power consumption.
- the display may show three different selection icons, for example “yes”, “no”, “cancel”. Only within these three icons, user input is possible. Determining the areas, where user input is possible, allows for detecting the three icons within the display. Only the areas, which are overlap with these icons may be activated for touch detection. Other areas may be temporally deactivated for touch detection.
- input is only accepted within the detected areas. For example, when the user presses the touch screen outside the detected areas, no action is triggered by the drive engine. Only when the user presses the touch screen at the determined areas, the appropriate program action is triggered by the drive engine.
- the touch screen or the touch screen engine may issue an interrupt for the drive engine, when sensing a user input.
- the interrupt provides for information for the drive engine to provide for processing power for processing user input from the touch screen.
- the drive engine may allocate processing power to the processing of touch screen information.
- the activation of the drive engine for processing touch screen engine signals may depend upon reception of the interrupt from the touch screen engine.
- Embodiments provide activating the drive engine for processing signals from the touch screen engine indicative of a user input upon reception of the interrupt.
- the drive engine may be activated for processing the signals from the touch screen engine only when the interrupt from the touch screen engine is received.
- the interrupt may be issued and received, only when a user touches the touch screen at the determined areas. Touching the touch screen at areas outside the determined areas may not cause triggering of the interrupt and the drive engine may not receive the interrupt and may not provide for the necessary processing power for processing touch detection.
- embodiments provide analysing display information within the touch screen and selecting areas from the display information within which a user input is possible.
- the user input may be possible within a user selection button, an icon, a character input field, a QWERTY input field, or any other field, which is capable of receiving user input. Determining these fields may, on the one hand be done by analysing the display content. On the other hand it may also be possible to receive from the respective program, which provides for the display content, information about the areas, within which user input is possible.
- a user interface API may provide the information, within which areas user input is possible.
- a user input may comprise obtaining press position information.
- the touch screen may allow for detecting the coordinates of a press position. Detecting the coordinates of the press position allows for detecting, whether the press position is within a determined area or not and initiating the respective operation.
- Touch detection may comprise, according to embodiments, making the touch screen sensible for haptic user input. For example, users may use their fingers to input information. Also, input pens may be used.
- the coordinates of the press position is detected.
- the detected coordinates may, according to embodiments, be converted within the touch screen engine into corresponding signals provided to the drive engine.
- the drive engine may thus control the software to operate in accordance with the user input.
- a pen may be used, when a resistive touch screen is used.
- the resistive touch screen may utilise a change in impedance of the touch screen, when the pen is pressed onto the touch screen.
- capacitive touch screens input may be possible using a finger.
- a capacitive touch screen utilises the change in capacitance of the touch screen. For example, when a finger approaches the touch screen, the capacitance of the touch screen changes, which may be evaluated and the press position may be detected.
- An optical touch detection may be operated using a finger or any other means touching the screen.
- the determination of areas, where a user input is possible may be provided in a normal state and/or in a power safe state of at least the drive engine.
- the activation is only possible within a small icon displayed on the touch screen. Only pressing this icon may allow for activating the terminal.
- activation of the drive engine i.e. by issuing the interrupt, shall only be possible, when the activation icon is pressed.
- the activation icon is determined, and user input is only possible within the activation icon. Interrupts are issued only when this icon is pressed.
- the drive engine thus consumes less energy in the power safe state, because interrupts are only issued when the activation area is touched.
- the drive engine is not activated, for example, by inadvertently pressing any other area of the touch screen.
- Another aspect of the specification is an apparatus comprising a touch screen, a touch screen engine, and a drive engine.
- the touch screen engine is arranged for determining areas, where a user input is possible.
- a touch screen engine is activated for sensing a user input only within the determined areas, where a user input is possible.
- a further aspect of the specification is a device comprising a means for driving a touch screen, and a means for driving the device, wherein the means for driving the touch screen are arranged for determining areas within a touch screen, where the user input is possible, and wherein the means for driving the touch screen are activated for sensing a user input only within the determined areas, where a user input is possible.
- the device may, according to embodiments, be, for example, a PDA, laptop, tablet PC, video player, music player, multimedia player, camera, mobile phone, or any other user device requiring user inputs.
- Another aspect of the specification is a computer-readable medium having a computer program stored thereon, the computer program comprising instructions operable to cause a processor to determine areas within a touch screen, where a user input is possible, and to activate touch detection of the touch screen for sensing a user input only within the determined areas, where user input is possible.
- a further aspect of the specification is a computer program comprising instructions operable to cause a processor to determine areas within a touch screen, where a user input is possible, and activate touch detection of the touch screen for sensing a user input only within the determined areas, where a user input is possible.
- FIG. 1 a block diagram of a mobile phone with its components
- FIG. 2 a side view of a touch screen
- FIG. 3 schematically a block diagram of a touch screen system
- FIG. 4 schematically a diagram of a touch screen system
- FIG. 5 schematically a display panel with pixel cells
- FIG. 6 schematically pixel cells with touch detection
- FIG. 7 a a screenshot of a display within a touch screen
- FIG. 7 b areas within which user input is possible of a screenshot as illustrated in FIG. 4 a;
- FIG. 7 c a combination of screenshot of FIG. 4 a and display of areas according to FIG. 4 b;
- FIG. 8 a flowchart of a method according to embodiments.
- FIG. 1 illustrates schematically a block diagram of a mobile device 2 .
- the mobile device 2 may be a terminal as previously described.
- different appliances and peripherals can be included within a mobile device 2 .
- a selection of possible appliances and peripherals are shown in FIG. 1 . It should be noted, that the selection of shown appliances and peripherals is illustrative only and shall not be understood as limiting.
- mobile device 2 is a mobile phone having a drive engine 4 .
- Drive engine 4 may be comprised of hardware and software.
- Drive engine 4 may be capable of operating all peripherals and any kind of software necessary for operating the peripherals.
- Drive engine 4 may be a microprocessor, which processes the mobile device 2 according to different standards, applications, and the like.
- Drive engine 4 may be understood as the core engine of the mobile device 2 , which is responsible for the operation and interoperation of programs and appliances, which are hereinafter explained.
- a touch screen 6 may comprise a touch screen panel 7 .
- Touch screen panel 7 may be placed in front of a display 8 .
- the touch screen panel may also be incorporated within display 8 .
- Touch screen panel 7 may be operated by a touch screen engine, i.e. a touch screen controller (not depicted).
- Touch screen panel 7 and display 8 may be connected to drive engine 4 .
- Touch screen panel may comprise a touch screen controller and may be a component, which is converting physical touches onto its surface or the surface of the display 8 into an electrical format, i.e. signals for drive engine 4 for operating programs and other appliances.
- Touch screen 6 will be further illustrated with reference to FIG. 2 .
- a display 8 may be arranged.
- Display 8 may be a component, which is converting electrical information received from the drive engine 4 into a readable format. This information may be any information generated from a software for controlling a user interface.
- Display 8 may be an LED display, OLED display, TFT display, CRT display, plasma display, or any other kind of display capable of converting information into a user readable format.
- the display 8 receives display information from drive engine 4 and puts out this information as optical information.
- the camera 10 may be a component, which is converting image information into a suitable format for further processing by drive engine 4 .
- Microphone 12 may be a component, which is converting audio information from acoustic waves into electrical information. Microphone 12 may receive user input via acoustic waves and may input these to drive engine 4 .
- GPS receiver 14 is a component for converting position information, i.e. from satellites into respective position information for drive engine 4 .
- keyboard 16 may be connected to drive engine 4 .
- Keyboard 16 may be a component, which is converting information from depressed keys into signals for drive engine 4 for receiving user input.
- a transmission and reception component 18 may allow for wired and wireless communication with any kind of other communication hardware.
- GSM and UMTS communication may be possible via component 18 .
- NFC, WiFi, or any other wireless communication may be possible.
- Component 18 may allow communicating via LAN, WAN or any other wired communication line.
- Loudspeaker 20 may be a component for converting electric information into acoustic waves.
- touch screen 6 i.e. touch screen panel 7 , display 8 and drive engine 4 .
- Power consumption of drive engine 4 shall be reduced by controlling touch screen panel 7 appropriately.
- Touch screen 6 is further illustrated in FIG. 2 .
- FIG. 2 illustrates a side view onto a touch screen 6 with a display 8 .
- Display 8 is arranged above a light guide 22 and covered by protection sheets 24 . Between protection sheets 24 and display 8 , there is arranged a touch detection sheet 26 , which enables the touch screen 6 , i.e. the touch screen controller, to detect a touch position of, for example, a touch pen 28 .
- the display 8 may driven by a display driver 30 .
- Display driver 30 may provide display 8 with display information, which is being displayed on display 8 and can be seen from a user's viewing direction 32 .
- the display information may be received from the drive engine 4 via a flex-foil connection (not depicted), or any other kind of electrical connection.
- Display 8 , light guide 22 , protection sheets 24 , and touch detection sheets 26 may in common or in any combination thereof be understood as touch screen 6 .
- Touch screen 6 may be connected to the drive engine 4 via an electrical connection, as will be shown in FIG. 3 .
- Touch screen panel 7 may be comprised of a touch screen engine and touch detection sheets 26 .
- Light guide 22 may be connected with a back lighting controller (not depicted) and provides the display 8 with back light, so that the content being displayed on display 8 and provided through display driver 30 can be seen even in dark viewing conditions.
- a back lighting controller not depicted
- a user may select a certain icon or item being displayed on display 8 . This may be done by detecting the press position of pen 28 on touch screen 6 using the touch detection sheet 26 .
- the touch detection and position detection is provided by a touch screen controller (not depicted), a.k.a. touch screen driver, being further illustrated in FIG. 3 .
- the touch screen driver may be a microprocessor running a program suitable for controlling the touch screen 6 , and/or the touch detection sheet 26 and for obtaining touch information from touch screen 6 and/or the touch detection sheet 26 .
- FIG. 3 illustrates a touch screen 6 being connected with touch screen controller 34 .
- Touch screen controller 34 is connected with drive engine 4 via interrupt line 36 .
- touch screen 6 provides for touch detection information to touch screen controller 34 .
- touch screen controller 34 provides for an interrupt via interrupt line 36 to drive engine 4 in order to activate drive engine 4 for processing user input through touch screen 6 .
- touch screen controller 34 issues an interrupt to drive engine 4 every time touch screen 6 is touched, no matter where the touch detection locates the area, where touch screen 6 is touched. This leads to issuing a plurality of interrupts on interrupt line 36 .
- Engine 4 is activated every time touch screen 6 is touched, even if the touch detection is not within areas, which allow or require user input. This leads to increased power consumption, as drive engine 4 needs to allocate processing power for detecting user input through touch screen 6 every time it receives an interrupt.
- FIG. 4 illustrates in more detail a touch screen controller 34 .
- touch screen controller 34 in connected to a drive engine 4 via an interface 36 , which may be a flex foil interface 36 .
- interface 36 may be a flex foil interface 36 .
- touch screen controller 36 may receive display information and may send touch detection signals.
- Touch screen controller 34 may be comprised of a frame memory 34 a .
- the image information is provided column by column through D/A converter 34 b to display panel 6 .
- a timing controller 34 c may provide clocking signals for selecting line addresses.
- the line addresses are provided to display 6 and also to frame memory 34 a by address coder 34 d . Through the line addresses, the display 6 is activated line by line and the respective pixel information for the respective lines is provided through frame memory 34 a.
- FIG. 5 illustrates several pixel cells 100 within a touch screen 6 .
- Each pixel cell 100 may represent one pixel.
- the pixel cell 100 may be comprised of transistor 100 a , capacitor 100 b , and liquid crystal cell 100 c.
- the column selection for a pixel cell 100 c may be done by activating the respective source line 102 (Source: Sn, Sn+1, Sn+2).
- the source lines 102 may be connected to DAC 34 b for receiving pixel data.
- the row selection may be done through gate line 104 (Gate: Gn, Gn+1, Gn+2, etc.) signals.
- Gate lines 104 may be connected to address coding 34 d.
- source line 102 and gate line 104 for a particular transistor 100 a are activated, the respective liquid crystal cell 100 c at pixel cell 100 is activated, and the pixel cell 100 c shows the image data, i.e. light intensity and color, for this respective pixel in the image.
- the block diagram of the pixel cells 100 as illustrated in FIG. 5 is working as follows
- Image data is input from interface 36 , which source is drive engine 4 , to the frame memory 34 a on the touch screen controller 34 .
- Timing controller 34 e sends timing information to address coding 34 d which generates control signals for controlling the line selection.
- the line selection within address coding 34 d may read location information from the frame memory 34 a by using a latch pointer and a line pointer.
- the digital image data is input to Digital-Analog-Converter (DAC) 34 b .
- DAC Digital-Analog-Converter
- the data is converted to an analogue image data for a certain column 102 , being represented by the source line 102 .
- the analogue image data is also inputted to the display panel for line selection.
- the location of each displayed pixel is controlled by address coding block 34 d via source lines 102 and gate line 204 control signals.
- the gate line control signal may have digital values (‘0’ or ‘1’), which may be used for selected a line of the pixel on the display panel.
- the pixel value of a certain column, being stored as digital information of the image data, may then be provided through source lines 102 , respectively.
- the source line 102 and gate line 104 are activated and the displayed pixel value represents the analogue value of the respective source line 102 .
- the analogue image data i.e. the current at source line 102
- the analogue image data can flow through transistor 100 a and load charging capacitor 100 b . This loading is continued until there is selected another gate line by setting another gate line 104 HIGH.
- the loading of capacitor 100 b controls the brightness of liquid crystal cell 100 c of the pixel cell 100 .
- the loaded capacitor 100 b keeps the analogue value, i.e. the visible grey level of the pixel cell 100 , until the same gate line 104 is selected again and a new loading is carried out.
- the pixel cell 100 which is visible, is working as follows
- Analogue image data is output on the source lines 102 (Sn, Sn+1, Sn+2, etc.).
- a selection which is the used gate line 104 , where all pixel cells 100 are updated, is further output by setting the respective gate line HIGH.
- the HIGH gate line 104 represents the line of the pixel cell 100 , which are updated at the same time.
- embodiments provide for determining areas within a touch screen, where a user input is possible and activating touch detection of the touch screen for sensing a user input only within the determined areas, where a user input is possible. This detection of areas, where user input is possible, is further illustrated in FIGS. 6-9 .
- FIG. 6 illustrates pixel cell 100 as illustrated in FIG. 5 , further comprising transistors for touch detection 106 .
- Pixel cell 100 further comprises touch detection sensors 108 .
- the selective touch detection works as follows:
- the gate driver includes same amount of the lines what are used for display panel 6 as illustrated in FIG. 5 . These lines are indicated as common gate lines 104 CGn, CGn+1, GCn+1, etc.
- the transistors for touch detection 106 are presumably activated in the same way and time when the transistors 100 a of pixel cell 100 on the display panel 6 are activated.
- a touch screen sensor 108 of pixel cell 100 For detecting touches on the display panel 6 , it is checked, whether a touch screen sensor 108 of pixel cell 100 is depressed. That means that only for those lines, where the CGn line 104 is active, touch sensors 108 are read out.
- touch screen gate lines 110 can be selected such that only those touch screen gate lines are HIGH, where user input is possible. This may be detected through analysing the content of the image.
- read out lines 112 may be used. Only those read out lines 112 may be read out, where touch detection is possible, or desired. This results in the possibility to selectively choose the pixel cells 100 , where touch detection is possible.
- FIG. 7 a illustrates a screenshot of a user interface 40 .
- the screenshot is a program window. Within this program window, it is possible to input user information only at certain areas.
- the program requires the user to input a selection of “yes”, “no”, or “cancel”.
- user interface 40 being displayed on display 8 , allows input only within the areas 42 , 44 , 46 , being input buttons.
- buttons 42 , 44 , 46 Touching the touch screen on any other position than the buttons 42 , 44 , 46 , would not result in a reaction of the program. Only touching of any of the buttons 42 , 44 , 46 , allows the program to move to its next state. In order to suppress interrupts being sent from touch screen controller 34 to engine 4 , when the display is touched at positions outside buttons 42 , 44 , 46 , it is necessary to determine these areas.
- FIG. 7 b is a representation of user interface 40 , where the locations of buttons 42 , 44 , 46 are highlighted.
- the highlighted areas of buttons 42 , 44 , 46 represent areas, within which touch screen 6 is activated, i.e. reacts on user input. In other areas the touch screen 6 is not sensible for touch detection, i.e. when areas other than the buttons 42 , 44 , 46 are touched there is no reaction of the touch screen 6 .
- the respective touch screen gate lines 110 where the buttons 42 , 44 , 46 are located, are set to HIGH. Further, the horizontal position of the buttons 42 , 44 , 46 determines, which read out lines 112 are actually read out. This results in only detecting touches on the display 6 in the areas of the buttons 42 , 44 , 46 .
- FIG. 7 c illustrated an overlay of the activated areas and buttons 42 , 44 , 46 in user interface 40 .
- User input is only possible at buttons 42 , 44 , 46 .
- the user can select one of buttons 42 , 44 , 46 and an interrupt is issued to engine 4 via touch screen controller 34 . Touching the touch screen 6 on any other position does not result in issuing such an interrupt.
- FIG. 8 illustrates a flowchart of a method according to embodiments.
- Display controller receives ( 52 ) display information for being displayed on display 8 .
- the display information is forwarded ( 54 ) to touch screen controller 34 .
- touch screen controller 34 the display information is analysed, and it is determined ( 56 ), where areas are located, where user input is possible. It is also possible, that a user interface API is requested by touch screen controller 34 to give information about where areas are located where a user input is possible.
- the information for being displayed is displayed ( 58 ) on display 8 .
- the touch screen 6 and touch screen controller 34 are arranged ( 60 ) such that they do only react on user input at the determined areas. If no user input at the determined areas is detected, the next image is being evaluated and displayed ( 52 - 58 ).
- touch screen controller 34 issues ( 62 ) an interrupt for drive engine 4 .
- the issuance of the interrupt initiates within drive engine 4 the appropriate program logic, and the program is further processed ( 64 ) according to the user input. This may be done by further detecting user inputs or by proceeding with the program logic. For example, proceeding program logic may result in storing certain results.
- issuing the interrupt and carrying out program logic consumes energy.
- issuing the interrupt should only occur, when the touch screen 6 is touched at areas, where user input is possible.
- touch detection is only carried out within the areas where user input is possible. Only touching the touch screen at these positions results in an issuance of an interrupt for engine 4 and further processing of a program logic. Power consumption is reduced only to cases where the touch screen is touched at areas, where user input is possible and expected. This results in a reduction of power consumption of device 2 .
- the logical blocks in the schematic block diagrams as well as the flowchart and algorithm steps presented in the above description may at least partially be implemented in electronic hardware and/or computer software, wherein it depends on the functionality of the logical block, flowchart step and algorithm step and on design constraints imposed on the respective devices to which degree a logical block, a flowchart step or algorithm step is implemented in hardware or software.
- the presented logical blocks, flowchart steps and algorithm steps may for instance be implemented in one or more digital signal processors, application specific integrated circuits, field programmable gate arrays or other programmable devices.
- the computer software may be stored in a variety of storage media of electric, magnetic, electro-magnetic or optic type and may be read and executed by a processor, such as for instance a microprocessor.
- a processor such as for instance a microprocessor.
- the processor and the storage medium may be coupled to interchange information, or the storage medium may be included in the processor.
Abstract
This specification relates to determining areas within touch screen, where a user input is possible. To reduce power consumption of a device with a touch screen, touch detection for sensing a user input is only activated within the determined areas, where a user input is possible.
Description
- This specification relates to activation of touch detection within a touch screen.
- Within portable information terminals, for example PDAs, laptops, tablet PCs, video players, music players, multimedia players, cameras, mobile phone, and the like, touch screens for receiving user input are emerging into the market.
- Touch screens for receiving user input may be understood as touch sensible input screens, which are arranged to detect a user input from depressing a screen which displays user information. Touch screens may be a combination of a display arranged below a touch sensitive sheet, which is capable of sensing the location of contact with a finger or a pen. Also, touch screens may be a combination of a display arranged with a touch sensitive switch matrix, e.g. a display integrated touch screen, which is capable of sensing the location of contact with a finger, a pen or any other object. A touch screen may receive user inputs, for example pressing a button or an icon, or selecting certain areas, writing memos, selecting programs, and the like within a user interface of a computer program.
- In order to process the user input, a microprocessor, e.g. a drive engine, which is responsible for the operation of the portable information terminal, i.e. the computer program running on the terminal, needs to receive the detected user inputs and convert them into the appropriate program instructions. In order to receive the user inputs, the drive engine responsible for the operation of the information terminal needs to receive the signals from the touch screen, and to convert these signals into the appropriate program logic.
- For being able to receive and process the signals received from the touch screen, respectively a microprocessor operating the touch screen, e.g. a touch screen engine, the drive engine needs to dedicate at least parts of its processing power to the touch screen engine. The drive engine may operate, besides the touch screen, also loudspeakers, transmission and reception antennas and modules, for example for wireless communication, e.g. GSM, UMTS, WiFi, Near Field Communication (NFC), Bluetooth and the like, keyboards, global positioning devices (GPS), microphones, camera devices, display devices, multimedia processors, and the like. All of these devices may be operated by the drive engine and the interoperation between the devices is controlled by the drive engine.
- In case the drive engine is required to process signals from the touch screen, the touch screen engine may issue an interrupt for the drive engine. Upon reception of the interrupt, the drive engine may dedicate at least a part of its processing power to the touch screen and/or the touch screen engine in order to receive and process the signals received from the touch screen and/or the touch screen engine. When receiving the interrupt from the touch screen engine, the drive engine may be activated, and power consumption of the drive engine may thus increase.
- For example, from U.S. Pat. No. 6,091,031, there is known a portable information terminal. The information terminal has a predetermined area of a touch screen panel, which covers a liquid crystal screen. Further, there is provided a program selection screen. The terminal may be activated by touching the program selection screen, which acts as a system activation area. The system is only activated after depressing the activation area for a predetermined time.
- However, the program selection area is a predetermined area, which covers at least 10-20% of the whole area of the touch screen. Depressing the program selection area for a predetermined time inadvertently may cause the panel to be activated. Activating the system upon depressing the program selection area for a predetermined time thus may not necessarily be user initiated. The power consumption of the system thus may increases due to faulty activation.
- In order to reduce power consumption, there is provided a method comprising determining areas within a touch screen, where user input is possible. Activating touch detection of the touch screen for sensing a user input only within the determined areas, where user input is possible is provided.
- Areas within the touch screen, where a user input is possible, may change during changing operational states of the terminal. For example, when the terminal is in a sleep mode, a.k.a. rest state, sleep state, power save mode, or the like, there is only a small area within the whole touch screen, which may be used as activation area. This activation area may change dynamically and may cover only less then 10% of the touch screen. The activation area might be indicated by an appropriate button or icon or any other means within the display. The software running on the device with the touch screen may control the display to show the activation icon at a certain position within the touch screen. The position of the activation area may change dynamically. Determination of the area, where user input is possible, may allow for detecting automatically, and dynamically, where user input is possible at all. In other areas than the activation areas, no user input may be required and possible. User input within other areas may not be detected and may not cause issuing an interrupt. Only those areas, where user input is possible, may be determined as activation areas. The determination of the area(s), where user input is possible, may depend on the content actually shown on the display. There may be means which analyse the content shown and which determine, which areas are used as activation areas. Touch detection of the touch screen may only be activated for areas, which are determined to be possible input areas. The touch detection itself may already consume energy, as, for example, the touch state of the display needs to be analysed almost in real time. The display area needs to be activated for touch detection. By determining the areas, where user input is possible, only those areas need to be analysed, providing reduced power consumption.
- For example, the display may show three different selection icons, for example “yes”, “no”, “cancel”. Only within these three icons, user input is possible. Determining the areas, where user input is possible, allows for detecting the three icons within the display. Only the areas, which are overlap with these icons may be activated for touch detection. Other areas may be temporally deactivated for touch detection. When a user uses a pen, or his finger, or any other device or means for inputting information into the device, input is only accepted within the detected areas. For example, when the user presses the touch screen outside the detected areas, no action is triggered by the drive engine. Only when the user presses the touch screen at the determined areas, the appropriate program action is triggered by the drive engine.
- When triggering a program action, it may be necessary that the drive engine provides for processing power to process the program action. The touch screen or the touch screen engine may issue an interrupt for the drive engine, when sensing a user input. The interrupt provides for information for the drive engine to provide for processing power for processing user input from the touch screen. When receiving the interrupt, the drive engine may allocate processing power to the processing of touch screen information.
- In order to further reduce power consumption, the activation of the drive engine for processing touch screen engine signals may depend upon reception of the interrupt from the touch screen engine. Embodiments provide activating the drive engine for processing signals from the touch screen engine indicative of a user input upon reception of the interrupt. The drive engine may be activated for processing the signals from the touch screen engine only when the interrupt from the touch screen engine is received. The interrupt may be issued and received, only when a user touches the touch screen at the determined areas. Touching the touch screen at areas outside the determined areas may not cause triggering of the interrupt and the drive engine may not receive the interrupt and may not provide for the necessary processing power for processing touch detection.
- In order to find out, within which areas of the display user input is possible, embodiments provide analysing display information within the touch screen and selecting areas from the display information within which a user input is possible. For example, the user input may be possible within a user selection button, an icon, a character input field, a QWERTY input field, or any other field, which is capable of receiving user input. Determining these fields may, on the one hand be done by analysing the display content. On the other hand it may also be possible to receive from the respective program, which provides for the display content, information about the areas, within which user input is possible. For example, a user interface API (UI API) may provide the information, within which areas user input is possible.
- According to embodiments, a user input may comprise obtaining press position information. The touch screen may allow for detecting the coordinates of a press position. Detecting the coordinates of the press position allows for detecting, whether the press position is within a determined area or not and initiating the respective operation.
- Touch detection may comprise, according to embodiments, making the touch screen sensible for haptic user input. For example, users may use their fingers to input information. Also, input pens may be used. When inputting user information, in a first step the coordinates of the press position is detected. The detected coordinates may, according to embodiments, be converted within the touch screen engine into corresponding signals provided to the drive engine. The drive engine may thus control the software to operate in accordance with the user input.
- A pen may be used, when a resistive touch screen is used. The resistive touch screen may utilise a change in impedance of the touch screen, when the pen is pressed onto the touch screen. When using capacitive touch screens, input may be possible using a finger. A capacitive touch screen utilises the change in capacitance of the touch screen. For example, when a finger approaches the touch screen, the capacitance of the touch screen changes, which may be evaluated and the press position may be detected. An optical touch detection may be operated using a finger or any other means touching the screen.
- According to embodiments, the determination of areas, where a user input is possible, may be provided in a normal state and/or in a power safe state of at least the drive engine. For example, in the power safe state, the activation is only possible within a small icon displayed on the touch screen. Only pressing this icon may allow for activating the terminal. To reduce power consumption, activation of the drive engine, i.e. by issuing the interrupt, shall only be possible, when the activation icon is pressed. The activation icon is determined, and user input is only possible within the activation icon. Interrupts are issued only when this icon is pressed. The drive engine thus consumes less energy in the power safe state, because interrupts are only issued when the activation area is touched. The drive engine is not activated, for example, by inadvertently pressing any other area of the touch screen.
- Another aspect of the specification is an apparatus comprising a touch screen, a touch screen engine, and a drive engine. The touch screen engine is arranged for determining areas, where a user input is possible. A touch screen engine is activated for sensing a user input only within the determined areas, where a user input is possible.
- A further aspect of the specification is a device comprising a means for driving a touch screen, and a means for driving the device, wherein the means for driving the touch screen are arranged for determining areas within a touch screen, where the user input is possible, and wherein the means for driving the touch screen are activated for sensing a user input only within the determined areas, where a user input is possible.
- The device may, according to embodiments, be, for example, a PDA, laptop, tablet PC, video player, music player, multimedia player, camera, mobile phone, or any other user device requiring user inputs.
- Another aspect of the specification is a computer-readable medium having a computer program stored thereon, the computer program comprising instructions operable to cause a processor to determine areas within a touch screen, where a user input is possible, and to activate touch detection of the touch screen for sensing a user input only within the determined areas, where user input is possible.
- A further aspect of the specification is a computer program comprising instructions operable to cause a processor to determine areas within a touch screen, where a user input is possible, and activate touch detection of the touch screen for sensing a user input only within the determined areas, where a user input is possible.
- These and other aspects of the specification will be apparent from and elucidated with reference to the detailed description presented hereinafter. The features of the present specification and of its exemplary embodiments as presented above are understood to be disclosed also in all possible combinations with each other.
- In the figures show:
-
FIG. 1 a block diagram of a mobile phone with its components; -
FIG. 2 a side view of a touch screen; -
FIG. 3 schematically a block diagram of a touch screen system; -
FIG. 4 schematically a diagram of a touch screen system; -
FIG. 5 schematically a display panel with pixel cells; -
FIG. 6 schematically pixel cells with touch detection; -
FIG. 7 a a screenshot of a display within a touch screen; -
FIG. 7 b areas within which user input is possible of a screenshot as illustrated inFIG. 4 a; -
FIG. 7 c a combination of screenshot ofFIG. 4 a and display of areas according toFIG. 4 b; -
FIG. 8 a flowchart of a method according to embodiments. -
FIG. 1 illustrates schematically a block diagram of amobile device 2. Themobile device 2 may be a terminal as previously described. Depending on which kind of devicemobile device 2 is, different appliances and peripherals can be included within amobile device 2. A selection of possible appliances and peripherals are shown inFIG. 1 . It should be noted, that the selection of shown appliances and peripherals is illustrative only and shall not be understood as limiting. - As illustrated in
FIG. 1 ,mobile device 2 is a mobile phone having adrive engine 4. Driveengine 4 may be comprised of hardware and software. Driveengine 4 may be capable of operating all peripherals and any kind of software necessary for operating the peripherals. Driveengine 4 may be a microprocessor, which processes themobile device 2 according to different standards, applications, and the like. Driveengine 4 may be understood as the core engine of themobile device 2, which is responsible for the operation and interoperation of programs and appliances, which are hereinafter explained. - A
touch screen 6 may comprise atouch screen panel 7.Touch screen panel 7 may be placed in front of adisplay 8. The touch screen panel may also be incorporated withindisplay 8.Touch screen panel 7 may be operated by a touch screen engine, i.e. a touch screen controller (not depicted).Touch screen panel 7 anddisplay 8 may be connected to driveengine 4. Touch screen panel may comprise a touch screen controller and may be a component, which is converting physical touches onto its surface or the surface of thedisplay 8 into an electrical format, i.e. signals fordrive engine 4 for operating programs and other appliances.Touch screen 6 will be further illustrated with reference toFIG. 2 . - Spatially beneath
touch screen panel 7, adisplay 8 may be arranged.Display 8 may be a component, which is converting electrical information received from thedrive engine 4 into a readable format. This information may be any information generated from a software for controlling a user interface.Display 8 may be an LED display, OLED display, TFT display, CRT display, plasma display, or any other kind of display capable of converting information into a user readable format. Thedisplay 8 receives display information fromdrive engine 4 and puts out this information as optical information. - Further connected to drive
engine 4 may becamera 10. Thecamera 10 may be a component, which is converting image information into a suitable format for further processing bydrive engine 4. -
Microphone 12 may be a component, which is converting audio information from acoustic waves into electrical information.Microphone 12 may receive user input via acoustic waves and may input these to driveengine 4. - Further, connected to drive
engine 4 isGPS receiver 14, which is a component for converting position information, i.e. from satellites into respective position information fordrive engine 4. - Further,
keyboard 16 may be connected to driveengine 4.Keyboard 16 may be a component, which is converting information from depressed keys into signals fordrive engine 4 for receiving user input. - Further connected to drive
engine 4 is a transmission andreception component 18. Thiscomponent 18 may allow for wired and wireless communication with any kind of other communication hardware. For example, GSM and UMTS communication may be possible viacomponent 18. Further, NFC, WiFi, or any other wireless communication may be possible.Component 18 may allow communicating via LAN, WAN or any other wired communication line. - Information from the
mobile device 2 may be output vialoudspeaker 20.Loudspeaker 20 may be a component for converting electric information into acoustic waves. - The specification relates to the operation of
touch screen 6, i.e.touch screen panel 7,display 8 and driveengine 4. Power consumption ofdrive engine 4 shall be reduced by controllingtouch screen panel 7 appropriately.Touch screen 6 is further illustrated inFIG. 2 . -
FIG. 2 illustrates a side view onto atouch screen 6 with adisplay 8.Display 8 is arranged above alight guide 22 and covered byprotection sheets 24. Betweenprotection sheets 24 anddisplay 8, there is arranged a touch detection sheet 26, which enables thetouch screen 6, i.e. the touch screen controller, to detect a touch position of, for example, atouch pen 28. Thedisplay 8 may driven by adisplay driver 30.Display driver 30 may providedisplay 8 with display information, which is being displayed ondisplay 8 and can be seen from a user'sviewing direction 32. The display information may be received from thedrive engine 4 via a flex-foil connection (not depicted), or any other kind of electrical connection. -
Display 8,light guide 22,protection sheets 24, and touch detection sheets 26 may in common or in any combination thereof be understood astouch screen 6.Touch screen 6 may be connected to thedrive engine 4 via an electrical connection, as will be shown inFIG. 3 . -
Touch screen panel 7 may be comprised of a touch screen engine and touch detection sheets 26. -
Light guide 22 may be connected with a back lighting controller (not depicted) and provides thedisplay 8 with back light, so that the content being displayed ondisplay 8 and provided throughdisplay driver 30 can be seen even in dark viewing conditions. - By means of a
pen 28, a user may select a certain icon or item being displayed ondisplay 8. This may be done by detecting the press position ofpen 28 ontouch screen 6 using the touch detection sheet 26. - The touch detection and position detection is provided by a touch screen controller (not depicted), a.k.a. touch screen driver, being further illustrated in
FIG. 3 . The touch screen driver may be a microprocessor running a program suitable for controlling thetouch screen 6, and/or the touch detection sheet 26 and for obtaining touch information fromtouch screen 6 and/or the touch detection sheet 26. -
FIG. 3 illustrates atouch screen 6 being connected withtouch screen controller 34.Touch screen controller 34 is connected withdrive engine 4 via interruptline 36. When the user touches thetouch screen 6, using thetouch pen 28 or his finger,touch screen 6 provides for touch detection information totouch screen controller 34. Upon touch detection,touch screen controller 34 provides for an interrupt via interruptline 36 to driveengine 4 in order to activatedrive engine 4 for processing user input throughtouch screen 6. - When
touch screen 6 is activated through its whole area, and user input is possible through the whole area oftouch screen 6,touch screen controller 34 issues an interrupt to driveengine 4 everytime touch screen 6 is touched, no matter where the touch detection locates the area, wheretouch screen 6 is touched. This leads to issuing a plurality of interrupts on interruptline 36. -
Engine 4 is activated everytime touch screen 6 is touched, even if the touch detection is not within areas, which allow or require user input. This leads to increased power consumption, asdrive engine 4 needs to allocate processing power for detecting user input throughtouch screen 6 every time it receives an interrupt. - In power safe mode, when the
touch screen 6 should only be activated upon a touching certain area, the commonly knowntouch screen 6 always activates driveengine 4 after touch detection, after which it is checked, whether a terminal is to be activated or not. This leads to increase power consumption. -
FIG. 4 illustrates in more detail atouch screen controller 34. As illustrated,touch screen controller 34 in connected to adrive engine 4 via aninterface 36, which may be aflex foil interface 36. Throughinterface 36,touch screen controller 36 may receive display information and may send touch detection signals.Touch screen controller 34 may be comprised of aframe memory 34 a. The image information is provided column by column through D/A converter 34 b to displaypanel 6. Atiming controller 34 c may provide clocking signals for selecting line addresses. The line addresses are provided todisplay 6 and also to framememory 34 a byaddress coder 34 d. Through the line addresses, thedisplay 6 is activated line by line and the respective pixel information for the respective lines is provided throughframe memory 34 a. -
FIG. 5 illustratesseveral pixel cells 100 within atouch screen 6. Eachpixel cell 100 may represent one pixel. - The
pixel cell 100 may be comprised oftransistor 100 a,capacitor 100 b, andliquid crystal cell 100 c. - The column selection for a
pixel cell 100 c may be done by activating the respective source line 102 (Source: Sn, Sn+1, Sn+2). The source lines 102 may be connected toDAC 34 b for receiving pixel data. The row selection may be done through gate line 104 (Gate: Gn, Gn+1, Gn+2, etc.) signals.Gate lines 104 may be connected to addresscoding 34 d. - When
source line 102 andgate line 104 for aparticular transistor 100 a are activated, the respectiveliquid crystal cell 100 c atpixel cell 100 is activated, and thepixel cell 100 c shows the image data, i.e. light intensity and color, for this respective pixel in the image. - The block diagram of the
pixel cells 100 as illustrated inFIG. 5 is working as follows - Image data is input from
interface 36, which source isdrive engine 4, to theframe memory 34 a on thetouch screen controller 34. Timing controller 34 e sends timing information to addresscoding 34 d which generates control signals for controlling the line selection. - The line selection within
address coding 34 d may read location information from theframe memory 34 a by using a latch pointer and a line pointer. - The digital image data is input to Digital-Analog-Converter (DAC) 34 b. The data is converted to an analogue image data for a
certain column 102, being represented by thesource line 102. - The analogue image data is also inputted to the display panel for line selection. The location of each displayed pixel is controlled by
address coding block 34 d viasource lines 102 and gate line 204 control signals. - The gate line control signal may have digital values (‘0’ or ‘1’), which may be used for selected a line of the pixel on the display panel. The pixel value of a certain column, being stored as digital information of the image data, may then be provided through
source lines 102, respectively. - For a visible pixel the
source line 102 andgate line 104 are activated and the displayed pixel value represents the analogue value of therespective source line 102. - When illuminating one
pixel cell 100 c, the analogue image data, i.e. the current atsource line 102, can flow throughtransistor 100 a andload charging capacitor 100 b. This loading is continued until there is selected another gate line by setting anothergate line 104 HIGH. - The loading of
capacitor 100 b controls the brightness ofliquid crystal cell 100 c of thepixel cell 100. The loadedcapacitor 100 b keeps the analogue value, i.e. the visible grey level of thepixel cell 100, until thesame gate line 104 is selected again and a new loading is carried out. - The
pixel cell 100, which is visible, is working as follows - Analogue image data is output on the source lines 102 (Sn, Sn+1, Sn+2, etc.). A selection, which is the used
gate line 104, where allpixel cells 100 are updated, is further output by setting the respective gate line HIGH. - The
HIGH gate line 104 represents the line of thepixel cell 100, which are updated at the same time. Thepixel cells 100 in other lines are not updated. This update is starting on one of the edges of thedisplay panel 6 and after the start, every next line (e.g. from Gn=>Gn+1=>Gn+2, etc.) is updated until the opposite side of thedisplay panel 6 is reached. Then, the updating can be started from the beginning again. - In order to reduce power consumption, the interrupts need to be reduced. Therefore, embodiments provide for determining areas within a touch screen, where a user input is possible and activating touch detection of the touch screen for sensing a user input only within the determined areas, where a user input is possible. This detection of areas, where user input is possible, is further illustrated in
FIGS. 6-9 . -
FIG. 6 illustratespixel cell 100 as illustrated inFIG. 5 , further comprising transistors fortouch detection 106.Pixel cell 100 further comprisestouch detection sensors 108. The selective touch detection works as follows: - The gate driver includes same amount of the lines what are used for
display panel 6 as illustrated inFIG. 5 . These lines are indicated ascommon gate lines 104 CGn, CGn+1, GCn+1, etc. - When the
common gate line 104 are set HIGH, the transistors fortouch detection 106 are presumably activated in the same way and time when thetransistors 100 a ofpixel cell 100 on thedisplay panel 6 are activated. - For detecting touches on the
display panel 6, it is checked, whether atouch screen sensor 108 ofpixel cell 100 is depressed. That means that only for those lines, where theCGn line 104 is active,touch sensors 108 are read out. - When it is desired that only selected areas of the
display panel 6 can be read out, i.e. are active for touch detection, it may be possible to omit the transistors fortouch detection 106 but to provide HIGH and LOW states to thetouch sensors 108 through separate touch screen gate lines 110 (TGn, Tgn+1). The state of touchscreen gate lines 110 can be selected such that only those touch screen gate lines are HIGH, where user input is possible. This may be detected through analysing the content of the image. By only activating the relevant touchscreen gate lines 110, only within thosetouch sensors 108, which are connected to the touch screen gate lines can be read out. - In order to further select, which column is capable of touch detection, read out lines 112 (R01, R02) may be used. Only those read out
lines 112 may be read out, where touch detection is possible, or desired. This results in the possibility to selectively choose thepixel cells 100, where touch detection is possible. -
FIG. 7 a illustrates a screenshot of auser interface 40. The screenshot is a program window. Within this program window, it is possible to input user information only at certain areas. In the displayeduser interface 40, the program requires the user to input a selection of “yes”, “no”, or “cancel”. As can be seen inFIG. 4 a,user interface 40, being displayed ondisplay 8, allows input only within theareas - Touching the touch screen on any other position than the
buttons buttons touch screen controller 34 toengine 4, when the display is touched at positions outsidebuttons - The result of this determination is illustrated in
FIG. 7 b.FIG. 7 b is a representation ofuser interface 40, where the locations ofbuttons buttons touch screen 6 is activated, i.e. reacts on user input. In other areas thetouch screen 6 is not sensible for touch detection, i.e. when areas other than thebuttons touch screen 6. In other words, the respective touchscreen gate lines 110, where thebuttons buttons lines 112 are actually read out. This results in only detecting touches on thedisplay 6 in the areas of thebuttons -
FIG. 7 c illustrated an overlay of the activated areas andbuttons user interface 40. User input is only possible atbuttons buttons engine 4 viatouch screen controller 34. Touching thetouch screen 6 on any other position does not result in issuing such an interrupt. -
FIG. 8 illustrates a flowchart of a method according to embodiments. Display controller receives (52) display information for being displayed ondisplay 8. The display information is forwarded (54) totouch screen controller 34. Withintouch screen controller 34, the display information is analysed, and it is determined (56), where areas are located, where user input is possible. It is also possible, that a user interface API is requested bytouch screen controller 34 to give information about where areas are located where a user input is possible. - After having determined (56) the areas, where user input is possible, the information for being displayed is displayed (58) on
display 8. Besides that thetouch screen 6 andtouch screen controller 34 are arranged (60) such that they do only react on user input at the determined areas. If no user input at the determined areas is detected, the next image is being evaluated and displayed (52-58). - If a user input is detected within the areas, where user input is possible,
touch screen controller 34 issues (62) an interrupt fordrive engine 4. The issuance of the interrupt initiates withindrive engine 4 the appropriate program logic, and the program is further processed (64) according to the user input. This may be done by further detecting user inputs or by proceeding with the program logic. For example, proceeding program logic may result in storing certain results. - It should be understood that issuing the interrupt and carrying out program logic (62, 64) consumes energy. Thus, issuing the interrupt should only occur, when the
touch screen 6 is touched at areas, where user input is possible. - With the touch screen according to the specification, touch detection is only carried out within the areas where user input is possible. Only touching the touch screen at these positions results in an issuance of an interrupt for
engine 4 and further processing of a program logic. Power consumption is reduced only to cases where the touch screen is touched at areas, where user input is possible and expected. This results in a reduction of power consumption ofdevice 2. - The specification has been described above by means of exemplary embodiments. It should be noted that there are alternative ways and variations which are obvious to a skilled person in the art and can be implemented without deviating from the scope and spirit of the appended claims.
- Furthermore, it is readily clear for a skilled person that the logical blocks in the schematic block diagrams as well as the flowchart and algorithm steps presented in the above description may at least partially be implemented in electronic hardware and/or computer software, wherein it depends on the functionality of the logical block, flowchart step and algorithm step and on design constraints imposed on the respective devices to which degree a logical block, a flowchart step or algorithm step is implemented in hardware or software. The presented logical blocks, flowchart steps and algorithm steps may for instance be implemented in one or more digital signal processors, application specific integrated circuits, field programmable gate arrays or other programmable devices. The computer software may be stored in a variety of storage media of electric, magnetic, electro-magnetic or optic type and may be read and executed by a processor, such as for instance a microprocessor. To this end, the processor and the storage medium may be coupled to interchange information, or the storage medium may be included in the processor.
Claims (23)
1-23. (canceled)
24. A method, comprising:
determining at least one area within a touch screen where user input is possible; and
activating touch detection of the touch screen to sense user input for the at least one area, wherein an apparatus comprises the touch screen, a touch screen engine and a drive engine.
25. A method according to claim 24 , further comprising issuing an interrupt for the drive engine when sensing a user input.
26. A method according to claim 24 , further comprising activating the drive engine to process signals from the touch screen engine indicative of the user input upon reception of an interrupt.
27. A method according to claim 24 , wherein the determining comprises analysing display information within the touch screen and selecting the at least one area from the display information for which user input is possible.
28. A method according to claim 24 , wherein the at least one area comprises at least one of a user selection button, character input field and QWERTY input field.
29. A method according to claim 24 , wherein the sensing the user input comprises obtaining press position information.
30. A method according to claim 24 , wherein the activating comprises enabling haptic user input sensing for the touch screen.
31. A method according to claim 24 , wherein the user input is converted within the touch screen engine into corresponding signals provided to the drive engine.
32. A method according to claim 24 , wherein sensing user input comprises at least one of capacitive, resistive and optical touch detection.
33. A method according to claim 24 , wherein determining the at least one area is provided in at least one of a normal state and a power save state of the drive engine.
34. An apparatus, comprising:
a touch screen;
a touch screen engine; and
a drive engine,
the touch screen engine being configured to determine at least one area within the touch screen where user input is possible and to be activated to sense user input within the at least one area.
35. An apparatus according to claim 34 , wherein the touch screen engine is configured to issue an interrupt for the drive engine when user input is sensed.
36. An apparatus according to claim 35 , wherein the drive engine is configured to be activated to process at least one signal from the touch screen engine indicative of a user input upon reception of the interrupt.
37. An apparatus according to claim 34 , wherein the touch screen engine is configured to analyze display information within the touch screen and to select the at least one area from the display information for which a user input is possible.
38. An apparatus according to claim 34 , wherein the at least one area comprises at least one of a user selection button, character input field and QWERTY input field.
39. An apparatus according to claim 34 , wherein the touch screen engine is configured to sense the user input by obtaining press position information.
40. An apparatus according to claim 34 , wherein the touch screen is configured to sense haptic user input.
41. An apparatus according to claim 34 , wherein the touch screen engine is configured to convert the user input into corresponding signals, and further to provide the converted signals to the drive engine.
42. An apparatus according to claim 34 , wherein the touch screen is configured for at least one of capacitive touch detection, resistive touch detection and optical touch detection.
43. An apparatus according to claim 34 , wherein the touch screen engine is further configured to determine the at least one area during a normal state and a power save state of the drive engine.
44. A computer-readable medium having a computer program stored thereon, the computer program comprising:
instructions operable to cause a processor to determine at least one area within a touch screen where user input is possible and activate touch detection of the touch screen to sense a user input within the at least one area.
45. A computer program product, comprising:
instructions operable to cause a processor to determine at least one area within a touch screen where user input is possible and activate touch detection of the touch screen to sense a user input within the at least one area.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2007/063390 WO2009071123A1 (en) | 2007-12-06 | 2007-12-06 | Power reduction for touch screens |
Publications (1)
Publication Number | Publication Date |
---|---|
US20100265209A1 true US20100265209A1 (en) | 2010-10-21 |
Family
ID=39810242
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/746,732 Abandoned US20100265209A1 (en) | 2007-12-06 | 2007-12-06 | Power reduction for touch screens |
Country Status (2)
Country | Link |
---|---|
US (1) | US20100265209A1 (en) |
WO (1) | WO2009071123A1 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100039404A1 (en) * | 2008-08-18 | 2010-02-18 | Sentelic Corporation | Integrated input system |
US20110063491A1 (en) * | 2009-09-14 | 2011-03-17 | Samsung Electronics Co., Ltd. | Digital photographing apparatus and method of controlling the same |
US20110199336A1 (en) * | 2010-02-12 | 2011-08-18 | Pixart Imaging Inc. | Optical touch device |
US20110205176A1 (en) * | 2008-11-05 | 2011-08-25 | Takashi Okada | Portable electronic device, and power saving method and power saving program for the same |
CN102855011A (en) * | 2011-06-27 | 2013-01-02 | 比亚迪股份有限公司 | Touch screen control method, touch device and mobile terminal |
WO2013132241A3 (en) * | 2012-03-05 | 2013-12-05 | Elliptic Laboratories As | Touchless user interfaces |
US20130321322A1 (en) * | 2011-02-25 | 2013-12-05 | Lg Electronics Inc. | Mobile terminal and method of controlling the same |
US20140253487A1 (en) * | 2011-10-18 | 2014-09-11 | Slyde Watch Sa | Method and circuit for switching a wristwatch from a first power mode to a second power mode |
CN104076996A (en) * | 2013-03-26 | 2014-10-01 | 株式会社日本显示器 | Display device and electronic apparatus |
EP2551747A3 (en) * | 2011-07-29 | 2015-02-18 | ACER Inc. | Power saving method and touch display apparatus |
US20150193031A1 (en) * | 2014-01-07 | 2015-07-09 | Qualcomm Incorporated | System and method for context-based touch processing |
CN105431803A (en) * | 2013-07-30 | 2016-03-23 | 三星电子株式会社 | Display apparatus and control method thereof |
US9552094B2 (en) | 2011-12-22 | 2017-01-24 | Optis Circuit Technology, Llc | User interface responsiveness in an electronic device having a touch screen display |
US9791959B2 (en) | 2014-01-07 | 2017-10-17 | Qualcomm Incorporated | System and method for host-augmented touch processing |
US10055037B2 (en) * | 2012-02-23 | 2018-08-21 | Pantech Inc. | Mobile terminal and method for operating a mobile terminal based on touch input |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
BR112012001334A2 (en) * | 2009-07-30 | 2016-03-15 | Sharp Kk | portable display device, portable display device control method, recording program and medium |
KR20110095586A (en) * | 2010-02-19 | 2011-08-25 | 삼성전자주식회사 | Collecting method and apparatus of touch event for device |
CN107924251B (en) * | 2015-08-24 | 2021-01-29 | 华为技术有限公司 | Method and device for reducing power consumption of touch screen device |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5260697A (en) * | 1990-11-13 | 1993-11-09 | Wang Laboratories, Inc. | Computer with separate display plane and user interface processor |
US5949408A (en) * | 1995-09-28 | 1999-09-07 | Hewlett-Packard Company | Dual orientation display handheld computer devices |
US6091031A (en) * | 1997-04-11 | 2000-07-18 | Samsung Electronics Co., Ltd. | Portable information terminal and an activating method thereof |
US6373472B1 (en) * | 1995-10-13 | 2002-04-16 | Silviu Palalau | Driver control interface system |
US6476797B1 (en) * | 1999-04-27 | 2002-11-05 | International Business Machines Corporation | Display |
US20030081860A1 (en) * | 1986-08-15 | 2003-05-01 | Danielson Arvin D. | Data capture apparatus with handwritten data receiving component |
US20030095105A1 (en) * | 2001-11-16 | 2003-05-22 | Johannes Vaananen | Extended keyboard |
US20030137495A1 (en) * | 2002-01-22 | 2003-07-24 | Palm, Inc. | Handheld computer with pop-up user interface |
US20040001049A1 (en) * | 2002-06-27 | 2004-01-01 | Oakley Nicholas W. | Multiple mode display apparatus |
US20050079896A1 (en) * | 2003-10-14 | 2005-04-14 | Nokia Corporation | Method and apparatus for locking a mobile telephone touch screen |
US20050085215A1 (en) * | 2003-10-21 | 2005-04-21 | Nokia Corporation | Method and related apparatus for emergency calling in a touch screen mobile phone from a touch screen and keypad lock active state |
US20050253817A1 (en) * | 2002-06-19 | 2005-11-17 | Markku Rytivaara | Method of deactivating lock and portable electronic device |
US6998856B2 (en) * | 2001-06-29 | 2006-02-14 | Ethertouch | Apparatus for sensing the position of a pointing object |
US20060132447A1 (en) * | 2004-12-16 | 2006-06-22 | Conrad Richard H | Method and apparatus for automatically transforming functions of computer keyboard keys and pointing devices by detection of hand location |
US20060161871A1 (en) * | 2004-07-30 | 2006-07-20 | Apple Computer, Inc. | Proximity detector in handheld device |
US20060238503A1 (en) * | 2004-09-18 | 2006-10-26 | Smith Jerome W | Deputy series-music cinema or transfortainment PC (phirst classic) or Digi-Tasking Mobile-Cycle devices for maximum digital mobilosity or the digital entertaining PC or the all digital activity center or satellite entertainment mogul or satellite entertainment PC (phirst classic) |
US20070128899A1 (en) * | 2003-01-12 | 2007-06-07 | Yaron Mayer | System and method for improving the efficiency, comfort, and/or reliability in Operating Systems, such as for example Windows |
US20070150842A1 (en) * | 2005-12-23 | 2007-06-28 | Imran Chaudhri | Unlocking a device by performing gestures on an unlock image |
US7256770B2 (en) * | 1998-09-14 | 2007-08-14 | Microsoft Corporation | Method for displaying information responsive to sensing a physical presence proximate to a computer input device |
US20080188267A1 (en) * | 2007-02-07 | 2008-08-07 | Sagong Phil | Mobile communication terminal with touch screen and information inputing method using the same |
US20090128504A1 (en) * | 2007-11-16 | 2009-05-21 | Garey Alexander Smith | Touch screen peripheral device |
-
2007
- 2007-12-06 US US12/746,732 patent/US20100265209A1/en not_active Abandoned
- 2007-12-06 WO PCT/EP2007/063390 patent/WO2009071123A1/en active Application Filing
Patent Citations (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030081860A1 (en) * | 1986-08-15 | 2003-05-01 | Danielson Arvin D. | Data capture apparatus with handwritten data receiving component |
US5260697A (en) * | 1990-11-13 | 1993-11-09 | Wang Laboratories, Inc. | Computer with separate display plane and user interface processor |
US5949408A (en) * | 1995-09-28 | 1999-09-07 | Hewlett-Packard Company | Dual orientation display handheld computer devices |
US6373472B1 (en) * | 1995-10-13 | 2002-04-16 | Silviu Palalau | Driver control interface system |
US6091031A (en) * | 1997-04-11 | 2000-07-18 | Samsung Electronics Co., Ltd. | Portable information terminal and an activating method thereof |
US7256770B2 (en) * | 1998-09-14 | 2007-08-14 | Microsoft Corporation | Method for displaying information responsive to sensing a physical presence proximate to a computer input device |
US6476797B1 (en) * | 1999-04-27 | 2002-11-05 | International Business Machines Corporation | Display |
US6998856B2 (en) * | 2001-06-29 | 2006-02-14 | Ethertouch | Apparatus for sensing the position of a pointing object |
US20030095105A1 (en) * | 2001-11-16 | 2003-05-22 | Johannes Vaananen | Extended keyboard |
US20030137495A1 (en) * | 2002-01-22 | 2003-07-24 | Palm, Inc. | Handheld computer with pop-up user interface |
US20050253817A1 (en) * | 2002-06-19 | 2005-11-17 | Markku Rytivaara | Method of deactivating lock and portable electronic device |
US20040001049A1 (en) * | 2002-06-27 | 2004-01-01 | Oakley Nicholas W. | Multiple mode display apparatus |
US20070247432A1 (en) * | 2002-06-27 | 2007-10-25 | Oakley Nicholas W | Multiple mode display apparatus |
US20070128899A1 (en) * | 2003-01-12 | 2007-06-07 | Yaron Mayer | System and method for improving the efficiency, comfort, and/or reliability in Operating Systems, such as for example Windows |
US20050079896A1 (en) * | 2003-10-14 | 2005-04-14 | Nokia Corporation | Method and apparatus for locking a mobile telephone touch screen |
US20050085215A1 (en) * | 2003-10-21 | 2005-04-21 | Nokia Corporation | Method and related apparatus for emergency calling in a touch screen mobile phone from a touch screen and keypad lock active state |
US20060161871A1 (en) * | 2004-07-30 | 2006-07-20 | Apple Computer, Inc. | Proximity detector in handheld device |
US20060238503A1 (en) * | 2004-09-18 | 2006-10-26 | Smith Jerome W | Deputy series-music cinema or transfortainment PC (phirst classic) or Digi-Tasking Mobile-Cycle devices for maximum digital mobilosity or the digital entertaining PC or the all digital activity center or satellite entertainment mogul or satellite entertainment PC (phirst classic) |
US20060132447A1 (en) * | 2004-12-16 | 2006-06-22 | Conrad Richard H | Method and apparatus for automatically transforming functions of computer keyboard keys and pointing devices by detection of hand location |
US20070150842A1 (en) * | 2005-12-23 | 2007-06-28 | Imran Chaudhri | Unlocking a device by performing gestures on an unlock image |
US20080188267A1 (en) * | 2007-02-07 | 2008-08-07 | Sagong Phil | Mobile communication terminal with touch screen and information inputing method using the same |
US8174496B2 (en) * | 2007-02-07 | 2012-05-08 | Lg Electronics Inc. | Mobile communication terminal with touch screen and information inputing method using the same |
US20090128504A1 (en) * | 2007-11-16 | 2009-05-21 | Garey Alexander Smith | Touch screen peripheral device |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100039404A1 (en) * | 2008-08-18 | 2010-02-18 | Sentelic Corporation | Integrated input system |
US20110205176A1 (en) * | 2008-11-05 | 2011-08-25 | Takashi Okada | Portable electronic device, and power saving method and power saving program for the same |
US20110063491A1 (en) * | 2009-09-14 | 2011-03-17 | Samsung Electronics Co., Ltd. | Digital photographing apparatus and method of controlling the same |
US8441567B2 (en) * | 2009-09-14 | 2013-05-14 | Samsung Electronics Co., Ltd. | Digital photographing apparatus and method of controlling the same |
US20110199336A1 (en) * | 2010-02-12 | 2011-08-18 | Pixart Imaging Inc. | Optical touch device |
US20130321322A1 (en) * | 2011-02-25 | 2013-12-05 | Lg Electronics Inc. | Mobile terminal and method of controlling the same |
CN102855011A (en) * | 2011-06-27 | 2013-01-02 | 比亚迪股份有限公司 | Touch screen control method, touch device and mobile terminal |
EP2551747A3 (en) * | 2011-07-29 | 2015-02-18 | ACER Inc. | Power saving method and touch display apparatus |
US10198085B2 (en) | 2011-10-18 | 2019-02-05 | Slyde Watch Sa | Method and circuit for switching a wristwatch from a first power mode to a second power mode |
US20140253487A1 (en) * | 2011-10-18 | 2014-09-11 | Slyde Watch Sa | Method and circuit for switching a wristwatch from a first power mode to a second power mode |
US9804678B2 (en) * | 2011-10-18 | 2017-10-31 | Slyde Watch Sa | Method and circuit for switching a wristwatch from a first power mode to a second power mode |
US9552094B2 (en) | 2011-12-22 | 2017-01-24 | Optis Circuit Technology, Llc | User interface responsiveness in an electronic device having a touch screen display |
US10055037B2 (en) * | 2012-02-23 | 2018-08-21 | Pantech Inc. | Mobile terminal and method for operating a mobile terminal based on touch input |
US9886139B2 (en) | 2012-03-05 | 2018-02-06 | Elliptic Laboratories As | Touchless user interface using variable sensing rates |
WO2013132241A3 (en) * | 2012-03-05 | 2013-12-05 | Elliptic Laboratories As | Touchless user interfaces |
US20170102807A1 (en) * | 2013-03-26 | 2017-04-13 | Japan Display Inc. | Display device and electronic apparatus |
US20160109987A1 (en) * | 2013-03-26 | 2016-04-21 | Japan Display Inc. | Display device and electronic apparatus |
US9715322B2 (en) * | 2013-03-26 | 2017-07-25 | Japan Display Inc. | Display device and electronic apparatus |
US9791954B2 (en) * | 2013-03-26 | 2017-10-17 | Japan Display Inc. | Display device and electronic apparatus |
US9251758B2 (en) * | 2013-03-26 | 2016-02-02 | Japan Display Inc. | Display device and electronic apparatus |
US20140292711A1 (en) * | 2013-03-26 | 2014-10-02 | Japan Display Inc. | Display device and electronic apparatus |
CN104076996A (en) * | 2013-03-26 | 2014-10-01 | 株式会社日本显示器 | Display device and electronic apparatus |
CN105431803A (en) * | 2013-07-30 | 2016-03-23 | 三星电子株式会社 | Display apparatus and control method thereof |
CN105900054A (en) * | 2014-01-07 | 2016-08-24 | 高通股份有限公司 | System and method for context-based touch processing |
US9710150B2 (en) * | 2014-01-07 | 2017-07-18 | Qualcomm Incorporated | System and method for context-based touch processing |
US9791959B2 (en) | 2014-01-07 | 2017-10-17 | Qualcomm Incorporated | System and method for host-augmented touch processing |
US20150193031A1 (en) * | 2014-01-07 | 2015-07-09 | Qualcomm Incorporated | System and method for context-based touch processing |
Also Published As
Publication number | Publication date |
---|---|
WO2009071123A1 (en) | 2009-06-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20100265209A1 (en) | Power reduction for touch screens | |
AU2018282404B2 (en) | Touch-sensitive button | |
WO2021143805A1 (en) | Widget processing method and related apparatus | |
US8739053B2 (en) | Electronic device capable of transferring object between two display units and controlling method thereof | |
US20130016046A1 (en) | Control method and system of touch panel | |
US20080259046A1 (en) | Pressure sensitive touch pad with virtual programmable buttons for launching utility applications | |
US20120176414A1 (en) | Methods circuits apparatus and systems for human machine interfacing with an electronic appliance | |
US20100201615A1 (en) | Touch and Bump Input Control | |
CN101779188A (en) | systems and methods for providing a user interface | |
CN108733298B (en) | Touch information processing method and device, storage medium and electronic equipment | |
TW200937270A (en) | Touch sensor for a display screen of an electronic device | |
US20140015783A1 (en) | Method and apparatus for detecting an attachable/detachable pen | |
US9176631B2 (en) | Touch-and-play input device and operating method thereof | |
CN101470575B (en) | Electronic device and its input method | |
CN106886351B (en) | Display method and device of terminal time information and computer equipment | |
US8643620B2 (en) | Portable electronic device | |
US10620820B2 (en) | Electronic devices having touch-sensing module and method for generating displayed image | |
WO2020010917A1 (en) | Split-screen display opening method and device, storage medium and electronic equipment | |
US20090201259A1 (en) | Cursor creation for touch screen | |
US20100283729A1 (en) | Control method for controlling handheld electronic device | |
CN108737656B (en) | Processing method and device of floating button, storage medium and electronic equipment | |
US20140085340A1 (en) | Method and electronic device for manipulating scale or rotation of graphic on display | |
KR101919515B1 (en) | Method for inputting data in terminal having touchscreen and apparatus thereof | |
TWI486742B (en) | Power management method for an electronic device and related device | |
US20130162567A1 (en) | Method for operating tool list and portable electronic device using the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NOKIA CORPORATION, FINLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NURMI, JUHA HARRI-PEKKA;SAARINEN, KAJ;RAUTANEN, TERO JUHANI;REEL/FRAME:024496/0769 Effective date: 20100531 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |