WO2011041547A1 - Systems and methods to facilitate active reading - Google Patents

Systems and methods to facilitate active reading Download PDF

Info

Publication number
WO2011041547A1
WO2011041547A1 PCT/US2010/050911 US2010050911W WO2011041547A1 WO 2011041547 A1 WO2011041547 A1 WO 2011041547A1 US 2010050911 W US2010050911 W US 2010050911W WO 2011041547 A1 WO2011041547 A1 WO 2011041547A1
Authority
WO
WIPO (PCT)
Prior art keywords
document
virtual workspace
gesture
magnification
response
Prior art date
Application number
PCT/US2010/050911
Other languages
French (fr)
Inventor
Craig S. Tashman
W. Keith Edwards
Original Assignee
Georgia Tech Research Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Georgia Tech Research Corporation filed Critical Georgia Tech Research Corporation
Priority to US13/876,463 priority Critical patent/US10268661B2/en
Priority to PCT/US2011/028595 priority patent/WO2012044363A1/en
Publication of WO2011041547A1 publication Critical patent/WO2011041547A1/en
Priority to US16/391,003 priority patent/US11704473B2/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0483Interaction with page-structured environments, e.g. book metaphor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04805Virtual magnifying lens, i.e. window or frame movable on top of displayed information to enlarge it for better reading or selection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Definitions

  • Various embodiments of the present invention relate to digital documents and, more particularly, to systems and methods to facilitate active reading of digital documents.
  • paper supports bimanual interaction and freeform annotation within the boundaries of a single page
  • paper lacks the flexibility to provide, for example, content rearrangement, document overviews, and annotation outside of page boundaries.
  • the tangibility of paper supports some rapid forms of navigation, such as dog-earing and bookmarking with a finger
  • paper provides little flexibility to create a customized navigational structure.
  • Modern pen-based computerized tablets do a fine job of imitating paper, which benefits users by providing a familiar medium, but as a result, these pen-based tablets suffer from the same limitations as paper. Thus, neither paper nor modern computer systems adequately facilitate active reading.
  • a document review system can provide a fluid-like environment in which users can freely and flexibly manipulate, rearrange, and annotate documents without many of the restriction inherent in paper. It is to such systems and related methods that various embodiments of the present invention are directed. Briefly described, various embodiments of the present invention are review systems and methods for facilitating active reading of documents, by providing a fluid-like interface with advantages over physical paper and conventional word processing systems. According to embodiments of the present invention, a document review system can provide a novel approach to representing and interacting with documents. In contrast to the paper model, which offers a stable but rigid representation, the document review system can provide a highly flexible, malleable document representation.
  • the document review system can provide high degree-of- freedom ways to navigate through and manipulate the document representation, control what document content is displayed and where, and create annotations and other structures related to the document.
  • the document review system can include a multi-touch, gesture- based user interface.
  • Embodiments of the document review system can provide improvements, as compared to paper and conventional word processing, to each of these processes.
  • Annotation can be generally defined as text embellishment, including highlighting and marginalia.
  • the review system can provide efficient annotation by enabling convenient switching between annotation tools, by supporting idiosyncratic markings, and by providing a convenient means for retrieving annotations made previously.
  • Content extraction generally includes copying or moving content from a document to a secondary location, such as when outlining or note-taking.
  • the review system can closely integrate extraction with the reading process, so that user can organize and view extracted content, as well as link extracted content back to the original document.
  • Navigation generally entails moving throughout a document and between multiple documents, such as when searching for text, turning a page, or flipping between selected locations for comparison.
  • the review system can support bookmarks and parallelism to facilitate these or other navigational tasks.
  • Layout generally refers to the visual or spatial arrangement of the document and related objects.
  • the review system can optimize layout according to the user's preferences by enabling distinct portions of the document to be viewed in parallel, while maintaining the document's linearity.
  • a document review system can comprise a virtual workspace, a document view region, a preview region, and optional document objects.
  • the system can be embodied in one or more computer-readable media and can be executable by one or more computer processors on a computing device.
  • the computing device can comprise a multi-touch interface by which a user can interact with the virtual workspace and the overall document review system.
  • the virtual workspace can be a working environment in which the user can review a document.
  • the virtual workspace can be, for example, a graphical user interface displayed in a virtual window or frame viewable through the multi-touch interface.
  • the virtual workspace can be designed to look and feel like a physical desktop or other physical workspace to which the user may be accustomed.
  • the virtual workspace can be a relatively unstructured environment, enabling users to place the document objects as desired throughout the virtual workspace.
  • the document view region can be contained fully or partially within the virtual workspace.
  • the view region can be configured to display a viewable portion of at least one document at a size that enables a user to easily read the text of the document.
  • the size of the document in the view region can, however, be increased or decreased as the user desired. If the document is too long to be contained fully within the view region at a given magnification state of the document, then only a portion of the document can be viewable in the view region.
  • the document can be displayed in a continuous layout, and in an exemplary embodiment, page breaks in the document can be hidden, so that the document appears to be seamless and unbounded by pagination.
  • the preview region can be contained fully or partially within the virtual workspace.
  • the preview region can display a larger portion of the document, at a smaller size, than the view region.
  • the magnification of the preview region can be such that the entire document can be displayed continuously in the preview region.
  • the magnification can be such that the general layout of the document can be determined by the preview region, although the text of the document need not be readable within the region.
  • the preview region can be linked to the document view region and can serve various navigational or other purposes.
  • the portion of the document displayed in the document view region can change, so as to center in the document view region the portion of the document touched in the preview region.
  • the preview region can be utilized to select a portion of the document that is displayed in the document view region.
  • the document objects can be moveable objects positioned throughout the virtual workspace as desired by the user. In some embodiments, however, such movement can be restricted to areas outside of one or both of the document view region and the preview region, so as not to obstruct these regions.
  • a document object can be created by the user to assist the user in actively reading of the document. For example, and not limitation, the user can create an excerpt of the document or an annotation, either of which can be encapsulated in a document object, which may be smaller and more easily manipulable than the document as a whole. Once created, the document object can be freely moved about the virtual workspace, so as to enable the user to arrange the virtual workspace in a manner that creates a customized active reading experience.
  • the document object can be linked to the portion or portions of the document to which the document object relates.
  • the document object can include a visual link, such as an arrow, that the user can touch to cause the one or more documents in the document view region to shift position, thus bringing the related portions into view.
  • the document review system can thus enable users to manipulate documents in a way that improves upon paper and other document manipulation systems.
  • Fig. 1 illustrates a review system, according to an exemplary embodiment of the present invention.
  • Fig. 2 illustrates an architecture of a computing device for providing the review system, according to an exemplary embodiment of the present invention.
  • Fig. 3 illustrates a transient bookmark of the review system, according to an exemplary embodiment of the present invention.
  • Fig. 4 illustrates a flow diagram of a method of creating a transient bookmark, according to an exemplary embodiment of the present invention.
  • Figs. 5A-5B illustrate collapsing of a document, according to an exemplary embodiment of the present invention.
  • Figs. 6A-6B illustrate an excerpt of the review system 100, according to an exemplary embodiment of the present invention.
  • Figs. 7A-7B illustrate an annotation of the review system 100, according to an exemplary embodiment of the present invention.
  • Fig. 1 illustrates a review system 100, or document review system, according to an exemplary embodiment of the present invention.
  • the review system 100 can comprise, for example, a touchscreen input device 110 of a computing device 200, a virtual workspace 120, a document view region 130, a preview region 140, an optional one or more document objects 150, and a toolbar 160.
  • the touchscreen input device 110 can be a multi-touch input device for interfacing with the virtual workspace 120 and other aspects of the review system 100.
  • the touchscreen input device 110 is a multi-touch device capable of receiving multiple simultaneous touches, thus enabling a user to interact with the review system 100 in a natural manner, using multiple hands and fingers simultaneously.
  • a detection system 115 can be integrated with or in communication with the touchscreen input device 110, to detect user interactions with the touchscreen input device 110. These user interactions, or gestures, can be interpreted as commands to the review system 100.
  • the review system 100 can alternatively comprise some other multi-point, bimanual, spatial input device capable of receiving a wide away of gestures interpretable as commands.
  • the review system 100 can be designed to improve four major processes that occur in active reading: annotation, content extraction, navigation, and layout. Conventional paper-like approaches fall short in facilitating these processes because of their fixed structure and lack of flexibility. Utilizing a multi-touch input device 110 can provide parallel and bimanual input, which are important parts of paper-based reading, and which also enable a flexible environment.
  • a mouse as used in most computer-based reading systems, is an inefficient control device because it provides only a single indicator or selector.
  • a keyboard also used in most computer- based reading systems, lacks a natural spatial mapping. The flexible interactions made possible by embodiments of the review system 100 inherently offer more degrees of freedom than traditionally offered by computer-based reading systems.
  • the multi-touch input device 110 can support multi-point spatial input and is thus capable of efficiently receiving a wide array of gestures for interacting with the review system 100.
  • the terms “touch,” “hold,” and the like need not refer only to physical contact between the user and the touchscreen input device 110. Such terms can refer to various interactions simulating a physical contact, such as pointing from a distance or bringing a finger, hand, or implement in close proximity to the touchscreen input device 110, so as to indicate a virtual touching, holding, or the like.
  • the definition of a "touch” can be implementation-dependent, wherein the type of touchscreen input device 110 used can determine how interactions are detected and thus how a "touch” or “hold” is defined.
  • the touchscreen input device 110 can utilize resistive, capacitive, or camera technologies.
  • a "touch” can be defined based on camera sensitivity, or on an instrument's being within a predetermined distance from the touchscreen input device 110. Additionally, “touch,” “hold,” and like terms need not refer only to interactions between the user's hands or fingers and the touchscreen input device 110, but can also refer to interactions with an instrument held by the user, such as a stylus, marker, or pen.
  • a display system 118 can be in communication with the detection system 115, the touchscreen input device 110, or both. The display system 118 can react to user gestures by displaying and refreshing a graphical user interface presented to the user, preferably through the touchscreen input device 110, which can perform as both an input and an output device. This graphical user interface can include the virtual workspace 120, the review region 130, the document view region 140, and the document objects 150, all of which will be described in more detail below.
  • the virtual workspace 120 can be accessible and manipulable through the touchscreen input device 110.
  • the virtual workspace 120 can simulate a physical desktop, in that the user can freely move document objects 150 throughout the workspace 120 without being bound by a fixed structure common in computer-based reading system.
  • the virtual workspace 120 can contain the preview region 130 and the document view region 140. In other embodiments, however, the virtual workspace 120 can comprise the useable space of the review system 100 outside of the preview region 130 and the document view region 140.
  • the 100 can present the user with the virtual workspace 120 containing a document 50, or configured to display a yet-to-be-opened document 50.
  • the user can control the document 50 and other document objects 150 in the virtual workspace 120 with a vocabulary of multi-touch gestures. Through these gestures, the user can navigate, annotate, and manipulate the virtual workspace 120, rarely having to explicitly select tools or otherwise shift attention away from the document 50 at hand.
  • Some basic interactions can be performed in the virtual workspace 120 as one might expect based on conventional touch applications. For example, objects can be repositioned by dragging the objects about the virtual workspace 120. Rescaling can be performed by a pinching or stretching gesture with two fingers, preferably in a horizontal orientation.
  • Other performable gestures and operations are new to the review system 100, as will be described below in detail.
  • the user can open a document 50 in the virtual workspace 120, and the open document 50 can be displayed in the one or both of the preview region 130 and the document view region 140.
  • various actions are described as being performed or performable on the "text" of the open document 50. It will be understood, however, that all or most of such actions can similarly be performed on imbedded objects in the document 50 that are not text, such as images or multimedia.
  • text throughout this disclosure is used for illustrative purposes only and is not restrictive.
  • the preview region 130 can be configured to display the document 50 at a magnification that enables the user to view the general layout of the document 50.
  • the entire document 50 can be viewable in the preview region 130, so as to present the general layout of the entire document 50 to the user.
  • the magnification of the preview region 130 can be adjustable, so that the user can select a magnification size that is best suited to the user's needs.
  • the document view region 140 can display at least a portion of the open document 50.
  • the document view region 140 can display the document 50 at a magnification enabling the user to easily read the text of the document 50.
  • the magnification of the document 50 in the document view region 140 can be modified by the user to enable to user to select a text size best suiting the user's needs.
  • the text of the document 50 can, in either or both of the preview region 130 and the document view region 140, be presented to the user in a continuous format, with or without pagination. If pagination is provided, then this provision can be for the user's reference only and need not restrict operations of the review system 100 to page boundaries. Some embodiments of the review system 100 can enable the user to select whether pagination is shown, to further customize the user's active reading experience. In the document view region 140 and in the preview region 130, if the entire document 50 is not visible, then the user can scroll vertically in the respective region 140 or 130 to adjust the visible portion of the document 50.
  • Scrolling can occur when the user performs a predetermined gesture, such as touching the representation of the document 50 and, while maintaining contact with the touchscreen input device 110, sliding the fingers upward or downward. Sliding downward can cause the document 50 to move downward, thus displaying a previously invisible portion above the previously displayed portion of the document 50. Analogously, sliding upward can cause the document 50 to move upward, thus displaying a previously invisible portion below the previously displayed portion of the document 50.
  • a predetermined gesture such as touching the representation of the document 50 and, while maintaining contact with the touchscreen input device 110, sliding the fingers upward or downward. Sliding downward can cause the document 50 to move downward, thus displaying a previously invisible portion above the previously displayed portion of the document 50.
  • sliding upward can cause the document 50 to move upward, thus displaying a previously invisible portion below the previously displayed portion of the document 50.
  • the review system 100 can also support "fast scrolling" in the preview region 130, the document view region 140, or both. Scrolling at normal speed can occur as described above, in which case the displayed portion of the document 50 can be adjusted up or down corresponding to the distance the user's finger slides while in contact with the touchscreen input device 110. For example, and not limitation, with normal scrolling, the document 50 can be moved by a distance equivalent to the distance moved by the user's finger while the user's finger is holding the touchscreen input device 110. While normal scrolling is thus an intuitive means to navigate a document, normal scrolling can be inefficient for long document, when the user seeks to navigate between portions of the document 50 separated by a great distance.
  • the review system 100 can also support fast scrolling.
  • Fast scrolling can take advantage of modern touch sensors by determining an amount of pressure applied by the user in touching and holding the document 50 within the review system 100. For example, if the user drags the document 50 with a light touch, the movement of the document 50 in response can simply follow the finger, resulting in normal- speed scrolling. In contrast, if a firmer touch is used, then the movement of the document 50 can correspond to the pressure of the user's touch.
  • the document 50 can move in the same direction as the finger, but at a speed corresponding to the pressure applied by the user, where increased pressure corresponds to increased speed and distance, and where decreased pressure corresponds to decreased speed and distance. For example, if the user drags his or her finger over a distance of one inch, the document 50 can move by one, two, three, or six inches, depending on how hard the user presses the touchscreen input device 110.
  • the document objects 150 can be objects created by the user to facilitate the user's active reading process.
  • a particular document object 150 can be created by the user, with tools of the review system 100, to represent and include an excerpt or annotation of the document 50.
  • the document object 150 can contain text, an image, or another annotation or portion of the document 50.
  • the document object 150 can also comprise a link to the portion of the document 50 to which the document object 150 refers.
  • an excerpt can contain a link back to the portion of the document 50 from which the excerpt was extracted.
  • the link 155 can have a visible representation, such as an arrow, which can point from the document object 150 to the document view region 140 to indicate that the linked portion of the document 50 can be displayed in the document view region 140.
  • the document 50 in the document view region 140 can automatically scroll to display the portion of the document 50 referred to by the document object 150.
  • selecting the link can cause the referred-to portion to be centered within the document view region 140.
  • Selection of the link 155 can occur when the user touches the visible representation of the link 155.
  • the toolbar 160 can provide some guidance to the user of how to operate the review system 100. Although, in an exemplary embodiment, some or all tools of the review system 100 can be accessible via gestures, without the user's having to use the toolbar 160, the toolbar 160 can act a reminder of the gestures and tools available to the user. For example, the toolbar 160 can display a plurality of representations of categories 165, tools, or functions. Each category 165 can be related to one or more tools that the user can use to manipulate and review the document 50. In an exemplary embodiment, the toolbar 160 can teach the user how to use the review system 100.
  • the toolbar 160 when the user selects the toolbar 160, such as by touching the toolbar 160 or performing some other predetermined selection gesture, the toolbar 160 can expand slightly for a temporary period.
  • each button or category 165 of the toolbar 160 can be larger than when the toolbar 160 is not expanded, and each button or category 165 can display a visual cue as to how a gesture can be performed to initiate one or more tools related to the corresponding category 165.
  • a visual cue for only the gesture corresponding to that category 165 can be displayed.
  • the visual cues of the toolbar 160 can be, for example, text, animations, or still images of one or more hands performing the associated gestures.
  • the user can utilize the toolbar 160 to learn gestures recognized by the review system 100.
  • the toolbar 160 can behave similarly to a toolbar 160 in a conventional word-processing application, whereby the user can initiate functions form the toolbar 160.
  • the toolbar 160 can both teach the related gestures and additionally initiate one or more tools of the review system 100.
  • the review system 100 can be embodied in a computer-readable medium and executed by a computer processor to provide one, some, or all aspects of the invention.
  • the review system 100 can be integrated into a computing device 200, such as by being embodied in a software application installed on the computing device.
  • Fig. 2 illustrates an architecture of an exemplary computing device into which the review system 100 can be integrated.
  • Those skilled in the art will recognize that the general architecture described with reference to Fig. 2 is for example only, and can be modified to accommodate various embodiments of the review system 100 and particular operational environments.
  • the review system 100 can be built on a custom, general-purpose, "query-based,” touch processing system.
  • An implementation of the review system 100 can be based on the recognition that touch input relevant to an operation might not be directed at the object of that operation. For example, holding a finger on a document 50 might mean the user wishes to drag the document 50, or it might mean the user wishes to keep the region under the user's finger from moving. More generally, with arbitrary numbers of fingers on the touchscreen input device 110, the review system 100 should be able to determine which gesture is indicated by the current number and arrangement of fingers.
  • the review system 100 can utilize a querying system, where touch-enabled objects can supply lambda functions describing (1) requirements for each individual touch point required for the operation; (2) mutual requirements among the touch points required for the operation; (3) negative requirements that must be satisfied by any touch points not involved in the operation; and possible additional functions, depending on the flexibility of the touch processing that is desired or required.
  • the querying system can consider all touch points on the display, and can look for a way of partitioning the touch points so as to satisfy all three or more sets of requirements.
  • a computing device 200 embodying the review system 100 can comprise a central processing unit 205 and one or more system memories 207, such as a random access memory 209 ("RAM") and a non-volatile memory, such as a read-only memory (“ROM”) 211.
  • the computing device 200 can further comprise a system bus 212 coupling together the memory 207, the processing unit 205, and various other components.
  • a basic input/output system containing routines to assist in transferring information between components of the computing device 200 can be stored in the ROM 211.
  • the computing device 200 can include a mass storage device 214 for storing an operating system 216, application programs, and other program modules.
  • the mass storage device 214 can be connected to the processing unit 205 through a mass storage controller (not shown) connected to the bus 212.
  • the mass storage device 214 and other computer-readable media can comprise computer storage media, which can include volatile and non- volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data.
  • Computer storage media includes, but is not limited to, RAM, ROM, EPROM, EEPROM, flash memory, other solid state memory technology, CD-ROM, digital versatile disks (“DVD”), other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage, other magnetic storage devices, or various other media used to store data accessible by the computing device 200.
  • a number of program modules and data files can be stored in the computer storage media and RAM 209 of the computing device 200.
  • Such program modules and data files can include an operating system 216 suitable for controlling operations of a networked personal computer.
  • a web browser application program, or web client 224 can also be stored on the computer storage media and RAM 209.
  • the web client 224 may comprise an application program for requesting and rendering web pages 226 created in Hypertext Markup Language ("HTML") or other types of markup languages.
  • HTML Hypertext Markup Language
  • the web client 224 can be capable of executing scripts through the use of a scripting host.
  • the scripting host executes program code expressed as scripts within the browser environment.
  • Computer-readable instructions on the storage media of the computing device 200 can include, for example, instructions for implementing processes of the review system 100 or for implementing a web client 224 for receiving instructions from the review system 100 when operated remotely. These instructions can be executed by the computer processor 205 to enable use of the review system 100.
  • the computing device 200 can operate in a networked environment using logical connections to remote computers over a network 250, such as the Internet.
  • the computing device 200 can connect to the network 250 and remote computers through a network interface unit 220 connected to the bus 212.
  • the computing device 200 can also include an input/output controller 222 for receiving and processing input from a number of input devices, including a keyboard, mouse, or electronic stylus. Interactions between the input devices and the review system 100 can be detected by the input/output controller 222 to provide meaningful input to the computing device 200.
  • the input/output controller 222 can additionally provide output to a display screen, a printer, or other type of input/output device, such as the multi-touch input device 110 or other appropriate input device of the review system 100.
  • the hardware and virtual components described above can work in combination to provide various aspects and operations of the review system 100, as will be described in detail below.
  • the review system 100 can provide various mechanisms by which the user can navigate the document 50 and modify the layout of the document 50 for the user's convenience during active reading.
  • dog-earing or bookmarking can be supported in a manner that is more convenient than in conventional computer-based systems.
  • bookmarking is supported by navigating to a desired page, selecting a bookmark icon or menu item, and then typing a name for the bookmark. Later, when the user wishes to return to a bookmarked location, the user can select the bookmark that was created. And when the bookmark is no longer needed, the user must explicitly delete the bookmark to remove it from the document.
  • This bookmarking process is inconvenient and time-consuming in situations where a user intends to create only a temporary bookmark, to facilitate flipping between sections for comparison. When a user desires simply to compare two or more sections of a document, the user must bookmark each section and cycle through the bookmark links to flip between the bookmarked sections.
  • Fig. 3 illustrates the use of transient bookmarks 300 in the review system 100, according to an exemplary embodiment of the present invention, which are an improvement over bookmarking in conventional computer-based systems.
  • the review system 100 can provide a much more convenient means of bookmarking, analogous to dog-earing and simply holding one's place in a book with a finger.
  • the user can perform a gesture to create a transient bookmark 300, which can be recalled by a later gesture.
  • a gesture to create a transient bookmark 300, which can be recalled by a later gesture.
  • the user can simply touch and hold a finger to the document 50 as the user navigates through the document 50.
  • a touch and hold can be interpreted as transient bookmarking only when occurring in a predetermined area of the document view region 140, such as near the lower left corner. This need not be the case, however, and in some other embodiments, the touch and hold can occur anywhere on the document 50 to create a transient bookmark 300.
  • the touch and hold can indicate to the review system 100 that the user is holding the currently visible place in the document 50, as the user continues to scroll through or otherwise navigate the document 50 in the document view region 140. Additional fingers can touch and hold on the document 50, next to the first finger, to indicate other transient bookmarks 300 within the document 50, as navigation continues.
  • a finger touches and holds to create a transient bookmark 300 a graphical representation 310 or link of the bookmark 300, such as an orb, an arrow, or an icon of a bookmark 300, can be created where the user touches.
  • the user can simply remove the finger corresponding to the desired position of the document 50.
  • the document 50 in the document view region 140 can automatically scroll to display the portion of the document 50 that was visible when the finger originally touched down to create the virtual dog-ear or bookmark 300.
  • the transient bookmark can disappear and be automatically deleted after a predetermined time period, such as several seconds.
  • Replacing the finger on the document 50, or on the graphical representation 310 of the bookmark 300 can cause the review system 100 to restore the bookmark 300 until the finger is lifted again. Accordingly, by placing and alternately lifting two or more fingers, the user can mark and switch between positions in the document 50.
  • the user need not waste time naming or deleting bookmarks 300, but can thus create transient bookmarks 300 by simply touching and holding the document 50.
  • a transient bookmark 300 can save and restore a state of the virtual workspace 120 or of the document view region 140, as opposed to merely a position within the document 50.
  • a transient bookmark 300 can save the current layout of the document 50 or the current layout of the entire virtual workspace 120. For example, and not limitation, if a portion of the document 50 includes highlighting or is collapsed, as will be described further below, these aspects of the document layout can be restored when a transient bookmark 300 is recalled, such as by the user's lifting a finger.
  • a bookmark 300 can capture the placement of document objects 150 or the magnification and rotation of the document view region 140 and document objects 150.
  • transient bookmark 300 If the user seeks to destroy a transient bookmark 300 without returning to the saved state corresponding to the bookmark 300, the user can drag the graphical representation 310 of the bookmark 300 away from the document view region 140 and, in some embodiments, into a virtual trash can or recycling bin. And as mentioned above, simply lifting a finger can also destroy a bookmark 300. Thus, by using transient bookmarks 300 in the review system 100, the user can rapidly flip between and discard layout states by placing, lifting, and moving fingers.
  • Fig. 4 illustrates a flow diagram of an exemplary method 400 of providing a transient bookmark 300, according to an exemplary embodiment of the present invention.
  • the method 400 depicted in this Fig. 4 is provided for illustrative purpose and is not limiting, and other methods toward a similar end can also be implemented.
  • the review system 100 can receive a transient bookmark 300 gesture, such as a touch and hold in the document view region 140.
  • the review system 100 can save the current state of the virtual workspace 120.
  • the review system 100 can then receive one or more other commands resulting in a change in the state of the virtual workspace 120.
  • the user can continue to navigate the document 50, thus changing the portion of the document 50 displayed in the document view region 140.
  • the review system 100 can receive a recall gesture for the bookmark 300, such as the user's releasing from the document view region 140 a finger corresponding to the bookmark 300.
  • the review system 100 can save the current state of the virtual workspace 120 and return the virtual workspace 120 to the previous state to which the bookmark 300 corresponds.
  • the method 400 of Fig. 4 results in creation and use of a transient bookmark 300 in the review system 100.
  • collapsing Another tool provided by the review system 100 is collapsing, which is not efficiently provided in either paper or conventional computer-based systems.
  • the review system 100 seeks to treat a document 50 in a fluid manner, instead of as a rigid structure. Collapsing is a tool to that end, enabling a user to focus on important parts of the document 50 in the context of the document's original layout, without being distracted by less important portions of the document 50.
  • collapsing is a process of squishing, minimizing, or squeezing an intermediate portion of the document 50, so as to bring together two portions of the document 50 separated by that intermediate portion.
  • Figs. 5A-5B illustrate an example of collapsing a document 50, where Fig. 5A shows the document 50 in an uncollapsed state, and Fig. 5B shows the document 50 after being collapsed.
  • an intermediate section C of the document 50 can be collapsed to bring separate sections A and C closer together.
  • the distinct sections A and C of the document 50 were both simultaneously viewable in the document view region 140 even before collapsing, this need not be the case.
  • a first section A may be far removed from a second section C within the document 50, such that both sections would not be simultaneously viewable in the document 50 at a readable magnification, without collapsing the document 50.
  • the review system 100 can collapse the document 50 in response to a collapse gesture received from the user.
  • the collapse gesture can be a pinching gesture, whereby the user places two fingers, usually a thumb and forefinger, on the touchscreen input device 110, and then moves the fingers closer together while maintaining the touch, thus creating a pinching motion. Pinching to initiate collapsing is intuitive because it corresponding to simultaneously scrolling in two directions, where the top finger of the pinch scrolls downward, while the bottom finger scrolls upward. As a result of this opposite-direction scrolling, the document 50 is collapsed.
  • magnification of the document view region 140 can also be adjusted with a pinching motion.
  • the gestures indicating collapse and magnification can be distinguished based on orientation of the pinching. For example, and not limitation, magnification can be initiated by a horizontal pinching gesture, while collapsing can be initiated by a vertical pinching gesture.
  • a subtlety of the pinching gesture is that the user can control many aspects of the collapse process by the manner of pinching. For example, and not limitation, if the user moves his or her top finger toward the bottom finger, then the portion of the document 50 below the fingers can remain stationary while the part above the fingers can move and collapse downward. Analogously, if the user moves his or her bottom finger while leaving the top finger stationary, the reverse can occur. If the user moves both fingers toward each other, then both the above and below portions of the document 50 can move toward each other and collapse together in the process. Further, the distance by which the user moves his or her fingers can control how much of the document is collapsed. Therefore, the user can perform a complex command, with many degrees of freedom, by way of a one-hand movement.
  • one or more other gestures can also be interpreted as a collapse command.
  • a collapse gesture performed on the preview region 130 can be used to initiate collapsing.
  • the review system 100 can interpret such touching as a collapse gesture.
  • Yet another collapse gesture can comprise the user's touching and holding a first section A of the document 50 in the document view region 140 and then touching a second section C in the preview region 130, or the user can touch and hold the first section A in the preview region 130 and then touch the second section C in the document view region 140.
  • the review system 100 can automatically collapse the document 50 and, more specifically, can collapse the intermediate section B between the separate sections A and C that were touched by the user in the preview region 130 or the document view region 140.
  • Performing a version of the collapse gesture, on the preview region 130 can be particularly useful when the sections A and C that the user desires to bring closer together are separated by a large amount of space within the document 50. In that case, when a large intermediate section B of the document 50 needs to be collapsed, pinching can become time-consuming.
  • the preview region 130 can be used to initiate collapsing in an efficient manner.
  • Collapsing can provide a number of benefits to the user during active reading. As shown in Fig. 5B, collapsing can enable the user to simultaneously view two distinct sections of the document 50 while retaining the linearity of the document 50 and the context of the two sections A and C. For example, although a portion of the intermediate section B between the distinct sections A and C may not be readable after collapsing, some of the intermediate section B can remain readable, so as to enable the user to see the context of the two sections A and C brought closer together by the collapsing. Retaining the document's linearity can be beneficial to the user because it can enable the user to maintain awareness of where he or she is within the document 50 and, thus, to maintain awareness of the general flow and organization of the document 50. Additionally, because the collapsed portion is still visible to the user, although not necessarily readable, collapsing can provide the user with a visual cue as to the amount of text lying between the two distinct sections A and C of the document 50.
  • collapsing within a single document 50 need not be limited to bringing two sections closer together. Rather, collapsing can also be used to reduce the distraction caused by multiple unimportant sections. Further, multiple collapsed sections can be present within the document 50 simultaneously, so as to enable the user to modify the spatial arrangement of the document 50 and view only the sections of the document 50 that hold interest for the user, while collapsing less interesting sections, maintaining the linearity of the document 50, and enabling the user to view the context of the sections that remain readable. To facilitate efficient collapsing of multiple sections of the document 50, the review system 100 can support collapsing based on a filter.
  • the collapse gesture can cause all text of the document 50 that is not highlighted to collapse, so as to enable the user to focus on the highlighted portions of the document 50.
  • other characteristics of document text can also be used as a filter during collapsing.
  • the review system 100 can uncollapse a portion of collapsed text upon receiving an uncollapse gesture.
  • an uncollapsed gesture can comprise the user's brushing or swiping a hand or finger upward or downward across the collapsed portion.
  • An upward swipe can cause the review system 100 to uncollapse the document 50 upward, so as to maintain the bottom position of the collapsed portion upon uncollapsing.
  • a downward swipe can initiate a downward uncollapsing.
  • Another important aspect of active reading is text selection and emphasis.
  • the user may wish to emphasize, extract, or otherwise manipulate portions of the document 50. In order for such manipulation to occur, however, the user can sometimes be required first to select the portion of the document 50 to be manipulated.
  • the review system 100 can provide a means for selecting text in a document 50.
  • the review system 100 can select a block of text in the document 50, preferably displayed in the document view region 140, in response to receiving a selection gesture from the user.
  • the selection gesture can comprise the user's touching a thumb and forefinger to the touchscreen input device 110 over the document view region 140, where the forefinger is positioned just below the starting point of the intended selection area in the document 50. While maintaining the touch, the user can slide the thumb and forefinger to the end of the text to be selected, and then the user can remove the thumb and index finder to end the touch.
  • the selection gesture can comprise user's touching with the thumb and forefinger approximately where the intended selection is to begin. In response, the review system 100 can display a cursor at the indicated location.
  • the user can move the thumb and forefinger while maintaining the touch, so as to direct the cursor to the correct start point.
  • the user can remove his or her thumb, maintaining the touch of the forefinger, and the user can drag the forefinger to the desired end point of the selection. This latter selection gesture may thus provide the user a more precise way to establish the start point of the selection.
  • the review system 100 can interpret either or both of the above gestures as a command to select the text between the start and end points of the touch. To confirm that the indicated text was selected, the review system 100 can temporarily emphasize the selected portion, such as by coloring, highlighting underlining, or enlarging the selected portion in the document view region 140. Unlike some conventional touch-based systems, the review system 100 need not rely on dwell time to detect that a selection gesture is occurring, and the user need not hold his hand or fingers in a single position for an extended period of time in order for the selection gesture to be recognized by the review system 100.
  • An alternative selection gesture particularly useful for users with long fingernails, can comprise the user's touching a selection button or icon in the virtual workspace 120 and then dragging a finger across the text that the user intends to select.
  • the user can select multiple sections of text, thus enabling the user to perform an action on the multiple selections simultaneously.
  • the review system 100 can create multiple selections in response to a multiple- selection gesture.
  • the multiple-selection gesture can comprise, for example, selecting a first section of text as discussed above, and then touching and holding that selected section while creating a second selection elsewhere in the document 50. Alternatively, however, the user need not hold a selected section to begin selecting other sections of the document 50.
  • the review system 100 can simply detect that multiple selections are being made in sequence, and can thus retain all selections. In that case, a multiple-selection gesture can simple be a sequence of selection gestures. All currently selected portions of the document 50 can be emphasized to indicate to the user that selection was successful.
  • the user can highlight that selected portion to maintain an emphasized state of the selected text.
  • highlighting can occur simultaneously with text selection using a highlighting gesture.
  • the highlighting gesture can comprise the user's brushing a finger upward or downward across selected text.
  • the distance that the finger travels during the upward or downward brushing can determine the intensity of the highlighting color used.
  • the color can be highly saturated as a result of a long stroke upward, or the color can be light as a result of quick or short stroke.
  • the highlighting gesture can comprise the user's touching a highlight button 180 (see Fig.
  • the user can be required to hold the highlight button 180 while selecting the text, but other embodiments can support highlighting when the user simply touches the highlight button 180 before beginning the selection.
  • the review system 100 can provide the user with one or more colors with which to highlight text in the document 50. If multiple colors are available, then the user can select a desired color, and that selected color can be the active highlighting color used to highlight text when the user so indicates. Using the toolbar 160 or the highlight button 180, the user can selectively collapse the document 50 based on highlighted portions and, in some embodiments, based on highlighted portion in a particular color. For example, and not limitation, while holding the highlight button 180, the user can perform the collapse gesture. In response, the review system 100 can collapse the portions of the document 50 that are not highlighted in the active highlighting color.
  • the review system 100 can enable the user to control the extent of the collapsing, so as to display more or less of the text around the highlighted portions as the document 50 is collapsed. If the collapse gesture is performed by vertical pinching, for example, then pinching two fingers very close together can collapse all or virtually all text that is not highlighted in the current color. In contrast, pinching only slightly so that the fingers remain some distance apart can leave some readable text on either side of each highlighted during collapsing. Thus, highlighting and collapsing can be used together to manipulate the document layout to the user's needs.
  • FIGs. 6A-6B illustrate creation of an excerpt 600 in the review system 100, according to an exemplary embodiment of the present invention. More specifically, Fig. 6A illustrates a selected section of text within the document 50, and Fig. 6B illustrates the virtual workspace 120 after the selected section as been extracted into an excerpt 600.
  • the review system 100 can create an excerpt 600 in response to an excerpt gesture, which can comprise a selection gesture in combination with an extraction gesture.
  • an excerpt gesture can comprise a selection gesture in combination with an extraction gesture.
  • the user can touch and hold the document 50 with one finger or hand, and then touch and drag the selected text from the document view region 140 into a portion of the virtual workspace 120 outside of the document view region 140.
  • This can be an intuitive gesture, because performing the gesture simply requires the user, after initial selection, to simulate holding the document 50 in place with one hand, while dragging a portion of the document 50 away with the other hand.
  • an excerpt 600 can be encapsulated or embodied in an excerpt object 650, a type of document object 150 moveable throughout the virtual workspace 120.
  • the excerpt object 650 can include the text extracted from the document 50 during the excerpt's creation. In an exemplary embodiment, this text is not removed from the document 50 in the document view region 140, but is simply duplicated into the excerpt objects 650 for the user's convenience, while maintaining the linearity and content of the document 50 in the document view region 140.
  • the excerpt object 650 can comprise a link 155 back to the portion of the document 50 from which it was extracted. That link 155 can have a graphical representation, such as an arrow, visible on or near the excerpt object 650 in the virtual workspace 120.
  • the document view region 140 can automatically return to the portion of the document 50 referred to bet the excerpt object 650.
  • the document view region 140 no longer displays the section of the document 50 from which the excerpt 600 was extracted, that section of the document 50 can automatically become centered in the document view region 140 when the user selects the arrow or other representation of the link 155 contained by the excerpt object 650.
  • the user can retrieve the portion of the document 50 referred to by an excerpt object 650 by simply selecting the link 155 of the excerpt object 650.
  • the portion of the document 50 that was extracted to the excerpt object 650 can contain a link 55 to the excerpt object 650.
  • the link 55 in the document view region 140 can have a graphical representation, such as an arrow. This arrow can be positioned on or near the extracted portion of the document 50 in the document view region 140.
  • the excerpt object 650 referred to by the link 55 can be emphasized by the review system 100, to enable the user to locate the excerpt object 650.
  • Emphasis can take various forms. For example, and not limitation, the excerpt object 650 can automatically be placed in front of other document objects 150 that may appear in the virtual workspace 120 and that may block the user's view of the excerpt object 650.
  • the excerpt object 650 can flash, change colors, or be emphasized in various other manner to enable the user to locate the excerpt object 650 as a result of the user's selection of the link 55 within the document 50.
  • the review system 100 can establish a pair of bidirectional links enabling the user to maintain a connection between the excerpt 600 and the portion of the document 50 from the excerpt 600 was extracted.
  • a large shortcoming of paper is the constraint that paper places on textual annotations, such as comments and notes.
  • Annotations on paper must generally be fit to the space of a small margin, and are typically only able to refer to text appearing within a single page.
  • software products like Microsoft Word® and Adobe Acrobat® avoid some of these difficulties, these software products still largely follow paper's paradigm.
  • annotations created by these software products are thus limited to a single referent on a single page, and the user is provided little control over the size and scale of annotations.
  • the review system 100 can overcome these difficulties by providing a flexible visual-spatial arrangement.
  • Figs. 7A-7B illustrate creation of an annotation 700 in the review system 100, according to an exemplary embodiment of the present invention. More specifically, Fig 7 A illustrates selection of text in the document 50 to which an annotation 700 will refer, and Fig. 7B illustrates an annotation object 750 referring back to the text selected in Fig. 7A.
  • Creation of an annotation 700 in the review system 100 can begin with selection of text in the document 50, as displayed in the document view region 140, or with selection of text in a preexisting document object 150. After text is selected, the user can simply begin typing, or the user can select an annotation button and then begin typing. The review system 100 can then interpret the typed text as an annotation 700, which can be encapsulated in an annotation object 750, a type of document object 150. The typed input received from the user can be displayed in the annotation object 750.
  • the annotation object 750 need not refer to only a single portion of text, in the document 50 or in another document object 150.
  • an annotation object 750 referring to multiple portions can be created when the user selects two or more sections of text, using the multiple selection gesture, and then types the annotation text.
  • an annotation 700 can be created for multiple sections by touching and holding each intended section within the preview region 130, the document view region 140, document objects 150, or some combination of these, and then typing or selecting an annotation button.
  • the annotation object 750 can have many similarities to an excerpt object 650, given that both are types of document objects 150, which will described in more detail below.
  • the review system 100 can create a bidirectional link between each annotation object 750 and the portion or portions of text referred to by the annotation object 750.
  • the annotation object 750 can thus contain a link 155 back to the one or more text portions of the document 50 or other document objects 150 to which the annotation object 750 refers. That link 155 can have a graphical representation, such as an arrow, linking the annotation object 750 back to the portions of text to which the annotation 700 refers.
  • the annotation object 750 can have a separate link 155 for each portion of text to which the annotation object 750 refers, while in other embodiments, a single link 155 can be used to refer back to all of the related portions of text in the document 50 or elsewhere.
  • a single link 155 is used, and when the user selects the single link 155 of the annotation object 750, the document 50 can automatically collapse to simultaneously display any portions of the document 50 linked to the annotation 700, and any document objects 150 linked to the annotation object 750 can automatically move into view in front of other document objects 150 in the virtual workspace 120.
  • the user can touch and hold multiple links 155 of an annotation object 750 to prompt the review system 100 to collapse the document 50 and recall the linked document objects 150, as needed to display the multiple linked portions of text.
  • Document objects 150 can be located in the virtual workspace 120 and manipulable in a manner similar to physical objects in a physical workspace.
  • a document object 150 can be freely moved about the virtual workspace 120 and positioned in the workspace 120 wherever the user desires.
  • Document objects 150 can be placed over one another, so as to hide each other or to bring one document object 150 into view at the expense of the visibility of another document object 150.
  • the size and number of document objects 150 that can be placed on the virtual workspace 120 need not have a predetermined limit, so the user can create and manipulate as many document objects 150 as the user desires to fit into the virtual workspace 120.
  • the user can resize individual document objects 150 or the entire virtual workspace 120, such as by performing a pinching gesture or other resizing gesture on a document object 150 or on an empty portion of the virtual workspace 120. Resizing the virtual workspace 120 can automatically resize, i.e., change the magnification of, all of the document objects 150 in the virtual workspace 120 and the spacings between the document objects 150. In some cases, however, there may be so many document objects 150 on the workspace 120 that the user needs to make the workspace 120 so small, such that text encapsulated in the document object 150 can no longer be legible. In that case, the user may desire to selectively and temporarily magnify individual or groups of document objects 150 in the virtual workspace 120. Individual document objects 150 can be selectively enlarged or shrunken by rescaling each document object 150, such as by use of the pinching gesture, but this can be an inconvenient process.
  • the review system 100 can provide a magnifier 190 (see Fig. 1) or fisheye.
  • the magnifier 190 can initially be created in response to a magnification gesture.
  • the magnification gesture can comprise the user's touching and holding three fingertips to the virtual workspace 120 until the review system 100 creates and displays the magnifier 190 where the user's fingers touch.
  • the initial size and placement of the magnifier 190 can depend on the spatial arrangement of the user's fingers when performing the magnification gesture, such that the magnifier 190 can be created to extend between and touch the user's fingers.
  • the user can spread two or three fingers away from each other, while touching the magnifier 190, to enlarge the magnifier 190.
  • the user can pinch his two or three finger inward, while touching the magnifier 190, to shrink the magnifier 190.
  • the size of the magnifier 190 can correspond to the extent of the workspace affected by the magnifier 190, because the area of the virtual workspace 120 covered by the magnifier can be enlarged.
  • the user can rotate the magnifier 190 in a clockwise or counterclockwise direction by, respectively, rotating the fingers touching the magnifier 190 in a clockwise or counterclockwise direction.
  • the orientation of the magnifier 190 can determine the degree or amount of magnification provided.
  • the user can drag the magnifier 190 throughout the workspace 120 in a manner similar to that by which the user can drag the document objects 150. For example, to reposition the magnifier 190, the user can touch and hold the magnifier 190 with a finger, and can then drag the magnifier 190 by moving the touching finger to a new location. Each document object 150 positioned under the magnifier 190 can be magnified according to the size and orientation of the magnifier 190.
  • a first document object 150 can contain a link or links 155 to one or more portions of the document 50 or other document objects 150 associated with the first document object 150.
  • the link 155 can be part of a bidirectional link, where the other part of the bidirectional link is associated with the document 50 in the document view region 140, or with another document object 150, and refers back to the first document object 150. Selecting a link 155 of the first document object 150 can cause the document 50 in the document view region 140 to scroll, so as to position the related portion of the document 50 at the vertical center of the document view region 140.
  • link 155 connects to another document object 155, then when the link is selected, that other document object 150 can be automatically brought into view over other document objects 150. If multiple portions of text in the document 50 or other document objects 150 are referred to by a selected link 155, or if multiple links 155 of the first document object 150 are selected, or if multiple links 155 of multiple document objects 150 are selected, then the document 50 in the document view region 140 can collapse, scroll, or collapse and scroll as needed to simultaneously display all portions of the document 50 referred to by the links 155. Analogously, linked document objects 150 can also be brought into view as necessary to display the text referred to by the links 155. Further analogously, if selected links 155 additionally refer to portions of a second document 50 in a second document view region 140, that second document 50 and second document view region 140 can also be modified as needed to display the text referred to by the selected links 155.
  • a bidirectional link between two or more document 50 portions can be created in response to a linking gesture.
  • a linking gesture can include, for example, selecting the desired document 50 portions and then touching the desired portions simultaneously.
  • the review system 100 can create a bidirectional link between the selected portions of the document 50.
  • selection of the link at one of the linked document 50 portions can automatically cause the other linked portions to come into view.
  • document objects 150 can also be attachable to one another, to enable the user to rearrange the document objects 150 and the virtual workspace 120 as needed.
  • the user can touch and drag one document object 150 until it contacts another.
  • the two document objects 150 can then be attached to each other, until the user touches both of them and drags them away from each other.
  • moving a primary one of those attached document objects 150 can cause all of the attached document objects 150 to move together, maintaining their spatial relationships with one another.
  • the primary document object 150 can be, for example, the document object 150 positioned at the highest point in the virtual workspace 120, as compared to the other grouped document objects 150.
  • the user can group annotation 700 and excerpts 600 together into a group to assist the user in performing the organizational aspects of active reading. Further, even after grouping document objects 150 together, the user can continue to rearrange the virtual workspace 120 to best suit the user's needs.
  • document objects 150 within a group can have a parent-child hierarchy, where a primary document object 150, such as the highest positioned or the first to become a member of the group, can be a parent of a lower positioned or later-grouped document object 150.
  • a parent document object 150 can control the movement of its child or children, such that when the user moves the parent document object 150, the child document object 150 automatically moves, thus maintaining its spatial relationship to its parent document object 150.
  • a child document object 150 is moved, its parent need not follow.
  • the same parent-child principles can apply to manipulations of document objects 150 other than repositioning.
  • resizing, magnification, and deletion can also be inherited by a child document object 150 from a parent document object 150, such that the child document object 150 can be resized, magnified, or deleted automatically along with its parent document object 150.
  • manipulations performed to a child document object 150 need not be inherited by a parent document object 150.
  • moving any one document object 150 belonging to a group can generally cause other objects 150 in that group to move as well, so as to maintain constant spatial relationships.
  • the user may place a finger on the document object 150 of the group that the user desires to remove, and the user may then place another finger on a different document object 150 in the group.
  • the user may drag the finger on the document object 150 to be removed, while leaving the other finger stationary on the other document object 150. This is analogous to pulling the group apart, and this gesture can be interpreted by the review system as a command to separate the moved document object 150 from the remainder of the group.
  • the review system 100 can enable the user to save the current state of the virtual workspace 120.
  • the review system 100 can export the virtual workspace 120 by printing to paper, printing to Adobe PDF, or exported to an image.
  • the review system 100 can be associated with a proprietary document format. If the user saves the virtual workspace 120 in this format, then the user can return to the virtual workspace 120 to continue active reading in the same state in which the virtual workspace 120 was saved.
  • Embodiments of the review system can thus be used to facilitate active reading, by providing a fluid-like, non-rigid, reading environment customizable by a user. While the review system has been disclosed in exemplary forms, many modifications, additions, and deletions may be made without departing from the spirit and scope of the system, method, and their equivalents, as set forth in the following claims.

Abstract

Review systems and methods facilitate active reading by providing a flexible environment in which users can examine documents. A review system can comprise a virtual workspace, a document view region, a preview region, and optional document objects. The virtual workspace can simulate a desktop or other physical workspace. The document view region can display a portion of an open document at a magnification sufficient for reading. The preview region can display the document at a magnification sufficient to enable the generally layout of the entire document to be determined. The document objects can be created from the document through tools of the review system. These document objects can be moveable throughout the virtual workspace as needed to facilitate the user's active reading process.

Description

SYSTEMS AND METHODS TO FACILITATE ACTIVE READING
CROSS-REFERENCE TO RELATED APPLICATIONS This application claims a benefit, under 35 U.S.C. § 119(e), of U.S. Provisional Application Serial No. 61/247,279, filed 30 September 2009, the entire contents and substance of which are hereby incorporated by reference.
TECHNICAL FIELD
Various embodiments of the present invention relate to digital documents and, more particularly, to systems and methods to facilitate active reading of digital documents.
BACKGROUND
From magazines and novels to review of important document, reading forms a critical part of our lives, and many reading tasks involve a rich interaction with the text. This rich interaction, known as active reading, can be conducted to answer questions, perform analysis, or obtain information. Active reading can involve highlighting, annotating, outlining, note-taking, comparing, and searching. As a result, active reading generally demands more of a reading medium than simply an ability to advance pages.
Although paper supports bimanual interaction and freeform annotation within the boundaries of a single page, paper lacks the flexibility to provide, for example, content rearrangement, document overviews, and annotation outside of page boundaries. Additionally, although the tangibility of paper supports some rapid forms of navigation, such as dog-earing and bookmarking with a finger, paper provides little flexibility to create a customized navigational structure. Modern pen-based computerized tablets do a fine job of imitating paper, which benefits users by providing a familiar medium, but as a result, these pen-based tablets suffer from the same limitations as paper. Thus, neither paper nor modern computer systems adequately facilitate active reading.
SUMMARY
There is a need for a document review system to provide a fluid-like environment in which users can freely and flexibly manipulate, rearrange, and annotate documents without many of the restriction inherent in paper. It is to such systems and related methods that various embodiments of the present invention are directed. Briefly described, various embodiments of the present invention are review systems and methods for facilitating active reading of documents, by providing a fluid-like interface with advantages over physical paper and conventional word processing systems. According to embodiments of the present invention, a document review system can provide a novel approach to representing and interacting with documents. In contrast to the paper model, which offers a stable but rigid representation, the document review system can provide a highly flexible, malleable document representation. The document review system can provide high degree-of- freedom ways to navigate through and manipulate the document representation, control what document content is displayed and where, and create annotations and other structures related to the document. To this end, the document review system can include a multi-touch, gesture- based user interface.
Earlier work has shown that active reading involves four core processes: annotation, content extraction, navigation, and layout. Embodiments of the document review system can provide improvements, as compared to paper and conventional word processing, to each of these processes. Annotation can be generally defined as text embellishment, including highlighting and marginalia. The review system can provide efficient annotation by enabling convenient switching between annotation tools, by supporting idiosyncratic markings, and by providing a convenient means for retrieving annotations made previously. Content extraction generally includes copying or moving content from a document to a secondary location, such as when outlining or note-taking. In an exemplary embodiment, the review system can closely integrate extraction with the reading process, so that user can organize and view extracted content, as well as link extracted content back to the original document. Navigation generally entails moving throughout a document and between multiple documents, such as when searching for text, turning a page, or flipping between selected locations for comparison. The review system can support bookmarks and parallelism to facilitate these or other navigational tasks. Layout generally refers to the visual or spatial arrangement of the document and related objects. The review system can optimize layout according to the user's preferences by enabling distinct portions of the document to be viewed in parallel, while maintaining the document's linearity.
More specifically, in an exemplary embodiment, a document review system can comprise a virtual workspace, a document view region, a preview region, and optional document objects. The system can be embodied in one or more computer-readable media and can be executable by one or more computer processors on a computing device. The computing device can comprise a multi-touch interface by which a user can interact with the virtual workspace and the overall document review system.
The virtual workspace can be a working environment in which the user can review a document. The virtual workspace can be, for example, a graphical user interface displayed in a virtual window or frame viewable through the multi-touch interface. In an exemplary embodiment, the virtual workspace can be designed to look and feel like a physical desktop or other physical workspace to which the user may be accustomed. The virtual workspace can be a relatively unstructured environment, enabling users to place the document objects as desired throughout the virtual workspace.
The document view region can be contained fully or partially within the virtual workspace. When a user opens one or more document in the workspace, at least part of the documents can be displayed in the view region. In an exemplary embodiment, the view region can be configured to display a viewable portion of at least one document at a size that enables a user to easily read the text of the document. The size of the document in the view region can, however, be increased or decreased as the user desired. If the document is too long to be contained fully within the view region at a given magnification state of the document, then only a portion of the document can be viewable in the view region. The document can be displayed in a continuous layout, and in an exemplary embodiment, page breaks in the document can be hidden, so that the document appears to be seamless and unbounded by pagination.
Like the document view region, the preview region can be contained fully or partially within the virtual workspace. The preview region can display a larger portion of the document, at a smaller size, than the view region. In an exemplary embodiment, the magnification of the preview region can be such that the entire document can be displayed continuously in the preview region. Alternatively, however, the magnification can be such that the general layout of the document can be determined by the preview region, although the text of the document need not be readable within the region. The preview region can be linked to the document view region and can serve various navigational or other purposes. For example, and not limitation, when a user touches a point in the document within the preview region, the portion of the document displayed in the document view region can change, so as to center in the document view region the portion of the document touched in the preview region. Thus, the preview region can be utilized to select a portion of the document that is displayed in the document view region.
The document objects can be moveable objects positioned throughout the virtual workspace as desired by the user. In some embodiments, however, such movement can be restricted to areas outside of one or both of the document view region and the preview region, so as not to obstruct these regions. A document object can be created by the user to assist the user in actively reading of the document. For example, and not limitation, the user can create an excerpt of the document or an annotation, either of which can be encapsulated in a document object, which may be smaller and more easily manipulable than the document as a whole. Once created, the document object can be freely moved about the virtual workspace, so as to enable the user to arrange the virtual workspace in a manner that creates a customized active reading experience. The document object can be linked to the portion or portions of the document to which the document object relates. For example, the document object can include a visual link, such as an arrow, that the user can touch to cause the one or more documents in the document view region to shift position, thus bringing the related portions into view.
The document review system can thus enable users to manipulate documents in a way that improves upon paper and other document manipulation systems. Other objects, features, and advantages of the review system, will become more apparent upon reading the following specification in conjunction with the accompanying drawing figures.
BRIEF DESCRIPTION OF THE FIGURES
Fig. 1 illustrates a review system, according to an exemplary embodiment of the present invention.
Fig. 2 illustrates an architecture of a computing device for providing the review system, according to an exemplary embodiment of the present invention.
Fig. 3 illustrates a transient bookmark of the review system, according to an exemplary embodiment of the present invention.
Fig. 4 illustrates a flow diagram of a method of creating a transient bookmark, according to an exemplary embodiment of the present invention.
Figs. 5A-5B illustrate collapsing of a document, according to an exemplary embodiment of the present invention. Figs. 6A-6B illustrate an excerpt of the review system 100, according to an exemplary embodiment of the present invention.
Figs. 7A-7B illustrate an annotation of the review system 100, according to an exemplary embodiment of the present invention.
DETAILED DESCRIPTION
To facilitate an understanding of the principles and features of the invention, various illustrative embodiments are explained below. In particular, the invention is described in the context of being a review system enabling a user to interact with documents in a fluid-like environment, thus facilitating active reading. Embodiments of the invention, however, are not limited to this context. Rather, embodiments of the invention can provide a freeform, fluid-like environment for performing a variety of tasks.
The components described hereinafter as making up various elements of the invention are intended to be illustrative and not restrictive. Many suitable components that can perform the same or similar functions as components described herein are intended to be embraced within the scope of the invention. Such other components not described herein can include, but are not limited to, similar or analogous components developed after development of the invention.
Various embodiments of the present invention are review systems to facilitate active reading. Referring now to the figures, in which like reference numerals represent like parts throughout the views, various embodiment of the review system will be described in detail.
Fig. 1 illustrates a review system 100, or document review system, according to an exemplary embodiment of the present invention. In an exemplary embodiment, the review system 100 can comprise, for example, a touchscreen input device 110 of a computing device 200, a virtual workspace 120, a document view region 130, a preview region 140, an optional one or more document objects 150, and a toolbar 160.
The touchscreen input device 110 can be a multi-touch input device for interfacing with the virtual workspace 120 and other aspects of the review system 100. In an exemplary embodiment, the touchscreen input device 110 is a multi-touch device capable of receiving multiple simultaneous touches, thus enabling a user to interact with the review system 100 in a natural manner, using multiple hands and fingers simultaneously. A detection system 115 can be integrated with or in communication with the touchscreen input device 110, to detect user interactions with the touchscreen input device 110. These user interactions, or gestures, can be interpreted as commands to the review system 100. Instead of a touchscreen input device 110, the review system 100 can alternatively comprise some other multi-point, bimanual, spatial input device capable of receiving a wide away of gestures interpretable as commands.
The review system 100 can be designed to improve four major processes that occur in active reading: annotation, content extraction, navigation, and layout. Conventional paper-like approaches fall short in facilitating these processes because of their fixed structure and lack of flexibility. Utilizing a multi-touch input device 110 can provide parallel and bimanual input, which are important parts of paper-based reading, and which also enable a flexible environment. A mouse, as used in most computer-based reading systems, is an inefficient control device because it provides only a single indicator or selector. A keyboard, also used in most computer- based reading systems, lacks a natural spatial mapping. The flexible interactions made possible by embodiments of the review system 100 inherently offer more degrees of freedom than traditionally offered by computer-based reading systems. Controlling these interactions with a mouse or a keyboard would be highly inefficient, requiring numerous sequential inputs to create a single command. In contrast, the multi-touch input device 110 can support multi-point spatial input and is thus capable of efficiently receiving a wide array of gestures for interacting with the review system 100.
As used through thus disclosure, the terms "touch," "hold," and the like need not refer only to physical contact between the user and the touchscreen input device 110. Such terms can refer to various interactions simulating a physical contact, such as pointing from a distance or bringing a finger, hand, or implement in close proximity to the touchscreen input device 110, so as to indicate a virtual touching, holding, or the like. The definition of a "touch" can be implementation-dependent, wherein the type of touchscreen input device 110 used can determine how interactions are detected and thus how a "touch" or "hold" is defined. For example, and not limitation, the touchscreen input device 110 can utilize resistive, capacitive, or camera technologies. If, for example, camera technology is used, then a "touch" can be defined based on camera sensitivity, or on an instrument's being within a predetermined distance from the touchscreen input device 110. Additionally, "touch," "hold," and like terms need not refer only to interactions between the user's hands or fingers and the touchscreen input device 110, but can also refer to interactions with an instrument held by the user, such as a stylus, marker, or pen. A display system 118 can be in communication with the detection system 115, the touchscreen input device 110, or both. The display system 118 can react to user gestures by displaying and refreshing a graphical user interface presented to the user, preferably through the touchscreen input device 110, which can perform as both an input and an output device. This graphical user interface can include the virtual workspace 120, the review region 130, the document view region 140, and the document objects 150, all of which will be described in more detail below.
The virtual workspace 120 can be accessible and manipulable through the touchscreen input device 110. The virtual workspace 120 can simulate a physical desktop, in that the user can freely move document objects 150 throughout the workspace 120 without being bound by a fixed structure common in computer-based reading system. In some exemplary embodiments, the virtual workspace 120 can contain the preview region 130 and the document view region 140. In other embodiments, however, the virtual workspace 120 can comprise the useable space of the review system 100 outside of the preview region 130 and the document view region 140.
When an application embodying the review system 100 is first opened, the review system
100 can present the user with the virtual workspace 120 containing a document 50, or configured to display a yet-to-be-opened document 50. Throughout the active reading process, the user can control the document 50 and other document objects 150 in the virtual workspace 120 with a vocabulary of multi-touch gestures. Through these gestures, the user can navigate, annotate, and manipulate the virtual workspace 120, rarely having to explicitly select tools or otherwise shift attention away from the document 50 at hand. Some basic interactions can be performed in the virtual workspace 120 as one might expect based on conventional touch applications. For example, objects can be repositioned by dragging the objects about the virtual workspace 120. Rescaling can be performed by a pinching or stretching gesture with two fingers, preferably in a horizontal orientation. Other performable gestures and operations, however, are new to the review system 100, as will be described below in detail.
The user can open a document 50 in the virtual workspace 120, and the open document 50 can be displayed in the one or both of the preview region 130 and the document view region 140. Throughout this disclosure, various actions are described as being performed or performable on the "text" of the open document 50. It will be understood, however, that all or most of such actions can similarly be performed on imbedded objects in the document 50 that are not text, such as images or multimedia. Thus, the term "text" throughout this disclosure is used for illustrative purposes only and is not restrictive.
The preview region 130 can be configured to display the document 50 at a magnification that enables the user to view the general layout of the document 50. In an exemplary embodiment, the entire document 50 can be viewable in the preview region 130, so as to present the general layout of the entire document 50 to the user. In some other embodiments, however, the magnification of the preview region 130 can be adjustable, so that the user can select a magnification size that is best suited to the user's needs.
The document view region 140 can display at least a portion of the open document 50. In an exemplary embodiment, the document view region 140 can display the document 50 at a magnification enabling the user to easily read the text of the document 50. In a further exemplary embodiment, the magnification of the document 50 in the document view region 140 can be modified by the user to enable to user to select a text size best suiting the user's needs.
The text of the document 50 can, in either or both of the preview region 130 and the document view region 140, be presented to the user in a continuous format, with or without pagination. If pagination is provided, then this provision can be for the user's reference only and need not restrict operations of the review system 100 to page boundaries. Some embodiments of the review system 100 can enable the user to select whether pagination is shown, to further customize the user's active reading experience. In the document view region 140 and in the preview region 130, if the entire document 50 is not visible, then the user can scroll vertically in the respective region 140 or 130 to adjust the visible portion of the document 50. Scrolling can occur when the user performs a predetermined gesture, such as touching the representation of the document 50 and, while maintaining contact with the touchscreen input device 110, sliding the fingers upward or downward. Sliding downward can cause the document 50 to move downward, thus displaying a previously invisible portion above the previously displayed portion of the document 50. Analogously, sliding upward can cause the document 50 to move upward, thus displaying a previously invisible portion below the previously displayed portion of the document 50.
The review system 100 can also support "fast scrolling" in the preview region 130, the document view region 140, or both. Scrolling at normal speed can occur as described above, in which case the displayed portion of the document 50 can be adjusted up or down corresponding to the distance the user's finger slides while in contact with the touchscreen input device 110. For example, and not limitation, with normal scrolling, the document 50 can be moved by a distance equivalent to the distance moved by the user's finger while the user's finger is holding the touchscreen input device 110. While normal scrolling is thus an intuitive means to navigate a document, normal scrolling can be inefficient for long document, when the user seeks to navigate between portions of the document 50 separated by a great distance.
To provide a more efficient scrolling mechanism, the review system 100 can also support fast scrolling. Fast scrolling can take advantage of modern touch sensors by determining an amount of pressure applied by the user in touching and holding the document 50 within the review system 100. For example, if the user drags the document 50 with a light touch, the movement of the document 50 in response can simply follow the finger, resulting in normal- speed scrolling. In contrast, if a firmer touch is used, then the movement of the document 50 can correspond to the pressure of the user's touch. The document 50 can move in the same direction as the finger, but at a speed corresponding to the pressure applied by the user, where increased pressure corresponds to increased speed and distance, and where decreased pressure corresponds to decreased speed and distance. For example, if the user drags his or her finger over a distance of one inch, the document 50 can move by one, two, three, or six inches, depending on how hard the user presses the touchscreen input device 110.
The document objects 150 can be objects created by the user to facilitate the user's active reading process. For example, and not limitation, a particular document object 150 can be created by the user, with tools of the review system 100, to represent and include an excerpt or annotation of the document 50. The document object 150 can contain text, an image, or another annotation or portion of the document 50. The document object 150 can also comprise a link to the portion of the document 50 to which the document object 150 refers. For example, and not limitation, an excerpt can contain a link back to the portion of the document 50 from which the excerpt was extracted. The link 155 can have a visible representation, such as an arrow, which can point from the document object 150 to the document view region 140 to indicate that the linked portion of the document 50 can be displayed in the document view region 140. When the user selects the link 155, the document 50 in the document view region 140 can automatically scroll to display the portion of the document 50 referred to by the document object 150. In an exemplary embodiment, selecting the link can cause the referred-to portion to be centered within the document view region 140. Selection of the link 155 can occur when the user touches the visible representation of the link 155. Various types and uses of the document objects 150 will be described in more detail later in this disclosure.
If included in the review system 100, the toolbar 160 can provide some guidance to the user of how to operate the review system 100. Although, in an exemplary embodiment, some or all tools of the review system 100 can be accessible via gestures, without the user's having to use the toolbar 160, the toolbar 160 can act a reminder of the gestures and tools available to the user. For example, the toolbar 160 can display a plurality of representations of categories 165, tools, or functions. Each category 165 can be related to one or more tools that the user can use to manipulate and review the document 50. In an exemplary embodiment, the toolbar 160 can teach the user how to use the review system 100. For example, in some embodiments of the review system 100, when the user selects the toolbar 160, such as by touching the toolbar 160 or performing some other predetermined selection gesture, the toolbar 160 can expand slightly for a temporary period. When expanded, each button or category 165 of the toolbar 160 can be larger than when the toolbar 160 is not expanded, and each button or category 165 can display a visual cue as to how a gesture can be performed to initiate one or more tools related to the corresponding category 165. Alternatively, when a particular category 165 of the toolbar 160 is selected, a visual cue for only the gesture corresponding to that category 165 can be displayed. The visual cues of the toolbar 160 can be, for example, text, animations, or still images of one or more hands performing the associated gestures. Thus, the user can utilize the toolbar 160 to learn gestures recognized by the review system 100. In other embodiments, the toolbar 160 can behave similarly to a toolbar 160 in a conventional word-processing application, whereby the user can initiate functions form the toolbar 160. In yet other embodiments, the toolbar 160 can both teach the related gestures and additionally initiate one or more tools of the review system 100.
The review system 100 can be embodied in a computer-readable medium and executed by a computer processor to provide one, some, or all aspects of the invention. For example, the review system 100 can be integrated into a computing device 200, such as by being embodied in a software application installed on the computing device. Fig. 2 illustrates an architecture of an exemplary computing device into which the review system 100 can be integrated. Those skilled in the art will recognize that the general architecture described with reference to Fig. 2 is for example only, and can be modified to accommodate various embodiments of the review system 100 and particular operational environments.
Architecturally, the review system 100 can be built on a custom, general-purpose, "query-based," touch processing system. An implementation of the review system 100 can be based on the recognition that touch input relevant to an operation might not be directed at the object of that operation. For example, holding a finger on a document 50 might mean the user wishes to drag the document 50, or it might mean the user wishes to keep the region under the user's finger from moving. More generally, with arbitrary numbers of fingers on the touchscreen input device 110, the review system 100 should be able to determine which gesture is indicated by the current number and arrangement of fingers. For this, the review system 100 can utilize a querying system, where touch-enabled objects can supply lambda functions describing (1) requirements for each individual touch point required for the operation; (2) mutual requirements among the touch points required for the operation; (3) negative requirements that must be satisfied by any touch points not involved in the operation; and possible additional functions, depending on the flexibility of the touch processing that is desired or required. The querying system can consider all touch points on the display, and can look for a way of partitioning the touch points so as to satisfy all three or more sets of requirements.
As shown in Fig. 2, a computing device 200 embodying the review system 100 can comprise a central processing unit 205 and one or more system memories 207, such as a random access memory 209 ("RAM") and a non-volatile memory, such as a read-only memory ("ROM") 211. The computing device 200 can further comprise a system bus 212 coupling together the memory 207, the processing unit 205, and various other components. A basic input/output system containing routines to assist in transferring information between components of the computing device 200 can be stored in the ROM 211. Additionally, the computing device 200 can include a mass storage device 214 for storing an operating system 216, application programs, and other program modules.
The mass storage device 214 can be connected to the processing unit 205 through a mass storage controller (not shown) connected to the bus 212. The mass storage device 214 and other computer-readable media can comprise computer storage media, which can include volatile and non- volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data. Computer storage media includes, but is not limited to, RAM, ROM, EPROM, EEPROM, flash memory, other solid state memory technology, CD-ROM, digital versatile disks ("DVD"), other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage, other magnetic storage devices, or various other media used to store data accessible by the computing device 200.
A number of program modules and data files can be stored in the computer storage media and RAM 209 of the computing device 200. Such program modules and data files can include an operating system 216 suitable for controlling operations of a networked personal computer. A web browser application program, or web client 224, can also be stored on the computer storage media and RAM 209. The web client 224 may comprise an application program for requesting and rendering web pages 226 created in Hypertext Markup Language ("HTML") or other types of markup languages. The web client 224 can be capable of executing scripts through the use of a scripting host. The scripting host executes program code expressed as scripts within the browser environment.
Computer-readable instructions on the storage media of the computing device 200 can include, for example, instructions for implementing processes of the review system 100 or for implementing a web client 224 for receiving instructions from the review system 100 when operated remotely. These instructions can be executed by the computer processor 205 to enable use of the review system 100.
The computing device 200 can operate in a networked environment using logical connections to remote computers over a network 250, such as the Internet. The computing device 200 can connect to the network 250 and remote computers through a network interface unit 220 connected to the bus 212.
The computing device 200 can also include an input/output controller 222 for receiving and processing input from a number of input devices, including a keyboard, mouse, or electronic stylus. Interactions between the input devices and the review system 100 can be detected by the input/output controller 222 to provide meaningful input to the computing device 200. The input/output controller 222 can additionally provide output to a display screen, a printer, or other type of input/output device, such as the multi-touch input device 110 or other appropriate input device of the review system 100. The hardware and virtual components described above can work in combination to provide various aspects and operations of the review system 100, as will be described in detail below.
The review system 100 can provide various mechanisms by which the user can navigate the document 50 and modify the layout of the document 50 for the user's convenience during active reading. For example, dog-earing or bookmarking can be supported in a manner that is more convenient than in conventional computer-based systems. In conventional systems, bookmarking is supported by navigating to a desired page, selecting a bookmark icon or menu item, and then typing a name for the bookmark. Later, when the user wishes to return to a bookmarked location, the user can select the bookmark that was created. And when the bookmark is no longer needed, the user must explicitly delete the bookmark to remove it from the document. This bookmarking process is inconvenient and time-consuming in situations where a user intends to create only a temporary bookmark, to facilitate flipping between sections for comparison. When a user desires simply to compare two or more sections of a document, the user must bookmark each section and cycle through the bookmark links to flip between the bookmarked sections.
Fig. 3 illustrates the use of transient bookmarks 300 in the review system 100, according to an exemplary embodiment of the present invention, which are an improvement over bookmarking in conventional computer-based systems. Through transient bookmarking, the review system 100 can provide a much more convenient means of bookmarking, analogous to dog-earing and simply holding one's place in a book with a finger.
In the document view region 140 of the review system 100, the user can perform a gesture to create a transient bookmark 300, which can be recalled by a later gesture. For example, the user can simply touch and hold a finger to the document 50 as the user navigates through the document 50. In some exemplary embodiments, a touch and hold can be interpreted as transient bookmarking only when occurring in a predetermined area of the document view region 140, such as near the lower left corner. This need not be the case, however, and in some other embodiments, the touch and hold can occur anywhere on the document 50 to create a transient bookmark 300.
The touch and hold can indicate to the review system 100 that the user is holding the currently visible place in the document 50, as the user continues to scroll through or otherwise navigate the document 50 in the document view region 140. Additional fingers can touch and hold on the document 50, next to the first finger, to indicate other transient bookmarks 300 within the document 50, as navigation continues. When a finger touches and holds to create a transient bookmark 300, a graphical representation 310 or link of the bookmark 300, such as an orb, an arrow, or an icon of a bookmark 300, can be created where the user touches.
When the user desires to return to a marked document position, the user can simply remove the finger corresponding to the desired position of the document 50. In response to the lifted finger, the document 50 in the document view region 140 can automatically scroll to display the portion of the document 50 that was visible when the finger originally touched down to create the virtual dog-ear or bookmark 300. If the user leaves his or her finger up after recalling a bookmark, the transient bookmark can disappear and be automatically deleted after a predetermined time period, such as several seconds. Replacing the finger on the document 50, or on the graphical representation 310 of the bookmark 300, can cause the review system 100 to restore the bookmark 300 until the finger is lifted again. Accordingly, by placing and alternately lifting two or more fingers, the user can mark and switch between positions in the document 50. The user need not waste time naming or deleting bookmarks 300, but can thus create transient bookmarks 300 by simply touching and holding the document 50.
Further, as a benefit over both paper and conventional computer-based systems, a transient bookmark 300 can save and restore a state of the virtual workspace 120 or of the document view region 140, as opposed to merely a position within the document 50. In some embodiments, a transient bookmark 300 can save the current layout of the document 50 or the current layout of the entire virtual workspace 120. For example, and not limitation, if a portion of the document 50 includes highlighting or is collapsed, as will be described further below, these aspects of the document layout can be restored when a transient bookmark 300 is recalled, such as by the user's lifting a finger. For another example, a bookmark 300 can capture the placement of document objects 150 or the magnification and rotation of the document view region 140 and document objects 150.
If the user seeks to destroy a transient bookmark 300 without returning to the saved state corresponding to the bookmark 300, the user can drag the graphical representation 310 of the bookmark 300 away from the document view region 140 and, in some embodiments, into a virtual trash can or recycling bin. And as mentioned above, simply lifting a finger can also destroy a bookmark 300. Thus, by using transient bookmarks 300 in the review system 100, the user can rapidly flip between and discard layout states by placing, lifting, and moving fingers.
Fig. 4 illustrates a flow diagram of an exemplary method 400 of providing a transient bookmark 300, according to an exemplary embodiment of the present invention. The method 400 depicted in this Fig. 4 is provided for illustrative purpose and is not limiting, and other methods toward a similar end can also be implemented. As shown in the illustrated method 400, at 410 the review system 100 can receive a transient bookmark 300 gesture, such as a touch and hold in the document view region 140. At 420, in response to the bookmarking gesture, the review system 100 can save the current state of the virtual workspace 120. At 430, the review system 100 can then receive one or more other commands resulting in a change in the state of the virtual workspace 120. For example, and not limitation, the user can continue to navigate the document 50, thus changing the portion of the document 50 displayed in the document view region 140. At 440, the review system 100 can receive a recall gesture for the bookmark 300, such as the user's releasing from the document view region 140 a finger corresponding to the bookmark 300. In response to this recall gesture, at 450, the review system 100 can save the current state of the virtual workspace 120 and return the virtual workspace 120 to the previous state to which the bookmark 300 corresponds. Thus, the method 400 of Fig. 4 results in creation and use of a transient bookmark 300 in the review system 100.
Another tool provided by the review system 100 is collapsing, which is not efficiently provided in either paper or conventional computer-based systems. The review system 100 seeks to treat a document 50 in a fluid manner, instead of as a rigid structure. Collapsing is a tool to that end, enabling a user to focus on important parts of the document 50 in the context of the document's original layout, without being distracted by less important portions of the document 50. In essence, collapsing is a process of squishing, minimizing, or squeezing an intermediate portion of the document 50, so as to bring together two portions of the document 50 separated by that intermediate portion.
Figs. 5A-5B illustrate an example of collapsing a document 50, where Fig. 5A shows the document 50 in an uncollapsed state, and Fig. 5B shows the document 50 after being collapsed. As shown by comparing Figs. 5A and 5B, an intermediate section C of the document 50 can be collapsed to bring separate sections A and C closer together. Although in Fig, 5A, the distinct sections A and C of the document 50 were both simultaneously viewable in the document view region 140 even before collapsing, this need not be the case. In some instances, a first section A may be far removed from a second section C within the document 50, such that both sections would not be simultaneously viewable in the document 50 at a readable magnification, without collapsing the document 50.
The review system 100 can collapse the document 50 in response to a collapse gesture received from the user. In some embodiments, the collapse gesture can be a pinching gesture, whereby the user places two fingers, usually a thumb and forefinger, on the touchscreen input device 110, and then moves the fingers closer together while maintaining the touch, thus creating a pinching motion. Pinching to initiate collapsing is intuitive because it corresponding to simultaneously scrolling in two directions, where the top finger of the pinch scrolls downward, while the bottom finger scrolls upward. As a result of this opposite-direction scrolling, the document 50 is collapsed.
As mentioned above, magnification of the document view region 140 can also be adjusted with a pinching motion. The gestures indicating collapse and magnification can be distinguished based on orientation of the pinching. For example, and not limitation, magnification can be initiated by a horizontal pinching gesture, while collapsing can be initiated by a vertical pinching gesture.
A subtlety of the pinching gesture, in those embodiments where it is used, is that the user can control many aspects of the collapse process by the manner of pinching. For example, and not limitation, if the user moves his or her top finger toward the bottom finger, then the portion of the document 50 below the fingers can remain stationary while the part above the fingers can move and collapse downward. Analogously, if the user moves his or her bottom finger while leaving the top finger stationary, the reverse can occur. If the user moves both fingers toward each other, then both the above and below portions of the document 50 can move toward each other and collapse together in the process. Further, the distance by which the user moves his or her fingers can control how much of the document is collapsed. Therefore, the user can perform a complex command, with many degrees of freedom, by way of a one-hand movement.
In addition, or alternatively, to the vertical pinching gesture, one or more other gestures can also be interpreted as a collapse command. For example, a collapse gesture performed on the preview region 130 can be used to initiate collapsing. When the user touches and holds on a first section A of the document 50 within the preview region 130, and while holding the first section A, the user also touches a separate second section C also in the preview region 130, the review system 100 can interpret such touching as a collapse gesture. Yet another collapse gesture can comprise the user's touching and holding a first section A of the document 50 in the document view region 140 and then touching a second section C in the preview region 130, or the user can touch and hold the first section A in the preview region 130 and then touch the second section C in the document view region 140.
In response to one or all of the above collapse gestures, the review system 100 can automatically collapse the document 50 and, more specifically, can collapse the intermediate section B between the separate sections A and C that were touched by the user in the preview region 130 or the document view region 140. Performing a version of the collapse gesture, on the preview region 130, can be particularly useful when the sections A and C that the user desires to bring closer together are separated by a large amount of space within the document 50. In that case, when a large intermediate section B of the document 50 needs to be collapsed, pinching can become time-consuming. Thus, the preview region 130 can be used to initiate collapsing in an efficient manner.
Collapsing can provide a number of benefits to the user during active reading. As shown in Fig. 5B, collapsing can enable the user to simultaneously view two distinct sections of the document 50 while retaining the linearity of the document 50 and the context of the two sections A and C. For example, although a portion of the intermediate section B between the distinct sections A and C may not be readable after collapsing, some of the intermediate section B can remain readable, so as to enable the user to see the context of the two sections A and C brought closer together by the collapsing. Retaining the document's linearity can be beneficial to the user because it can enable the user to maintain awareness of where he or she is within the document 50 and, thus, to maintain awareness of the general flow and organization of the document 50. Additionally, because the collapsed portion is still visible to the user, although not necessarily readable, collapsing can provide the user with a visual cue as to the amount of text lying between the two distinct sections A and C of the document 50.
It will be understood that collapsing within a single document 50 need not be limited to bringing two sections closer together. Rather, collapsing can also be used to reduce the distraction caused by multiple unimportant sections. Further, multiple collapsed sections can be present within the document 50 simultaneously, so as to enable the user to modify the spatial arrangement of the document 50 and view only the sections of the document 50 that hold interest for the user, while collapsing less interesting sections, maintaining the linearity of the document 50, and enabling the user to view the context of the sections that remain readable. To facilitate efficient collapsing of multiple sections of the document 50, the review system 100 can support collapsing based on a filter. For example, and not limitation, if a highlighting filter is set, then the collapse gesture can cause all text of the document 50 that is not highlighted to collapse, so as to enable the user to focus on the highlighted portions of the document 50. In addition or alternatively to highlighting, other characteristics of document text can also be used as a filter during collapsing.
The review system 100 can uncollapse a portion of collapsed text upon receiving an uncollapse gesture. In an exemplary embodiment, for example, an uncollapsed gesture can comprise the user's brushing or swiping a hand or finger upward or downward across the collapsed portion. An upward swipe can cause the review system 100 to uncollapse the document 50 upward, so as to maintain the bottom position of the collapsed portion upon uncollapsing. Analogously, a downward swipe can initiate a downward uncollapsing.
Another important aspect of active reading is text selection and emphasis. The user may wish to emphasize, extract, or otherwise manipulate portions of the document 50. In order for such manipulation to occur, however, the user can sometimes be required first to select the portion of the document 50 to be manipulated. Thus, the review system 100 can provide a means for selecting text in a document 50.
The review system 100 can select a block of text in the document 50, preferably displayed in the document view region 140, in response to receiving a selection gesture from the user. In an exemplary embodiment, the selection gesture can comprise the user's touching a thumb and forefinger to the touchscreen input device 110 over the document view region 140, where the forefinger is positioned just below the starting point of the intended selection area in the document 50. While maintaining the touch, the user can slide the thumb and forefinger to the end of the text to be selected, and then the user can remove the thumb and index finder to end the touch. Alternatively, the selection gesture can comprise user's touching with the thumb and forefinger approximately where the intended selection is to begin. In response, the review system 100 can display a cursor at the indicated location. If the user is unsatisfied with the cursor position, the user can move the thumb and forefinger while maintaining the touch, so as to direct the cursor to the correct start point. After positioning the cursor on the desired start point, the user can remove his or her thumb, maintaining the touch of the forefinger, and the user can drag the forefinger to the desired end point of the selection. This latter selection gesture may thus provide the user a more precise way to establish the start point of the selection.
The review system 100 can interpret either or both of the above gestures as a command to select the text between the start and end points of the touch. To confirm that the indicated text was selected, the review system 100 can temporarily emphasize the selected portion, such as by coloring, highlighting underlining, or enlarging the selected portion in the document view region 140. Unlike some conventional touch-based systems, the review system 100 need not rely on dwell time to detect that a selection gesture is occurring, and the user need not hold his hand or fingers in a single position for an extended period of time in order for the selection gesture to be recognized by the review system 100. An alternative selection gesture, particularly useful for users with long fingernails, can comprise the user's touching a selection button or icon in the virtual workspace 120 and then dragging a finger across the text that the user intends to select.
In some embodiments of the review system 100, the user can select multiple sections of text, thus enabling the user to perform an action on the multiple selections simultaneously. The review system 100 can create multiple selections in response to a multiple- selection gesture. The multiple-selection gesture can comprise, for example, selecting a first section of text as discussed above, and then touching and holding that selected section while creating a second selection elsewhere in the document 50. Alternatively, however, the user need not hold a selected section to begin selecting other sections of the document 50. In some embodiments, for example, the review system 100 can simply detect that multiple selections are being made in sequence, and can thus retain all selections. In that case, a multiple-selection gesture can simple be a sequence of selection gestures. All currently selected portions of the document 50 can be emphasized to indicate to the user that selection was successful.
After a portion of a document 50 is selected, the user can highlight that selected portion to maintain an emphasized state of the selected text. In some embodiments, highlighting can occur simultaneously with text selection using a highlighting gesture. For example, and not limitation, the highlighting gesture can comprise the user's brushing a finger upward or downward across selected text. In some embodiments, the distance that the finger travels during the upward or downward brushing can determine the intensity of the highlighting color used. The color can be highly saturated as a result of a long stroke upward, or the color can be light as a result of quick or short stroke. Alternatively, the highlighting gesture can comprise the user's touching a highlight button 180 (see Fig. 1) in the virtual workspace 120 or on the toolbar 160 while dragging another finger across the text to be selected, or while performing the selection gesture. In some embodiments, the user can be required to hold the highlight button 180 while selecting the text, but other embodiments can support highlighting when the user simply touches the highlight button 180 before beginning the selection.
The review system 100 can provide the user with one or more colors with which to highlight text in the document 50. If multiple colors are available, then the user can select a desired color, and that selected color can be the active highlighting color used to highlight text when the user so indicates. Using the toolbar 160 or the highlight button 180, the user can selectively collapse the document 50 based on highlighted portions and, in some embodiments, based on highlighted portion in a particular color. For example, and not limitation, while holding the highlight button 180, the user can perform the collapse gesture. In response, the review system 100 can collapse the portions of the document 50 that are not highlighted in the active highlighting color. In some further embodiments, the review system 100 can enable the user to control the extent of the collapsing, so as to display more or less of the text around the highlighted portions as the document 50 is collapsed. If the collapse gesture is performed by vertical pinching, for example, then pinching two fingers very close together can collapse all or virtually all text that is not highlighted in the current color. In contrast, pinching only slightly so that the fingers remain some distance apart can leave some readable text on either side of each highlighted during collapsing. Thus, highlighting and collapsing can be used together to manipulate the document layout to the user's needs.
In addition to highlighting, various other tasks can be performed on a block of selected text. For example, Figs. 6A-6B illustrate creation of an excerpt 600 in the review system 100, according to an exemplary embodiment of the present invention. More specifically, Fig. 6A illustrates a selected section of text within the document 50, and Fig. 6B illustrates the virtual workspace 120 after the selected section as been extracted into an excerpt 600.
The review system 100 can create an excerpt 600 in response to an excerpt gesture, which can comprise a selection gesture in combination with an extraction gesture. To perform the extraction portion of the gesture, the user can touch and hold the document 50 with one finger or hand, and then touch and drag the selected text from the document view region 140 into a portion of the virtual workspace 120 outside of the document view region 140. This can be an intuitive gesture, because performing the gesture simply requires the user, after initial selection, to simulate holding the document 50 in place with one hand, while dragging a portion of the document 50 away with the other hand.
Once created, an excerpt 600 can be encapsulated or embodied in an excerpt object 650, a type of document object 150 moveable throughout the virtual workspace 120. The excerpt object 650 can include the text extracted from the document 50 during the excerpt's creation. In an exemplary embodiment, this text is not removed from the document 50 in the document view region 140, but is simply duplicated into the excerpt objects 650 for the user's convenience, while maintaining the linearity and content of the document 50 in the document view region 140.
The excerpt object 650 can comprise a link 155 back to the portion of the document 50 from which it was extracted. That link 155 can have a graphical representation, such as an arrow, visible on or near the excerpt object 650 in the virtual workspace 120. When the user selects the link 155, such as by touching the graphical representation, the document view region 140 can automatically return to the portion of the document 50 referred to bet the excerpt object 650. In other words, if the document view region 140 no longer displays the section of the document 50 from which the excerpt 600 was extracted, that section of the document 50 can automatically become centered in the document view region 140 when the user selects the arrow or other representation of the link 155 contained by the excerpt object 650. Thus, the user can retrieve the portion of the document 50 referred to by an excerpt object 650 by simply selecting the link 155 of the excerpt object 650.
In the document view region 140, the portion of the document 50 that was extracted to the excerpt object 650 can contain a link 55 to the excerpt object 650. Like the link 155 comprised in the excerpt object 650, the link 55 in the document view region 140 can have a graphical representation, such as an arrow. This arrow can be positioned on or near the extracted portion of the document 50 in the document view region 140. When the link 55 is selected, the excerpt object 650 referred to by the link 55 can be emphasized by the review system 100, to enable the user to locate the excerpt object 650. Emphasis can take various forms. For example, and not limitation, the excerpt object 650 can automatically be placed in front of other document objects 150 that may appear in the virtual workspace 120 and that may block the user's view of the excerpt object 650. Alternatively, for example, the excerpt object 650 can flash, change colors, or be emphasized in various other manner to enable the user to locate the excerpt object 650 as a result of the user's selection of the link 55 within the document 50. Thus, when an excerpt is created, the review system 100 can establish a pair of bidirectional links enabling the user to maintain a connection between the excerpt 600 and the portion of the document 50 from the excerpt 600 was extracted.
A large shortcoming of paper is the constraint that paper places on textual annotations, such as comments and notes. Annotations on paper must generally be fit to the space of a small margin, and are typically only able to refer to text appearing within a single page. While software products like Microsoft Word® and Adobe Acrobat® avoid some of these difficulties, these software products still largely follow paper's paradigm. As a result, annotations created by these software products are thus limited to a single referent on a single page, and the user is provided little control over the size and scale of annotations. The review system 100 can overcome these difficulties by providing a flexible visual-spatial arrangement.
Figs. 7A-7B illustrate creation of an annotation 700 in the review system 100, according to an exemplary embodiment of the present invention. More specifically, Fig 7 A illustrates selection of text in the document 50 to which an annotation 700 will refer, and Fig. 7B illustrates an annotation object 750 referring back to the text selected in Fig. 7A.
Creation of an annotation 700 in the review system 100 can begin with selection of text in the document 50, as displayed in the document view region 140, or with selection of text in a preexisting document object 150. After text is selected, the user can simply begin typing, or the user can select an annotation button and then begin typing. The review system 100 can then interpret the typed text as an annotation 700, which can be encapsulated in an annotation object 750, a type of document object 150. The typed input received from the user can be displayed in the annotation object 750.
In some embodiments, the annotation object 750 need not refer to only a single portion of text, in the document 50 or in another document object 150. For example, an annotation object 750 referring to multiple portions can be created when the user selects two or more sections of text, using the multiple selection gesture, and then types the annotation text. For another example, an annotation 700 can be created for multiple sections by touching and holding each intended section within the preview region 130, the document view region 140, document objects 150, or some combination of these, and then typing or selecting an annotation button.
The annotation object 750 can have many similarities to an excerpt object 650, given that both are types of document objects 150, which will described in more detail below. For example, like an excerpt object 650, the review system 100 can create a bidirectional link between each annotation object 750 and the portion or portions of text referred to by the annotation object 750. The annotation object 750 can thus contain a link 155 back to the one or more text portions of the document 50 or other document objects 150 to which the annotation object 750 refers. That link 155 can have a graphical representation, such as an arrow, linking the annotation object 750 back to the portions of text to which the annotation 700 refers. In some embodiments, the annotation object 750 can have a separate link 155 for each portion of text to which the annotation object 750 refers, while in other embodiments, a single link 155 can be used to refer back to all of the related portions of text in the document 50 or elsewhere. When a single link 155 is used, and when the user selects the single link 155 of the annotation object 750, the document 50 can automatically collapse to simultaneously display any portions of the document 50 linked to the annotation 700, and any document objects 150 linked to the annotation object 750 can automatically move into view in front of other document objects 150 in the virtual workspace 120. Likewise, if multiple links 155 are used, the user can touch and hold multiple links 155 of an annotation object 750 to prompt the review system 100 to collapse the document 50 and recall the linked document objects 150, as needed to display the multiple linked portions of text.
Document objects 150, such as excerpt objects 650 and annotation objects 750, can be located in the virtual workspace 120 and manipulable in a manner similar to physical objects in a physical workspace. For example, and not limitation, a document object 150 can be freely moved about the virtual workspace 120 and positioned in the workspace 120 wherever the user desires. Document objects 150 can be placed over one another, so as to hide each other or to bring one document object 150 into view at the expense of the visibility of another document object 150. The size and number of document objects 150 that can be placed on the virtual workspace 120 need not have a predetermined limit, so the user can create and manipulate as many document objects 150 as the user desires to fit into the virtual workspace 120. In some embodiments of the review system 100, the user can resize individual document objects 150 or the entire virtual workspace 120, such as by performing a pinching gesture or other resizing gesture on a document object 150 or on an empty portion of the virtual workspace 120. Resizing the virtual workspace 120 can automatically resize, i.e., change the magnification of, all of the document objects 150 in the virtual workspace 120 and the spacings between the document objects 150. In some cases, however, there may be so many document objects 150 on the workspace 120 that the user needs to make the workspace 120 so small, such that text encapsulated in the document object 150 can no longer be legible. In that case, the user may desire to selectively and temporarily magnify individual or groups of document objects 150 in the virtual workspace 120. Individual document objects 150 can be selectively enlarged or shrunken by rescaling each document object 150, such as by use of the pinching gesture, but this can be an inconvenient process.
To assist the user in modifying the scale of individual document objects 150, the review system 100 can provide a magnifier 190 (see Fig. 1) or fisheye. The magnifier 190 can initially be created in response to a magnification gesture. For example, and not limitation, the magnification gesture can comprise the user's touching and holding three fingertips to the virtual workspace 120 until the review system 100 creates and displays the magnifier 190 where the user's fingers touch. The initial size and placement of the magnifier 190 can depend on the spatial arrangement of the user's fingers when performing the magnification gesture, such that the magnifier 190 can be created to extend between and touch the user's fingers. After the magnifier 190 is initially created, the user can spread two or three fingers away from each other, while touching the magnifier 190, to enlarge the magnifier 190. Analogously, the user can pinch his two or three finger inward, while touching the magnifier 190, to shrink the magnifier 190. The size of the magnifier 190 can correspond to the extent of the workspace affected by the magnifier 190, because the area of the virtual workspace 120 covered by the magnifier can be enlarged. Additionally, the user can rotate the magnifier 190 in a clockwise or counterclockwise direction by, respectively, rotating the fingers touching the magnifier 190 in a clockwise or counterclockwise direction. The orientation of the magnifier 190 can determine the degree or amount of magnification provided.
The user can drag the magnifier 190 throughout the workspace 120 in a manner similar to that by which the user can drag the document objects 150. For example, to reposition the magnifier 190, the user can touch and hold the magnifier 190 with a finger, and can then drag the magnifier 190 by moving the touching finger to a new location. Each document object 150 positioned under the magnifier 190 can be magnified according to the size and orientation of the magnifier 190.
As discussed above with respect to excerpt objects 650 and annotation objects 750, a first document object 150 can contain a link or links 155 to one or more portions of the document 50 or other document objects 150 associated with the first document object 150. The link 155 can be part of a bidirectional link, where the other part of the bidirectional link is associated with the document 50 in the document view region 140, or with another document object 150, and refers back to the first document object 150. Selecting a link 155 of the first document object 150 can cause the document 50 in the document view region 140 to scroll, so as to position the related portion of the document 50 at the vertical center of the document view region 140. Alternatively, if the link 155 connects to another document object 155, then when the link is selected, that other document object 150 can be automatically brought into view over other document objects 150. If multiple portions of text in the document 50 or other document objects 150 are referred to by a selected link 155, or if multiple links 155 of the first document object 150 are selected, or if multiple links 155 of multiple document objects 150 are selected, then the document 50 in the document view region 140 can collapse, scroll, or collapse and scroll as needed to simultaneously display all portions of the document 50 referred to by the links 155. Analogously, linked document objects 150 can also be brought into view as necessary to display the text referred to by the links 155. Further analogously, if selected links 155 additionally refer to portions of a second document 50 in a second document view region 140, that second document 50 and second document view region 140 can also be modified as needed to display the text referred to by the selected links 155.
In the same or similar manner by which a document object 150 can be linked to a portion of the document 50 or to another document object 150, two or more portions of a document 50 or in different documents 50 can be linked together. A bidirectional link between two or more document 50 portions can be created in response to a linking gesture. A linking gesture can include, for example, selecting the desired document 50 portions and then touching the desired portions simultaneously. In response to this linking gesture, the review system 100 can create a bidirectional link between the selected portions of the document 50. Like the links associated with document object 150, selection of the link at one of the linked document 50 portions can automatically cause the other linked portions to come into view.
In addition to being moveable throughout the workspace 120, document objects 150 can also be attachable to one another, to enable the user to rearrange the document objects 150 and the virtual workspace 120 as needed. To attach two or more document objects 150 together, the user can touch and drag one document object 150 until it contacts another. The two document objects 150 can then be attached to each other, until the user touches both of them and drags them away from each other. In some exemplary embodiments, when a group of document objects 150 are attached together, moving a primary one of those attached document objects 150 can cause all of the attached document objects 150 to move together, maintaining their spatial relationships with one another. The primary document object 150 can be, for example, the document object 150 positioned at the highest point in the virtual workspace 120, as compared to the other grouped document objects 150. Thus, the user can group annotation 700 and excerpts 600 together into a group to assist the user in performing the organizational aspects of active reading. Further, even after grouping document objects 150 together, the user can continue to rearrange the virtual workspace 120 to best suit the user's needs.
In some other exemplary embodiments, document objects 150 within a group can have a parent-child hierarchy, where a primary document object 150, such as the highest positioned or the first to become a member of the group, can be a parent of a lower positioned or later-grouped document object 150. A parent document object 150 can control the movement of its child or children, such that when the user moves the parent document object 150, the child document object 150 automatically moves, thus maintaining its spatial relationship to its parent document object 150. In contrast, when a child document object 150 is moved, its parent need not follow. The same parent-child principles can apply to manipulations of document objects 150 other than repositioning. For example, and not limitation, resizing, magnification, and deletion can also be inherited by a child document object 150 from a parent document object 150, such that the child document object 150 can be resized, magnified, or deleted automatically along with its parent document object 150. In contrast, manipulations performed to a child document object 150 need not be inherited by a parent document object 150.
In yet other exemplary embodiments, moving any one document object 150 belonging to a group can generally cause other objects 150 in that group to move as well, so as to maintain constant spatial relationships. In such an embodiment, to remove a document object 150 from the group, the user may place a finger on the document object 150 of the group that the user desires to remove, and the user may then place another finger on a different document object 150 in the group. The user may drag the finger on the document object 150 to be removed, while leaving the other finger stationary on the other document object 150. This is analogous to pulling the group apart, and this gesture can be interpreted by the review system as a command to separate the moved document object 150 from the remainder of the group.
When the user seeks to exit the review system 100 but would like to retain the state of the virtual workspace 100, the review system 100 can enable the user to save the current state of the virtual workspace 120. For example, the review system 100 can export the virtual workspace 120 by printing to paper, printing to Adobe PDF, or exported to an image. For further example, the review system 100 can be associated with a proprietary document format. If the user saves the virtual workspace 120 in this format, then the user can return to the virtual workspace 120 to continue active reading in the same state in which the virtual workspace 120 was saved.
Embodiments of the review system can thus be used to facilitate active reading, by providing a fluid-like, non-rigid, reading environment customizable by a user. While the review system has been disclosed in exemplary forms, many modifications, additions, and deletions may be made without departing from the spirit and scope of the system, method, and their equivalents, as set forth in the following claims.

Claims

CLAIMS What is claimed is:
1. A system comprising:
a virtual workspace;
a multi-touch interface to the virtual workspace;
a first document view region for displaying a first document; and
two or more document objects displayed in the virtual workspace, each comprising a corresponding text block and each relating to a corresponding portion of the first document, the document objects being manipulable and moveable about the virtual workspace in response to user interactions with the multi-touch interface;
wherein a first document object is linked to a corresponding first portion of the document, wherein selection of the first document object initiates an automatic modification of the first document view region to display the corresponding first portion of the document.
2. The system of Claim 1, the virtual workspace being resizable in response to a resize gesture received on the multi-touch interface, wherein resizing the virtual workspace resizes all of the two or more document objects.
3. The system of Claim 1, the corresponding text block of the first document object being an excerpt from the first portion of the document.
4. The system of Claim 1, the corresponding text block of the first document object being an annotation to the first portion of the document.
5. The system of Claim 1, further comprising a preview region for displaying the document at a first magnification, the first document view region displaying the document at a second magnification differing from the first magnification.
6. The system of Claim 5, the preview region being configured to guide navigation of the document in the first document view region, wherein, in response to receiving a selection of a first portion of the document in the preview region, the first document view region displays the first portion of the document.
7. The system of Claim 5, the preview region being configured to display the entire document simultaneously.
8. The system of Claim 7, the first document view region being configured to display the document at a higher magnification than the document is displayed in the preview region.
9. The system of any of Claims 1, the first document object being attachable to a second document object in the virtual workspace.
10. The system of Claim 9, wherein a command to move the first document object about the virtual workspace is interpreted as a command to the move the second document object attached to the first document object, to maintain a fixed spatial relationship between the first document object and the second document objects.
11. The system of either of Claims 9 or 10, the corresponding text block of the first document object being an excerpt from the first portion of the document, and the corresponding text block of the second document object being an annotation to the first portion of the document.
12. The system of any of Claims 1-10, further comprising a magnification object in the virtual workspace, wherein placement of the magnification object over the first document object enlarges at least a portion of the first document object.
13. The system of Claim 12, wherein rotating the magnification object increases the degree of magnification of the first document object, when the magnification object is placed over the first document object.
14. The system of Claim 12, wherein enlarging the magnification object modifies the range of the magnification object.
15. The system of Claim 12, the magnification object being provided in the virtual workspace in response to detection of a magnification gesture.
16. The system of Claim 15, the magnification gesture being a simultaneous touching of three fingers to a multi-touch interface.
17. The system of Claim 16, wherein an initial size of the magnification gesture is selected to correspond to the placement of the three fingers in the magnification gesture.
18. The system of any of Claims 1-10, the first document object corresponding to a second portion of the first document, in addition to the first portion of the first document, wherein selection of the first document object initiates an automatic modification of the first document view region to display the corresponding first portion of the first document and the corresponding second portion of the document.
19. The system of Claim 18, the first portion of the document and the second portion of the document being separated by an intermediate portion of the document, wherein of the first document object initiates a size reduction of the intermediate portion of the document to fit both the first and second portions of the document in the first document view region.
20. The system of any of Claims 1-10, further comprising a second document view region for displaying a second document.
21. The system of Claim 20, the first document object corresponding to a portion of the second document, in addition to the first portion of the first document, wherein selection of the first document object initiates an automatic modification of the first document view region to display the corresponding first portion of the document and an automatic modification of the second document view region to display the corresponding portion of the second document.
22. A method comprising:
providing a virtual workspace;
displaying a document in the virtual workspace;
receiving a collapse gesture indicating a first portion of the document and second portion of the document, wherein a first intermediate portion of the document is positioned between the first portion and the second portion; and
repositioning at least one of the first and second portions of the document by modifying the first intermediate portion of the document in response to the collapse gesture;
wherein the first and second portions of the document are closer together after the repositioning than before the repositioning.
23. The method of Claim 22, further comprising automatically reducing the size of the first intermediate portion of the document in response to the collapse gesture.
24. The method of Claim 23, further comprising:
receiving a magnification gesture;
providing a magnification object in the virtual workspace in response to the
magnification gesture; positioning the magnification object over the reduced first intermediate portion of the document; and
magnifying at least a portion of the reduced intermediate portion of the document in response to the positioning of the magnification object over the shrunken intermediate portion of the document.
25. The method of Claim 24, wherein receiving the magnification gesture comprises receiving a three-finger touch on a multi-touch interface.
26. The method of Claim 24, further comprising:
rotating the magnification object; and
increasing the magnification of the intermediate portion of the document under the magnification object in response to the rotation of the magnification object.
27. The method of any of Claims 24-26, further comprising:
expanding the magnification object; and
increasing the amount of the intermediate portion of the document that is magnified under the magnification object, in response to the expansion of the magnification object.
28. The method of Claim 22, further comprising:
providing a preview region in the virtual workspace, the document being displayed in the preview region; and
providing a view region in the virtual workspace, the document being displayed in the view region at a larger magnification than in the preview region.
29. The method of Claim 28, wherein the collapse gesture comprises a selecting the first and second portions of the document in the preview region.
30. The method of Claim 28, wherein the collapse gesture comprises a selecting the first portion of the document in the preview region and selecting the second portion of the document in the view region.
31. The method of Claim 22, the collapse gesture comprising a vertical pinching gesture.
32. The method of Claim 22, further comprising:
applying a first characteristic to the first portion of the document; and
applying the first characteristic to the second portion of the document; wherein the collapse gesture comprises a filter gesture indicating a command to filter the document based on the first characteristic.
33. The method of Claim 32, the first characteristic being highlighting, wherein non- highlighted portions of the document are reduced in size in response to the collapse gesture.
34. The method of any of Claims 22, further comprising:
receiving an uncollapse gesture; and
expanding the first intermediate portion of the document in response to the uncollapse gesture.
35. The method of Claim 34, the uncollapse gesture comprising an upward swiping over the first intermediate portion of the document.
36. The method of Claim 35, wherein expanding the first intermediate portion of the document in response to the uncollapse gesture comprises expanding the first intermediate portion of the document in an upward direction.
37. The method of Claim 34, the uncollapse gesture comprising a downward swiping over the first intermediate portion of the document.
38. The method of Claim 37, wherein expanding the first intermediate portion of the document in response to the uncollapse gesture comprises expanding the first intermediate portion of the document in a downward direction.
39. The method of any of Claims 28-38, the collapse gesture further indicating a third portion of the document, wherein a second intermediate portion of the document separates the third portion of the document from the second portion, and wherein the method further comprising repositioning the third portion of the document by modifying the second intermediate portion of the document in response to the collapse gesture.
40. The method of Claim 39, further comprising automatically shrinking the second intermediate portion of the document in response to the collapse gesture.
41. A method comprising:
providing a virtual workspace;
displaying at least a portion of a document in a document view region of the virtual workspace; receiving an extraction gesture indicating a first portion of the document and a first location in the virtual workspace, the first location being in the virtual workspace outside of the document; and
creating an independent excerpt object at the first location containing a duplicate of the first portion of the document, in response to the extraction gesture, the placement of the excerpt object being independent of the document and moveable about the virtual workspace.
providing a first visible link between the excerpt object and the first portion of the document, the visible link being positioned proximate the excerpt object;
receiving a selection of the first visible link; and
modifying the document view region to display the first portion of the document, if the first portion of the document is not currently displayed in the document view region, in response to the selection of the first visible link.
42. The method of Claim 41, further comprising:
providing a second visible link between the excerpt object and the first portion of the document, the second visible link being positioned proximate the first portion of the document; receiving a selection of the second visible link; and
recalling the excerpt object in response to the selection of the second visible link.
43. The method of Claim 42, further comprising providing a document object positioned in the virtual workspace to obstruct the view of the excerpt object, wherein recalling the excerpt object comprises repositioning the excerpt object to avoid the obstruction of the document object.
44. The method of any of Claims 41-43, further comprising:
receiving a relocation gesture indicating the excerpt object and the a second location in the virtual workspace outside of the document; and
relocating the excerpt object from the first location to the second location in response to the relocation gesture, while maintaining a current position of the document.
45. A method comprising:
providing a virtual workspace; displaying a first document view region in the virtual workspace, wherein a viewable portion of a first document is visible in the document view region, and wherein a hidden portion of the first document is hidden from the document view region;
receiving a selection of a first portion of the first document within the viewable portion of the first document;
receiving an annotation input;
creating an annotation object at a first location in the virtual workspace outside of the first document, in response to receiving the annotation input, the annotation object comprising the annotation input;
providing proximate the annotation object a first graphical link to the first portion of the first document;
modifying the first document view region to hide the first portion of the first document from the document view region;
receiving a selection of the first graphical link proximate the annotation object; and modifying the first document view region to display the first portion of the first document, in response to the selection of the first graphical link.
46. The method of Claim 45, the annotation object being moveable about the virtual workspace independent of the placement of the first document.
47. The method of Claims 45, further comprising:
receiving a relocation gesture indicating the annotation object and a second location in the virtual workspace outside of the document; and
relocating the annotation object from the first location to the second location in response to the relocation gesture, while maintaining a current position of the document.
48. The method of Claim 45, further comprising:
providing a second graphical link between the annotation object and a second portion of the first document, the second graphical link being positioned proximate the annotation object; receiving a selection of the second graphical link in combination with the selection of the first graphical link; and modifying the first document view region to display simultaneously both the first portion of the first document and the second portion of the first document, in response to the selections of the first and second graphical links.
49. The method of Claim 45, wherein the first graphical link additionally links the annotation object to a second portion of the document, the method further comprising modifying the first document view region to display simultaneously both the first portion of the first document and the second portion of the first document, in response to the selection of the first graphical link.
50. The method of either of Claims 48 or 49, the first portion of the document and the second portion of the document being separated by an intermediate portion of the document, wherein modifying the first document view region to display simultaneously both the first portion of the first document and the second portion of the first document comprises modifying the
intermediate portion of the document.
51. The method of Claim 50, wherein modifying the first document view region to display simultaneously both the first portion of the first document and the second portion of the first document comprises reducing the vertical breadth of the intermediate portion of the document.
52. The method of Claim 45, further comprising a second document view region for displaying a second document.
53. The method of Claim 52, further comprising:
providing a second graphical link between the annotation object and a linked portion of the second document, the second graphical link being positioned proximate the annotation object;
receiving a selection of the second graphical link in combination with the selection of the first graphical link; and
modifying the first document view region to display the first portion of the first document, and modifying the second document view region to display the linked portion of the second document, in response to the selections of the first and second graphical links.
54. The method of Claim 45, further comprising: providing a second graphical link between the annotation object and the first portion of the first document, the second graphical link being positioned proximate the first portion of the first document;
receiving a selection of the second graphical link; and
recalling the annotation object in response to the selection of the second graphical link.
55. The method of Claim 54, further comprising providing a document object positioned in the virtual workspace to obstruct the view of the annotation object, wherein recalling the excerpt object comprises repositioning the excerpt object to avoid the obstruction of the document object.
56. A method comprising:
displaying a virtual workspace;
displaying a first object in the virtual workspace, the first object having a first initial size and comprising a first text block;
displaying a second object in the virtual workspace, the second object having a second initial size and comprising a second text block;
receiving a magnification gesture;
providing a magnification object in the virtual workspace in response to the
magnification gesture;
positioning the magnification object over the first object; and
modifying the size of at least a portion of the first object in response to the positioning of the magnification object over the first object, while maintaining the second initial size of the second object.
57. The method of Claim 56, wherein modifying the size of at least a portion of the first object comprises modifying the size of the entire first object.
58. The method of Claim 56, wherein modifying the size of at least a portion of the first object comprises modifying the size of less than the entire first object.
59. The method of Claim 56, wherein receiving the magnification gesture comprises receiving a simultaneous touch of three fingers on a multi-touch interface.
60. The method of Claim 59, further comprising establishing an initial size of the magnification object to fit a layout of the three fingers.
61. The method of any of Claims 56-60, wherein rotating the magnification object modifies a magnification of the at least a portion of the first object, when the magnification object is positioned over the first object.
62. The method of any of Claims 56-60, further comprising receiving an enlargement gesture directed at the magnification object.
63. The method of 62, wherein receiving the enlargement gesture comprises detecting an increase in a space between three fingers on a multi-touch interface.
64. The method of Claim 62, further comprising enlarging the magnification object in response to the enlargement gesture.
65. A method comprising:
providing a virtual workspace;
displaying a first document in a first document view region of the virtual workspace; displaying a second document in a second document view region of the virtual workspace;
receiving a linking gesture indicating a first section in the first document and a second section in the second document;
displaying a first graphical link proximate the first section in the first document, in response to the linking gesture;
displaying a second graphical link proximate the second section in second document, in response to the linking gesture;
receiving a selection of the first graphical link; and
modifying the second document view region to make the second section in the second document visible in the second document region, in response to the selection of the first graphical link, if the second section is not visible in the second document view region at the time of the selection of the first graphical link.
66. The method of Claim 65, further comprising:
receiving a selection of the second graphical link; and modifying the first document view region to make the first section in the first document visible in the first document region, in response to the selection of the second graphical link, if the first section is not visible in the first document view region at the time of the selection of the second graphical link.
67. The method of either of Claims 65 or 66, the linking gesture comprising a simultaneous touching directed at the first and second sections on a multi-touch interface.
68. A method comprising:
providing a virtual workspace;
receiving a first continuous touch in a predefined region of the virtual workspace;
saving the current state of the virtual workspace as a first saved state, in response to the first continuous touch;
modifying the virtual workspace;
receiving a second continuous touch in the predefined region of the virtual workspace, while the first continuous touch is ongoing;
saving the current state of the virtual workspace as a second saved state, in response to the second continuous touch;
detection a termination of the first continuous touch; and
restoring the first saved state of the virtual workspace, in response to termination of the first continuous touch.
69. The method of Claim 68, further comprising displaying a document in the predefined region of the virtual workspace, wherein modifying the virtual workspace comprises modifying a layout of the document, and wherein saving the current state of the virtual workspace as the first state comprises saving the current state of the layout of the document.
70. The method of either of claims 68 or 69, further comprising providing one or more document objects in the virtual workspace and freely moveable about the virtual workspace, wherein modifying the virtual workspace comprises modifying document objects, and wherein saving the current state of the virtual workspace as the first state comprises saving the current state of the document objects.
71. A system comprising:
a virtual workspace; a touchscreen input device for interacting with the virtual workspace;
a document displayed in the virtual workspace;
a detection system for detecting a gesture on the touchscreen input device, the gesture indicating a first portion of the document and second portion of the document, wherein an intermediate portion of the document is positioned between the first portion and the second portion; and
a display system configured to reduce the size of the intermediate portion of the document in response to the detected gesture.
72. A system comprising:
a virtual workspace;
a touchscreen input device for interacting with the virtual workspace;
a document displayed in a document view region of the virtual workspace;
a detection system for detecting a gesture on the touchscreen input device, the gesture indicating a first portion of the document and a first location in the virtual workspace, the first location being in the virtual workspace outside of the document; and
a display system for creating an independent excerpt object at the first location containing a duplicate of the first portion of the document, in response to the extraction gesture, the placement of the excerpt object being independent of the document and moveable throughout the virtual workspace;
wherein the display system is further configured to provide a first visible link between the excerpt object and the first portion of the document, the visible link being positioned proximate the excerpt object;
wherein the detection system is further configured to receive a selection of the first visible link; and
wherein the display system is further configured to modify the document view region to display the first portion of the document, in response to the selection of the first visible link, if the first portion of the document is not currently displayed in the document view region.
73. A system comprising:
a virtual workspace;
a touchscreen input device for interacting with the virtual workspace;
a document displayed in a document view region of the virtual workspace; a detection system for receiving a selection of a first portion of the document and for receiving an annotation input; and
a display system for displaying an annotation object at a first location in the virtual workspace outside of the document, in response to receiving the annotation input, the annotation object comprising the annotation input;
wherein the display system is further configured to provide proximate the annotation object a first graphical link to the first portion of the document;
wherein the detection system is further configured to receiving a selection of the first graphical link; and
wherein the display system is further configured to modify the document view region to display the first portion of the document, in response to the selection of the first graphical link, if the first portion of the document is not currently displayed.
74. A system comprising:
a virtual workspace;
a touchscreen input device for interacting with the virtual workspace;
a display system for displaying a first object in the virtual workspace, the first object having a first initial size and comprising a first text block, and for displaying a second object in the virtual workspace, the second object having a second initial size and comprising a second text block; and
a detection system for receiving a magnification gesture;
wherein the display system is further configured to provide a magnification object in the virtual workspace in response to the magnification gesture, and to modify the size of at least a portion of the first object in response to positioning of the magnification object over the first object, while maintaining the second initial size of the second object.
75. The system of any of Claims 71-74, the touch screen input device being configured to detect multiple simultaneous touches.
PCT/US2010/050911 2009-09-30 2010-09-30 Systems and methods to facilitate active reading WO2011041547A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US13/876,463 US10268661B2 (en) 2010-09-30 2011-03-16 Systems and methods to facilitate active reading
PCT/US2011/028595 WO2012044363A1 (en) 2010-09-30 2011-03-16 Systems and methods to facilitate active reading
US16/391,003 US11704473B2 (en) 2009-09-30 2019-04-22 Systems and methods to facilitate active reading

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US24727909P 2009-09-30 2009-09-30
US61/247,279 2009-09-30

Related Child Applications (2)

Application Number Title Priority Date Filing Date
PCT/US2011/028595 Continuation WO2012044363A1 (en) 2009-09-30 2011-03-16 Systems and methods to facilitate active reading
US13/876,463 Continuation US10268661B2 (en) 2010-09-30 2011-03-16 Systems and methods to facilitate active reading

Publications (1)

Publication Number Publication Date
WO2011041547A1 true WO2011041547A1 (en) 2011-04-07

Family

ID=43826654

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/050911 WO2011041547A1 (en) 2009-09-30 2010-09-30 Systems and methods to facilitate active reading

Country Status (1)

Country Link
WO (1) WO2011041547A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103019531A (en) * 2011-09-22 2013-04-03 汉王科技股份有限公司 On-line display method and device for electronic newspaper content
EP2575019A1 (en) * 2011-09-30 2013-04-03 France Telecom Method for bookmarking a section of an electronic content
CN103076957A (en) * 2013-01-17 2013-05-01 上海斐讯数据通信技术有限公司 Display control method and mobile terminal
EP2631757A1 (en) * 2012-02-24 2013-08-28 Research In Motion Limited Navigation of content media displayed on a touch screen of an electronic device
US20130265246A1 (en) * 2012-04-06 2013-10-10 Lg Electronics Inc. Electronic device and method of controlling the same
JP2014238667A (en) * 2013-06-06 2014-12-18 株式会社リコー Information terminal, information processing program, information processing system, and information processing method
US9250328B2 (en) 2009-09-30 2016-02-02 Javad Gnss, Inc. Graphics-aided remote position measurement with handheld geodesic device
CN108874264A (en) * 2018-06-13 2018-11-23 奇酷互联网络科技(深圳)有限公司 Application program image target aligning method, device, readable storage medium storing program for executing and terminal

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050055628A1 (en) * 2003-09-10 2005-03-10 Zheng Chen Annotation management in a pen-based computing system
US20050052427A1 (en) * 2003-09-10 2005-03-10 Wu Michael Chi Hung Hand gesture interaction with touch surface
US20050183005A1 (en) * 2004-02-12 2005-08-18 Laurent Denoue Systems and methods for freeform annotations

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050055628A1 (en) * 2003-09-10 2005-03-10 Zheng Chen Annotation management in a pen-based computing system
US20050052427A1 (en) * 2003-09-10 2005-03-10 Wu Michael Chi Hung Hand gesture interaction with touch surface
US20050183005A1 (en) * 2004-02-12 2005-08-18 Laurent Denoue Systems and methods for freeform annotations

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9250328B2 (en) 2009-09-30 2016-02-02 Javad Gnss, Inc. Graphics-aided remote position measurement with handheld geodesic device
CN103019531A (en) * 2011-09-22 2013-04-03 汉王科技股份有限公司 On-line display method and device for electronic newspaper content
EP2575019A1 (en) * 2011-09-30 2013-04-03 France Telecom Method for bookmarking a section of an electronic content
EP2631757A1 (en) * 2012-02-24 2013-08-28 Research In Motion Limited Navigation of content media displayed on a touch screen of an electronic device
US20130265246A1 (en) * 2012-04-06 2013-10-10 Lg Electronics Inc. Electronic device and method of controlling the same
CN103365591A (en) * 2012-04-06 2013-10-23 Lg电子株式会社 Electronic device and method of controlling same
EP2648085A3 (en) * 2012-04-06 2015-02-18 LG Electronics Inc. Electronic device with a touch screen and method of controlling the same
AU2013201324B2 (en) * 2012-04-06 2015-06-11 Lg Electronics Inc. Electronic device and method of controlling the same
CN103365591B (en) * 2012-04-06 2016-06-29 Lg电子株式会社 Electronic equipment and the method controlling electronic equipment
US9720529B2 (en) 2012-04-06 2017-08-01 Lg Electronics Inc. Electronic device and method of controlling the same
CN103076957A (en) * 2013-01-17 2013-05-01 上海斐讯数据通信技术有限公司 Display control method and mobile terminal
JP2014238667A (en) * 2013-06-06 2014-12-18 株式会社リコー Information terminal, information processing program, information processing system, and information processing method
CN108874264A (en) * 2018-06-13 2018-11-23 奇酷互联网络科技(深圳)有限公司 Application program image target aligning method, device, readable storage medium storing program for executing and terminal

Similar Documents

Publication Publication Date Title
US10268661B2 (en) Systems and methods to facilitate active reading
US10417309B2 (en) Facilitating active reading of digital documents
US5347295A (en) Control of a computer through a position-sensed stylus
JP4063246B2 (en) Page information display device
US9424241B2 (en) Annotation mode including multiple note types for paginated digital content
US6891551B2 (en) Selection handles in editing electronic documents
US11120203B2 (en) Editing annotations of paginated digital content
JP4637455B2 (en) User interface utilization method and product including computer usable media
WO2011041547A1 (en) Systems and methods to facilitate active reading
US20180101509A1 (en) Filtering and searching annotations of paginated digital content
US7818691B2 (en) Zeroclick
US9411461B2 (en) Moveable interactive shortcut toolbar and unintentional hit rejecter for touch input devices
US8941600B2 (en) Apparatus for providing touch feedback for user input to a touch sensitive surface
Hinckley et al. InkSeine: In Situ search for active note taking
RU2557463C2 (en) Dual screen portable touch sensitive computing system
US20140189593A1 (en) Electronic device and input method
US20110216015A1 (en) Apparatus and method for directing operation of a software application via a touch-sensitive surface divided into regions associated with respective functions
US10915698B2 (en) Multi-purpose tool for interacting with paginated digital content
JP2003531428A (en) User interface and method of processing and viewing digital documents
Matulic et al. Pen and touch gestural environment for document editing on interactive tabletops
JPH06502734A (en) Computer document as a notebook-type compound document
JP2007334910A (en) Graphical user interface for help system
JP2004355106A (en) Touch interface of computer
JP3256994B2 (en) Display target movement method by touch input
JPH06175775A (en) Information processor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10821255

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10821255

Country of ref document: EP

Kind code of ref document: A1