US8276088B2 - User interface for three-dimensional navigation - Google Patents
User interface for three-dimensional navigation Download PDFInfo
- Publication number
- US8276088B2 US8276088B2 US11/827,530 US82753007A US8276088B2 US 8276088 B2 US8276088 B2 US 8276088B2 US 82753007 A US82753007 A US 82753007A US 8276088 B2 US8276088 B2 US 8276088B2
- Authority
- US
- United States
- Prior art keywords
- user interface
- pose
- change
- image
- feature
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 claims abstract description 188
- 238000000605 extraction Methods 0.000 claims abstract description 98
- 230000008859 change Effects 0.000 claims description 56
- 230000015654 memory Effects 0.000 claims description 26
- 238000003384 imaging method Methods 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 5
- 230000004044 response Effects 0.000 claims description 4
- 238000004590 computer program Methods 0.000 claims description 3
- 238000009877 rendering Methods 0.000 abstract description 2
- 230000008569 process Effects 0.000 description 43
- 230000009466 transformation Effects 0.000 description 30
- 230000002093 peripheral effect Effects 0.000 description 29
- 238000012545 processing Methods 0.000 description 28
- 238000004891 communication Methods 0.000 description 25
- 230000033001 locomotion Effects 0.000 description 23
- 238000010586 diagram Methods 0.000 description 12
- 230000007704 transition Effects 0.000 description 12
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 11
- 238000013139 quantization Methods 0.000 description 11
- 230000000007 visual effect Effects 0.000 description 7
- 238000013500 data storage Methods 0.000 description 5
- 230000007246 mechanism Effects 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 230000002829 reductive effect Effects 0.000 description 4
- 238000013519 translation Methods 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 3
- 230000001413 cellular effect Effects 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000005192 partition Methods 0.000 description 3
- 230000000737 periodic effect Effects 0.000 description 3
- 238000000513 principal component analysis Methods 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 230000007812 deficiency Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 238000007639 printing Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 239000000872 buffer Substances 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 230000000593 degrading effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
- G06V30/142—Image acquisition using hand-held instruments; Constructional details of the instruments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
- G06V30/1444—Selective acquisition, locating or processing of specific regions, e.g. highlighted text, fiducial marks or predetermined fields
- G06V30/1456—Selective acquisition, locating or processing of specific regions, e.g. highlighted text, fiducial marks or predetermined fields based on user interactions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Definitions
- the present invention relates to processing techniques for patch recognition. More particularly, the present invention relates to systems and methods for using a new feature referred to herein as invisible junctions for image-based document patch recognition. Still more particularly, the present invention relates to a system and methods for using invisible junctions to provide a user interface for three-dimensional navigation.
- PDA personal digital assistant
- cellular telephones e.g., cellular camera phones
- digital cameras over the past few years
- SIFT Scale-Invariant Feature Transform
- the SIFT key points are chosen from the extrema in scale space. More specifically, all scales and image locations are scanned for local maxima in scale space; these local maxima are chosen as key point candidates. This makes the SIFT poor at discrimination between text and the SIFT is not stable and repeatable in noisy environments.
- Other prior art approaches focus on geometric features of the text block but they are not suitable for Asian or ideographic languages.
- One additional problem with the prior art is that there is not a method for indicating the region of an electronic document that corresponds to a query image being input. In particular, there are not prior art methods for indicating the viewing region and a camera look at point on the electronic document.
- the present invention overcomes the deficiencies of the prior art with a system using invisible junctions for image-based document patch recognition.
- the system is advantageous because it uses invisible junctions of a low quality image patch to identify and retrieve a corresponding electronic document, page, look-at point and viewing region.
- the system is advantageous because it is very fast, works with both western and eastern languages, and works images that mix text and image components.
- the present invention uses invisible junctions which are a set of local features unique to every page of the electronic document to match the captured image to a part of an electronic document.
- the system of the present invention includes: an image capture device, a feature extraction and recognition system and database.
- the feature extraction and recognition system further comprises a feature extraction unit, a feature indexing unit, a feature retrieval unit and a geometric estimation unit.
- the feature extraction and recognition system When an electronic document is printed, the feature extraction and recognition system, for example in print-capture software, captures an image of the document page. The features in the captured image are then extracted, indexed and stored in the database. Given a query image, usually a small patch of some document page captured by a low resolution image capture device such as a web-cam, the features in the query image are extracted and compared against those stored in the database to identify the query image.
- the feature extraction and recognition system advantageously uses geometric estimation to reduce the query result to a single one or a few candidate matches. In one embodiment, the feature extraction and recognition system uses two separate geometric estimations to rank and verify matching candidates.
- the present invention also includes methods for feature extraction, feature indexing, feature retrieval and geometric estimation.
- the present invention also includes methods for recognizing and tracking the viewing region and look at point corresponding to the input query image. Moreover, the present invention is particularly advantageous because this information is combined with a rendering of the original input document to generate a new graphical user interface for the user. This user interface can be displayed on a conventional browser or even on the display of an image capture device.
- FIG. 1 is a functional diagram of an embodiment of a system for feature extraction and recognition in accordance with present invention.
- FIG. 2 illustrates a block diagram of an embodiment of a system configured in accordance with an embodiment of the present invention.
- FIG. 3 illustrates a block diagram of a memory of the system of FIG. 2 configured in accordance with an embodiment of the present invention.
- FIG. 4 is a block diagram of an embodiment of a Feature Extraction and Recognition system of FIG. 1 in accordance with the present invention.
- FIGS. 5A to 5E are graphical representations of the steps of feature extraction being performed on an example image in accordance with the present invention.
- FIG. 6 is a flow chart of an embodiment of a method for performing feature extraction in accordance with the present invention.
- FIG. 7A is a flow chart of an embodiment of a method for performing feature indexing in accordance with the present invention.
- FIG. 7B is a flow chart of an alternate embodiment of a method for performing feature indexing in accordance with the present invention.
- FIG. 8 is a flow chart of an embodiment of a method for feature retrieval in accordance with the present invention.
- FIG. 9A is a flow chart of a first embodiment of a method for geometric estimation in accordance with the present invention.
- FIG. 9B is a flow chart of a second embodiment of a method for geometric estimation in accordance with the present invention.
- FIG. 9C is a flow chart of a second embodiment of a method for geometric estimation in accordance with the present invention.
- FIG. 9D is a graphical representation of example geometric constraints showing a paper plane, an imaging plane and a query.
- FIG. 10A is a flow chart of an embodiment of a method for combining recognizing and tracking in accordance with the present invention.
- FIG. 10B is a graphical representation of a user interface generated by the method for combining recognizing and tracking in accordance with the present invention.
- FIG. 10C is a state diagram of a process used by the method for combining recognizing and tracking in accordance with the present invention.
- FIGS. 10D-10K are graphical representations of other embodiment for a user interface generated by the combined recognizing and tracking in accordance with the present invention.
- FIG. 11 is a functional block diagram of multifunction peripheral including of an embodiment of the system for feature extraction and recognition in accordance with present invention.
- FIG. 12 is a flow chart of an embodiment of a method for using invisible junction recognition as a part of document security in accordance with the present invention.
- FIG. 13 is a flow chart of an embodiment of a method for using invisible junction recognition as a part of database updating and document processing in accordance with the present invention.
- FIG. 14 is a flow chart of an embodiment of a method for using invisible junction recognition as a part of mixed media reality database annotation and document processing in accordance with the present invention.
- FIG. 15 is a flow chart of an embodiment of a method for using invisible junction recognition as a part of a user interface to control display of a three-dimensional model in accordance with the present invention.
- FIG. 16 is a flow chart of an embodiment of a method for updating a display of a three-dimensional model based on a change in input device position.
- FIGS. 17A to 17I are graphical representations of user interfaces generated by the method of FIG. 15 in accordance with the present invention.
- FIG. 18 is a flow chart of an embodiment of a method for updating a display of navigation information based on a change in input device position.
- FIGS. 19A to 19H are graphical representations of user interfaces generated by the method of FIG. 18 in accordance with the present invention.
- the present invention also relates to an apparatus for performing the operations herein.
- This apparatus may be specially constructed for the required purposes, or it may comprise a general-purpose computer selectively activated or reconfigured by a computer program stored in the computer.
- a computer program may be stored in a computer readable storage medium, such as, but is not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, each coupled to a computer system bus.
- the system 100 comprises: an image capture device 106 , a feature extraction and recognition system 108 and a database 110 configured to capture an image of a portion 104 of a printed or paper document 102 and retrieve the corresponding electronic document, page, look-at point and viewing region.
- the correspondence between the original area 104 of the paper document 102 and the electronic representation 112 is shown by line 114 .
- FIG. 1 illustrates an example of the system 100 capturing 116 a query image such as a small patch 104 of some paper document 102 with the image capture device 106 .
- the feature extraction and recognition system 108 extracts the features in the query image and compares them against those stored in the database 110 to identify the electronic representation 112 .
- the image capture device 106 is any conventional type known in the art.
- the image capture device 106 may be a web camera, a camera that is part of cell phone or a digital camera.
- the present invention is particularly advantageous because it is operable with low resolution image capture devices.
- the image capture device 106 is coupled for communication with the feature extraction and recognition system 108 . This communication may be via a wired or wireless network.
- the image captured device 106 is a scanner of a multifunction peripheral or software for generating a printed document.
- the image captured device 106 could be print capture software in such an embodiment.
- the feature extraction and recognition system 108 performs the functions of feature extraction, indexing and retrieval for invisible junctions.
- the feature extraction and recognition system 108 further comprises a feature extraction unit 310 , a feature indexing unit 312 , a feature retrieval unit 314 and a geometric estimation unit 316 as will be described in more detail below with reference to FIGS. 2-4 .
- the feature extraction and recognition system 108 is coupled to the image capture device 106 to receive an image of a paper document, a model image of the paper document or an image patch.
- the feature extraction and recognition system 108 is coupled to the database 110 to store feature descriptions of document pages and electronic document or indices to electronic documents.
- the feature extraction and recognition system 108 is also coupled to the database 110 to retrieve electronic document or indices to electronic documents during retrieval.
- the feature extraction and recognition system 108 is also coupled to output the retrieved electronic document, page, look-at point and viewing region to other computing systems for display with user interfaces.
- the database 110 is a conventional type and stores indices, electronic documents, feature descriptions and other information used in the indexing and retrieval process.
- the database may be a federation of databases.
- the database 110 is stored within a multifunction peripheral device.
- the system 100 of the present invention is particularly advantageous because it provides a number of benefits.
- the invisible junction provides stability or independence from noise and viewing condition; is repeatable in both the query image and the model image; provides strong discrimination, and is low dimensional.
- Second, the present invention provides a very fast recognition and retrieval at a rate of nearly 4 frames per second on a database of 10000 pages. The recognition and retrieval is fast enough for real-time interactive applications.
- Third, the present invention can use noisy low quality input image such as available with a cell phone camera for recognition and retrieval.
- the retrieval operates accurately despite image rotation, scale changes and occlusions.
- the present invention can perform recognition and retrieval on both eastern and western languages.
- the present invention can perform recognition and retrieval on input images that include both text and graphic images in any portion for text to image.
- the feature extraction and recognition system 108 preferably comprises a control unit 250 , a display device 210 , an input device 212 , cursor control 214 , a communication device 216 , and one or more input/output (I/O) devices 218 .
- the control unit 250 comprises an arithmetic logic unit, a microprocessor, a general purpose computer or some other information appliance equipped to provide electronic display signals to display device 210 .
- the control unit 250 comprises a general purpose computer having a graphical user interface, which may be generated by, for example, a program written in Java running on top of an operating system like WINDOWS® or UNIX® based operating systems.
- one or more application programs are executed by control unit 250 including, without limitation, drawing applications, word processing applications, electronic mail applications, financial applications and web browser applications.
- control unit 250 is shown including processor 202 , main memory 204 and data storage device 206 , all of which are communicatively coupled to system bus 208 .
- Processor 202 processes data signals and may comprise various computing architectures including a complex instruction set computer (CISC) architecture, a reduced instruction set computer (RISC) architecture, or an architecture implementing a combination of instruction sets. Although only a single processor is shown in FIG. 2 , multiple processors may be included.
- CISC complex instruction set computer
- RISC reduced instruction set computer
- Main memory 204 stores instructions and/or data that may be executed by processor 202 .
- the instructions and/or data may comprise code for performing any and/or all of the techniques described herein.
- Main memory 204 may be a dynamic random access memory (DRAM) device, a static random access memory (SRAM) device, or some other memory device known in the art.
- DRAM dynamic random access memory
- SRAM static random access memory
- the memory 204 is described in more detail below with reference to FIG. 3 .
- Data storage device 206 stores data and instructions for processor 202 and comprises one or more devices including a hard disk drive, a floppy disk drive, a CD-ROM device, a DVD-ROM device, a DVD-RAM device, a DVD-RW device, a flash memory device, or some other mass storage device known in the art.
- data storage device 206 includes database 110 .
- the database is coupled to the feature extraction and recognition system 108 via communication device 216 and signal line 220 representing a conventional network.
- System bus 208 represents a shared bus for communicating information and data throughout control unit 250 .
- System bus 208 may represent one or more buses including an industry standard architecture (ISA) bus, a peripheral component interconnect (PCI) bus, a universal serial bus (USB), or some other bus known in the art to provide similar functionality.
- Additional components coupled to control unit 250 through system bus 208 include the display device 210 , the input device 212 , cursor control 214 , the communication device 216 and the I/O device(s) 218 .
- Display device 210 represents any device equipped to display electronic images and data as described herein.
- the display device 210 is a liquid crystal display (LCD) and light emitting diodes (LEDs) to provide status feedback, operation settings and other information to the user.
- the display device 210 may be, for example, a cathode ray tube (CRT) or any other similarly equipped display device, screen or monitor.
- display device 210 is equipped with a touch screen and/or includes a digitizer in which a touch-sensitive, transparent panel covers the screen of display device 210 .
- the input device 212 is a keyboard.
- the keyboard can be a QWERTY keyboard, a key pad, or representations of such created on a touch screen.
- Cursor control 214 represents a user input device equipped to communicate positional data as well as command selections to processor 202 .
- Cursor control 214 may include a mouse, a trackball, a stylus, a pen, a touch screen, cursor direction keys or other mechanisms to cause movement of a cursor.
- Communication device 216 links control unit 250 to a network 220 that may include multiple processing systems and in one embodiment is a network controller.
- the network of processing systems may comprise a local area network (LAN), a wide area network (WAN) (e.g., the Internet), and/or any other interconnected data path across which multiple devices may communicate.
- the control unit 250 also has other conventional connections to other systems such as a network for distribution of files (media objects) using standard network protocols such as TCP/IP, http, https, and SMTP as will be understood to those skilled in the art.
- One or more I/O devices 218 are coupled to the bus 208 . These I/O devices may be part of the other systems (not shown).
- the I/O device 218 can include an image scanner for capturing an image of a document.
- the I/O device 218 may also includes a printer for generating documents.
- the I/O device 218 may also include audio input/output device equipped to receive audio input via a microphone and transmit audio output via speakers.
- audio device is a general purpose; audio add-in/expansion card designed for use within a general purpose computer system.
- I/O audio device may contain one or more analog-to-digital or digital-to-analog converters, and/or one or more digital signal processors to facilitate audio processing.
- feature extraction and recognition system 108 may include more or less components than those shown in FIG. 2 without departing from the spirit and scope of the present invention.
- feature extraction and recognition system 108 may include additional memory, such as, for example, a first or second level cache, or one or more application specific integrated circuits (ASICs).
- the feature extraction and recognition system 108 may include multiple data buses rather than the single bus 208 . Multiple buses allow fast transfer of image data from and between the components of the feature extraction and recognition system 108 , and simultaneous data transfer of user interface information to a display device, for example.
- control unit 250 may be coupled to control unit 250 including, for example, an RFID tag reader, digital still or video cameras, or other devices that may or may not be equipped to capture and/or download electronic data to control unit 250 .
- control unit 250 may include, for example, an RFID tag reader, digital still or video cameras, or other devices that may or may not be equipped to capture and/or download electronic data to control unit 250 .
- One or more components could also be eliminated such as the keyboard & cursor control 212 .
- FIG. 3 is a block diagram of one embodiment of the memory unit 204 for the feature extraction and recognition system 108 .
- the memory unit 204 preferably comprises: an operating system 302 , an application 304 , a control module 306 , a browser 308 , a feature extraction unit 310 , a feature indexing unit 312 , a feature retrieval unit 314 , a geometric estimation unit 316 , a video tracker 320 , a paper tracker 322 , a recognizer 324 , a drift corrector 326 and a user interface module 328 .
- the memory 204 also includes buffers for storing data and other information temporarily during the indexing or retrieval process.
- the memory unit 204 stores instructions and/or data that may be executed by processor 202 .
- the instructions and/or data comprise code for performing any and/or all of the techniques described herein.
- These modules 302 - 314 are coupled by bus 208 to the processor 202 for communication and cooperation to system 100 .
- Those skilled in the art will recognized that while the present invention will now be described as modules or portions of a memory unit 204 of a computer, the modules or portions thereof may also be stored in other media such as permanent data storage device 206 and may be distributed across a network 220 having a plurality of different computers such as in a client/server environment.
- the operating system 302 is preferably a custom operating system that is accessible to user via an application interface.
- the operating system 302 is one of a conventional type such as, WINDOWS®, SOLARIS® or LINUX® based operating systems.
- the memory unit 204 also includes one or more application programs 304 including, without limitation, drawing applications, word processing applications, electronic mail applications, financial applications and web browser applications.
- the applications 304 specifically utilize the unique capabilities of the other modules or units of memory 204 to bridge the gap between electronic documents and paper documents by enabling users to establish and identify links between a printed paper and its corresponding electronic document. More specifically, the applications 304 access electronic data from a paper and attach/embed electronic data to specific locations on a paper so that a user can later retrieval that electronic data from the paper, using cameras as the tool for the above operations.
- one application 304 retrieves an electronic file (Microsoft Word, Adobe PDF, etc) of a printed paper in response to the user simply pointing a web-cam and capturing an image of the printed paper.
- Another example application 304 is a paper-based photo-wiki, which lets the user attach the pictures you have taken to specific locations on a paper map (and its electronic version), by simply pointing the camera to the desired locations on the physical paper map.
- There are many other applications 304 such as novel user interfaces to access or browse a document database, digitized pen and paper, map navigator, security in copiers and scanners, camera games and duplicate detection for copyright protection.
- the control module 306 is used to control the other modules of the memory 204 .
- the control module 306 is adapted for control of and communication with the application 304 , the browser 308 , the feature extraction unit 310 , the feature indexing unit 312 , the feature retrieval unit 314 and the geometric estimation unit 316 .
- the operation of the control module 306 will be apparent from the description of FIGS. 4 and 6 - 9 below. While the control module 306 is shown as a separate module of the memory 204 , those skilled in the art will recognize that the control module 306 in another embodiment may be distributed as routines in the other modules 310 - 316 .
- the browser 308 is a conventional type such as Internet Explorer by Microsoft® or Firefox by Mozilla.
- the Web browser is a software application that enables a user to display and interact with text, images, and other information typically located on a Web page at a website on the World Wide Web or a local area network.
- the feature extraction unit 310 is software and routines for extracting the invisible junction features from original electronic documents and scanned images of paper documents.
- the feature extraction unit 310 includes an input coupled to receive original electronic documents and an output to provide a feature description to the feature indexing unit 312 .
- the feature extraction unit 310 includes an input coupled to receive an image of a paper document and provide a feature description to the feature retrieval unit 314 .
- the feature extraction unit 310 includes a location determination module, a size determination module, an orientation determination module and a descriptor creator module coupled in series to produce the feature description. The operation of the feature extraction unit 310 is described below in detail with reference to FIGS. 5A-6 .
- the feature extraction unit 310 is coupled for communication with the feature indexing unit 312 and the feature retrieval unit 314 .
- the feature indexing unit 312 is software and routines for generating one or more indices of invisible junction feature descriptions and storing them in the database 110 .
- the feature indexing unit 312 uses a quantization tree to index the feature descriptions.
- the feature indexing unit 312 is coupled to an output of the feature extraction unit 310 to receive invisible junction feature descriptions.
- the feature indexing unit 312 is coupled to the database 110 . The operation of the feature indexing unit 312 is described in detail below with reference to FIGS. 7A and 7B .
- the feature retrieval unit 314 is software and routines for creating a query from the invisible junction feature descriptions of a paper document and communicating with the database 110 to retrieve a document, a point and a viewing region of a matching electronic document from the database 110 .
- the feature retrieval unit 314 is coupled to an output of the feature extraction unit 310 to receive invisible junction feature descriptions, and to the database 110 to perform queries and receive results.
- the possible matching documents, points and viewing regions are output by the feature retrieval unit 314 .
- the operation of the feature retrieval unit 314 is described in detail below with reference to FIG. 8 .
- the geometric estimation unit 316 is software and routines for sorting the candidate documents output by the feature retrieval unit 314 and selecting a document that includes the largest number of feature points consistent with the geometric constraint.
- the geometric estimation unit 316 is adapted for communication with the feature retrieval unit 314 and has an output that provides a document, a point and a viewing region that matches the input image patch. The operation of the geometric estimation unit 316 is described in detail below with reference to FIG. 9A .
- the present invention also includes the capability to recognize and track the look at position and viewing region base on an input query image. In addition to generating information indicating the document, look at position and the viewing region, the present invention merges that information with the original electronic document to generate novel user interfaces that provide the user with precise information to link very specific locations within the electronic document for actions such as adding annotations or controlling operations.
- This recognition and tracking capability is provides by the video tracker 320 , the paper tracker 322 , the recognizer 324 and the drift corrector 326 .
- the control module 306 also controls the operation of the video tracker 320 , the paper tracker 322 , the recognizer 324 and the drift corrector 326 consistent with the state diagram of FIG. 10C .
- another module such as a recognition and tracking module (not shown) could be use in place of the control module 306 . This functionality is described in more detail below with reference to FIGS. 10A and 10C .
- the video tracker 320 is software and routines for tracking camera motion as a projective transformation across video frames.
- the video tracker 320 uses the first video frame as the reference frame, and then outputs information indicating movement the camera. In essence, the video tracker 320 provides the relative motion of the camera between frames.
- the video tracker 320 is adapted for communication with the feature retrieval unit 314 to receive feature descriptors and has an output to provide generated camera motion information. The operation of the video tracker 320 is described in detail below with reference to FIGS. 10A and 10C .
- the paper tracker 322 is software and routines for tracking camera motion with respect to the position of the paper.
- the paper tracker 322 uses the paper plane as a reference frame, and then outputs information indicating the camera position relative to the paper plane. This information can then be used to translate the information from the video tracker 322 (relative or local coordinate space) to move with reference to the paper (a global coordinate space).
- the paper tracker 322 is adapted for communication with the feature retrieval unit 314 to receive feature descriptors and has an output to provide paper position information. The operation of the paper tracker 322 is described in detail below with reference to FIGS. 10A and 10C .
- the recognizer 324 is software and routines for recognize the document page, location and viewing area, the current video frame, and for initializing the link between paper and the video frame.
- the recognizer 324 is adapted for communication with the feature retrieval unit 314 , the video tracker and the paper tracker 322 .
- the recognizer 324 provides the current video frame and links between paper and the video frame to the paper tracker 322 .
- the operation of the recognizer 324 is described in detail below with reference to FIGS. 10A and 10C .
- the drift corrector 326 is software and routines for correcting the link between paper and the video frame.
- the drift corrector 326 is adapted for communication with the paper tracker 322 and sends signal to make sure the document page, location and viewing area are properly aligned with the paper.
- the drift corrector 326 provides this information the paper tracker 322 .
- the operation of the paper tracker 322 is described in detail below with reference to FIGS. 10A and 10C .
- the user interface module 328 is software and routines for generating the user interfaces shown and described below with reference to FIGS. 10A-10K and 15 - 19 H.
- the user interface module 328 works with the video tracker 320 , the paper tracker 322 , the recognizer 324 and the drift corrector 326 to generate the visual feedback to the user showing the view region and the camera look-at point.
- the user interface module 328 works with these components to determine the camera position and uses changes in the camera position as a control mechanism for interfacing the with three-dimensional models and for providing three-dimensional navigation. The operation of the user interface module 328 in this other embodiment is described in more detail below with reference to FIGS. 15 through 19H .
- the feature extraction and recognition system 108 includes two operational modes: a first mode in which the original electronic document is rendered and input and a feature index is stored for the invisible junction features from the input; and a second mode in which a image patch is input, and a feature description is created and used to retrieve a corresponding document, point and viewing region are output.
- first mode in which the original electronic document is rendered and input and a feature index is stored for the invisible junction features from the input
- a second mode in which a image patch is input, and a feature description is created and used to retrieve a corresponding document, point and viewing region are output.
- This second embodiment of the system 108 includes the feature extraction unit 310 , the feature indexing unit 312 , the feature retrieval unit 314 and the geometric estimation unit 316 .
- FIG. 4 illustrates the flow of information between the feature extraction unit 310 , the feature indexing unit 312 , the feature retrieval unit 314 and the geometric estimation unit 316 .
- the two modes of operation are show with distinct couplings 420 , 430 .
- the first mode is shown with the path of double lines 420 in which the original electronic document 402 is rendered and input via signal line 420 to the feature extraction unit 310 .
- the feature extraction unit 310 outputs feature descriptions for the invisible junctions in the document 402 to the feature indexing unit 312 via signal line 402 and the feature indexing unit 312 creates indexes and stores them in the database 110 .
- the second mode is shown with the path of signal lines 430 in which a patch of video data or a video frame 404 is input via signal line 430 to the feature extraction unit 310 .
- the feature extraction unit 310 outputs feature descriptions for the invisible junctions in the video frame 404 to the feature retrieval unit 314 via signal line 430 .
- the feature retrieval unit 314 queries the database 110 via signal line 430 using the feature descriptions and outputs a list of documents having matching features to the geometric estimation unit 316 via signal line 430 .
- the geometric estimation unit 316 applies geometric constraints to select a matching document, point and viewing region 406 .
- the present invention has created and utilizes a novel set of features for document patch recognition.
- the feature points used for document representation and retrieval are the junctions of the intrinsic skeleton existing in document pages. Since these junctions are not visible, i.e., there are in the blank space of the text document or regions with homogeneous colors, we refer to them here as “invisible junctions” or “invisible junction features”. There are four key components in an invisible junction feature:
- the feature content will be determined by all of these key factors, it is critical to have them stable and repeatable.
- the descriptor feature description must also be discriminative.
- the process begins by receiving 602 a document.
- the feature extraction unit 310 receives 602 an ideal model of an electronic document 402 .
- the feature extraction unit 310 receives 602 a captured image of a printed document such as a video frame or web camera image 404 .
- the process determines 604 a location for one or more invisible junctions.
- the feature extraction unit 310 detects features at the junctions of the text document skeleton. As shown in FIGS. 5A and 5B for this embodiment, the feature extraction unit 310 binarizes the input image, and then applies a distance transform to the binary image.
- An example input image 501 , the binarization 502 of the input image 501 and the distance transform 503 of the binarization 502 are shown in FIG. 5A .
- a skeleton 504 is detected as boundary of the label fields from the distance transformation.
- the skeleton 504 alternatively can be the ridge lines of the distance field. In one embodiment, ridge lines are detected as local maxima in one dimension. For the example input of FIG.
- the distance transform 503 and skeleton 504 are shown in FIG. 5B .
- the features are detected as the junction points 506 as shown in FIG. 5B .
- the junction points 506 alternatively can be the local maxima (in 2D) of the distance transform fields, or the local maxima on the skeleton 504 .
- the use of invisible junctions is particularly advantageous because they are stable and repeatable.
- the text document contains mostly high frequency image content in the text, which are easily contaminated and hard to filter out without degrading the text.
- the white space in the text page is mostly stable and contains mostly DC components in the frequency domain. Noise in the white space can be easily filtered out without affecting the feature locations.
- the distance transform field depends on the text locations which is desirable since the text content is eventually used for classification, the locations of invisible junction points 506 are robust to noise since the text around each junction are usually contaminated in similar ways.
- the process next determines 606 a junction size or scale for the invisible junction.
- the distance transformation value at each feature point 506 is intrinsic to the feature point 506 .
- the scale of the feature point 506 is thus assigned as such distance value multiplied by a factor ⁇ .
- ⁇ is selected as equal to 2 for this example.
- the content inside the circle 510 (centered at the feature point 506 with radius equal to its scale) is invariant to camera in-plane rotation and zoom (or distance to the paper document when the camera optical axis is perpendicular to the paper).
- the feature is local, i.e., the circle is small, the feature is also robust to changes in camera viewing angles with respect to the paper (out-of-plane rotation).
- the scale of the invisible junction is stable and repeatable. The scale is computed based on the skeleton, and is therefore stable. However, compared to location and orientation, scale is mostly subject to the changes of the nearby texts. In one embodiment, this effect is reduced by applying a weight that is inversely-proportional to the distance value when computing the feature descriptors.
- each feature point 506 is assigned a local orientation.
- a 2D local coordinate system for each feature point 506 is then determined with the feature location as the origin and its local orientation as the x-axis.
- the feature descriptor is computed based on this local coordinate system.
- the feature extraction unit 310 calculates a circle 511 centered at the feature point 506 , with a radius determined by the feature scale. This circle 511 intersects the skeleton at several positions. The two positions that are furthest apart (from among all pairs of positions on the circle that intersect the skeleton) determine the local orientation of this feature.
- FIG. 5D illustrates an example of estimation of feature orientation.
- the circle 511 centered at the feature point 506 with radius equal to feature scale intersects the skeleton 504 at three points A, B, and C.
- the two points A and B has largest distance, the feature orientation is therefore assigned as the direction of the line segment AB.
- the feature extraction unit 310 uses a more global orientation. Specifically, feature points with local orientations vote for a global orientation for the query image. The voting is done by counting the number of local feature orientations in 360 bins (one degree for each bin).
- the dominant mode of the orientation histogram is the selected as the global orientation. After voting, every feature point is re-assigned its orientation to be consistent with the selected global orientation.
- the feature extraction unit 310 uses a soft split scheme when counting the numbers for the histogram to ensure that the histogram will smoothly change when disturbed by noise.
- the global orientation is stable since there is indeed a dominant orientation, which is also true for most other non-text documents.
- the present invention can also be adjusted to achieve a tradeoff between pure local orientation and pure global orientation.
- the feature extraction unit 310 can aggregate orientations from its neighbors (as compared to from all feature points in the same image) to determine its orientation. This will resolve the issue that in some languages, such as Chinese or Japanese, the document page could contain both vertical and horizontal text lines.
- the orientation of a junction is determined by this skeleton which is stable as discussed above, and is therefore also stable.
- This is in contrast to the prior art SIFT, where the local orientation is determined by gradients of pixels around the feature point. It is well known that gradient operation amplifies noise, and therefore makes the feature orientation sensitive to noise.
- Our local orientation is derived from the skeleton and is much more stable.
- the global voting step makes the feature orientation even more stable, as it utilizes the fact that the camera is looking at a planar page, and also the fact that such page has some dominant orientation in its invisible skeleton.
- the process creates 610 a descriptor or feature description for the invisible junction.
- the feature extraction unit 310 assigns a 2D coordinate frame 516 for each feature point 506 , and computes its descriptor 518 inside such coordinate frame 516 .
- the feature extraction unit 310 computes a square with edge length equal to two times the feature scale. This square is divided into 4 ⁇ 4 cells.
- the feature extraction unit 310 can use concentric circles to construct the cells. Inside each cell, the feature extraction unit 310 accumulates a weighted gradient orientation histogram that has 8 bins (i.e., 8 directions). The gradient can be calculated by several techniques including the well known Roberts or Sobel operators.
- the feature descriptor is the concatenation of such orientation histograms, resulting in a 128-dimensional feature vector. This feature vector is normalized to accommodate linear lighting changes.
- the feature extraction unit 310 applies dimension reduction scheme using Principal Component Analysis (PCA), and reserves the first 14 dimensions. While this might reduce the discriminative power of the features, it helps reduce the noise. For this embodiment, since the cells nearby the center are mostly blank, the dimension reduction is more sensible. This advantageously produces features that are intrinsically low-dimensional, considering that for each class there is only one prototype in our case.
- the descriptor of the present invention is particularly advantageous because it is discriminative and repeatable. Repeatability is a result of the repeatability of feature location, orientation and scale. As can be seen, the content of each feature point contains pixels from three or more words in the text document.
- the process determines 612 whether there are more invisible junctions in the document. If so, the method selects an unprocessed junction and continues in step 604 to process it. If there are no more invisible junctions, the processing of the document is complete and ends.
- the feature indexing unit 312 generates one or more indices of invisible junction feature descriptors and stores them in the database 110 . For each model image, the feature indexing unit 312 extracts the feature points and stores them in the database 110 . The feature indexing unit 312 also stores metadata associated with each feature point, including location (x, y), orientation, scale, feature descriptor and a unique ID of the document page that contains this feature point. In order to allow fast real time matching and retrieval, the feature indexing unit 312 also indexes this data for efficient storage in and retrieval from the database 110 .
- the process begins by creating 702 a quantization tree.
- the quantization tree is trained by a set of training features using hierarchical clustering.
- the training set is recursively divided into K clusters.
- K L leaf nodes i.e., the vocabulary size is K L .
- the feature indexing unit 312 uses k-means algorithm with multiple random initializations. At each node of the tree, the feature indexing unit 312 records the cluster center given by the points belonging to this node.
- Each node of the quantization tree is also associated with a node, using term-frequency-inverse-document-frequency.
- the quantization tree partitions the feature space in a hierarchical way. Once the quantization tree is created, it is stored 704 in the database 110 . Next, a feature point is retrieved or received 706 from the feature extraction unit 310 , and is pushed down the tree and added 708 to the database. At each level of the tree, the feature point chooses the branch with a cluster center closest to the feature point. Once the feature point reaches the leaf node, an entry for this feature point is inserted into the hit list of this leaf node, recording the page ID and coordinates (x, y) of this feature point.
- the feature indexing unit 312 does not store the original feature descriptors. Once the feature indexing unit 312 quantizes the feature space, it can approximate the feature descriptor by its index of its residing cell in the quantization space. For each feature, the feature indexing unit 312 only stores the document page ID that contains this feature, as well as the coordinate of the feature inside the document page. The storage space requires is thereby dramatically reduced.
- the feature indexing unit 312 uses use the traditional k-d tree to organize 722 the feature points.
- the traditional k-d tree can be modified to extend the k-d tree for fast disk I/O.
- the k-d tree partitions the feature space into cells where each cell contains one or more feature points (actually pointers to feature points). During k-nearest neighbor search, it is often required to back-track the tree.
- the feature indexing unit 312 uses a depth-first linearization of the tree to order the feature points; in such a way that memory/disk locality during back-tracking is preserved.
- the cache miss rate is reduced (cache could be the CPU cache if the features reside in the main memory, or memory cache if the features reside in external hard disk drive), and the speed performance be increased.
- the feature indexing unit 312 uses PCA to reduce the feature dimension 724 , or more exactly, to find the useful dimensions. Even though our data is intrinsically low dimensional, the low-dimensional space needs to be estimated so that the result k-d tree is well-balanced. The significance of such dimension reduction is not only to meet the low-dimension requirement, but also to order the feature dimensions in such a way that the space partition is most effective.
- the space cut is ordered 726 in such a way that the dimension with most energy will be cut first, so that the system 108 can achieve a k-d tree that is not only balanced, but also with cell boundaries that are stable (i.e., small noises will not disturb which cell should the feature belong to).
- FIG. 8 an embodiment of a method for feature retrieval in accordance with the present invention is shown.
- the feature retrieval unit 314 sends the feature points detected in the query image down to the quantization, and collects the hit lists at the leaf nodes. This produces a table of candidate document pages, along with a list of coordinates (x, y) for each candidate page. This can be send to the geometric estimation unit for further processing.
- the retrieval process begins with the feature retrieval unit 314 receiving 802 a feature descriptor from the feature extraction unit 310 .
- the feature retrieval unit 314 performs pattern recognition based on local features.
- the feature retrieval unit 314 searches for the page document whose feature points best corresponded to the feature points of the given query image patch. To establish the correspondence (in other words to recognize), the feature retrieval unit 314 recognizes each feature point in the given image. More specifically, the feature retrieval unit 314 searches 804 the database 110 for the k nearest neighbors (kNN) of a feature point from the query image.
- kNN k nearest neighbors
- the collection of all such points from the database is then divided 806 into sets where points in one set share the same document page ID (recorded as meta-data with the feature point in the database). So for each candidate document page, the feature retrieval unit 314 produces a set of feature correspondences. Note that many of such correspondences are outliers, since for each query point there is at most one true correspondence in its k nearest neighbors.
- On alternate method also monitors the distance between the query feature and the current neighbor. Once there is a big jump in the distance, the feature retrieval unit 314 returns k ⁇ 10 neighbors as the following neighbors are probably outliers (i.e., the true correspondence is probably already in the current k ⁇ 10 points).
- the feature retrieval unit 314 may include scoring using tree-intersection. For each query image, the feature retrieval unit 314 includes a tree representation given the paths that the query points have gone through. The feature retrieval unit 314 constructs a similar tree for each candidate page. The similarity of the query image and a model image is then computed based on the intersection of the two trees. Thereafter the tree intersections can be ranking.
- FIG. 9A is a flow chart of a first embodiment of a method for geometric estimation in accordance with the present invention.
- FIG. 9A is a flow chart of a first embodiment of a method for geometric estimation in accordance with the present invention.
- FIG. 9B shows the steps of method
- FIG. 9D shows the imaging plane 956 , the paper plane 952 and the transformation between them.
- the method begins by receiving the candidate page identifications have been identified by the feature retrieval unit 314 , and then the geometric estimation unit 316 sorts 910 the candidate document pages based on the number of feature points in each candidate document page.
- the geometric estimation unit 316 basically uses a geometric constraint to pick the document that contains the largest number of feature points consistent with that geometric constraint. As shown in FIG.
- an example geometric constraint comes from the fact that the document page is a plane 952 ; thus, there is a perspective transformation between the paper plane 952 and the camera imaging plane 956 . Since perspective transformation involves 8 independent parameters, and is a non-linear transformation, its estimation is not stable and more expensive before outliers are removed.
- the geometric estimation unit 316 therefore removes 912 the outliers with an estimated affine transformation.
- a RANdom SAmple Consensus (RANSAC) algorithm is used to estimate the affine transformation (a realistic approximation of perspective transformation) and to detect the outliers in the feature correspondences.
- the geometric estimation unit 316 also performs a validity checking on the affine transformation estimated by RANSAC.
- an affine transformation is a generic full-rank 3 ⁇ 3 matrix (with last row being [0, 0, 1]), which contains transformations that are physically invalid for our configuration.
- the geometric estimation unit 316 enforces the constraint such that the rotation, scale, skew (the ratio ⁇ 1 / ⁇ 2 ) and translation (the third column of affine transformation H) must be bound. For example, it is impossible for the translation to be out of the scope of the document page.
- the geometric estimation unit 316 refines 914 the affine transformation using the inliers to a projective transformation.
- the geometric estimation unit 316 uses non-linear optimization (such as Levenberg-Marquardt optimization) with the initialization by the affine transformation to identify the best candidate documents. Since the projective transformation (as well as affine transformation) also gives an estimation of the camera orientation and image scale with respect to the paper, the geometric estimation unit 316 enforces 916 this orientation and scale to be consistent to the orientation and scale voted by features at the feature detection stage. This selection 916 of the best candidate using the geometric constraints from feature extraction effectively reduces the false positive, in some case essentially to zero.
- the steps of refining 914 the affine transformation using the inliers to a projective transformation can be terminated once a close match is found. If the geometric estimation unit 316 identifies a document page that results in a valid geometric transformation with enough inlier support, it can safely return the identified document page, without examining other candidate pages. This further increases the performance of the recognition method of the present invention without significant affects on accuracy.
- FIG. 9C is a flow chart of a third embodiment of a method for geometric estimation in accordance with the present invention.
- This embodiment illustrates the use of two separate geometric estimations, and their operation in parallel for increased performance.
- the process begins by receiving the candidate page identifications that have been identified by the feature retrieval unit 314 , and then sorts 920 the candidate document pages based on the number of feature points in each candidate document page. The sorted list of candidate pages is then provided for two separate geometric estimations.
- One geometric estimation ranks 922 the candidate pages using feature detection with geometric constraints. For example, the rotation or orientation and scale attributes of feature detection are enforced in this geometric estimation.
- a second geometric estimation ranks 924 the candidate pages using a projective transformation with geometric constraints such as discussed above with reference to FIG. 9B .
- the camera viewing region and look at point attributes in the projective transform are enforced in this second geometric estimation.
- these two geometric estimations may be performed in parallel or serially.
- the two page rankings are then verified or reconciled 926 .
- the verification or validation requires that the features be physically meaningful and that they are consistent with the scale and orientation estimated by the feature extraction unit 310 .
- the geometric estimation unit 316 outputs the document identification, page, point and viewing region.
- the present invention does not need to issue a fresh query for each video frame. Instead, the present invention utilizes the previous recognition results for inferring the ID of current frame as well as the camera look-at point on the paper. In order to do this, the present invention tracks the camera to infer its motion with respect to the document page being imaged. Directly estimating the camera 3D motion between two video frames is not reliable, since there are several ambiguities among the 3D camera motion parameters (e.g., out-of-plane rotation vs. translation, zooming in/out vs. moving to/from the paper).
- 3D camera motion parameters e.g., out-of-plane rotation vs. translation, zooming in/out vs. moving to/from the paper).
- the present invention directly tracks the 2D camera look-at point on the paper and infers the camera motion state utilizing temporal redundancy. Tracking 2D camera look-at point is realized by tracking feature points across video frames, and then estimating planar motions (projective transformation) between video frames. Furthermore, in addition to using previous results for the recognition of current frame, by tracking the present invention also accumulates information from multiple frames for a fresh recognition, when previous recognition results are not available.
- the process begins by recognizing 1002 a document page.
- This initial recognition step 1002 can be performed using any one of the recognition methods described above with reference to FIGS. 9A-9C .
- the feature extraction and recognition system 108 identifies the document page, it tracks the camera look-at point by registering the query image and the result document page. In other words, the feature extraction and recognition system 108 first tries to “recognize” the query image using the previously retrieved page. Furthermore, the feature extraction and recognition system 108 combines frame tracking and recognition. Recognition acts as initialization and the drift corrector 326 tracks continuous video frames.
- the process of the present invention estimates 1004 the mode of the camera.
- the feature extraction and recognition system 108 maintains a state machine for determining when to track, to recognize or to correct the drift for tracking.
- the state machine is illustrated in FIG. 10C and described more detail below.
- the changes in the mode of the camera cause transition between states.
- the feature extraction and recognition system 108 estimates the dynamics mode of the video camera and is defined to have the following states:
- the feature extraction and recognition system 108 determines 1006 the current state of tracking.
- the present invention provides four different states of tracking in which either the video tracker 320 , the paper tracker 322 , the recognizer 324 or the drift coordinator 326 are active. Depending on the current state of tracking and the camera mode, the process transitions between the video tracking state, recognizing state, the paper tracking state or the drift correcting state.
- the method of the present invention processes 1008 a video frame based on the estimated camera mode and the determined state.
- the information provided by the video tracker 320 , the paper tracker 322 , the recognizer 324 or the drift coordinator 326 can be used to generate or modify 1010 a graphical user interface.
- This graphical user interface can be combined with a representation of the electronic document.
- the graphical user interface includes a plurality of visual indications 1022 , 1024 of viewing regions, a visual indication 1026 of a transition path generated from the from each frames captured by the image capture device 106 .
- These components 1022 , 1024 , 1026 of the graphical user interface can be overlaid upon an image of the recognized electronic document 1020 .
- a graphical representation of an example user interface generated by the method for combined recognizing and tracking in accordance with the present invention is shown in FIG. 10B .
- step of modifying 1010 the graphical user interface is shown in FIG. 10A with dashed lines indicating that this step is optional.
- the information generated in recognizing and tracking could be used for variety of different applications including a user input control device or an annotation tool for use with a desktop computer, and those skilled in the art will recognize that generating and presenting a graphical user interface is just one application for the combined recognition and tracking method of the present invention.
- the method of the present invention determines 1012 whether there are additional input image frames to process. If so the method retrieves the next image frame and returns to step 1004 , and performs steps 1004 , 1006 , 1008 and 1010 on the next image frame. If there are no additional input image frames to process, the method is complete and ends.
- FIG. 10C a state diagram of the process used for recognizing and tracking in accordance with the present invention is shown.
- the process begins in state one 1052 , where the video tracker 320 tracks the camera motion (as a projective transformation) across video frames with the first video frame being the reference frame.
- the process of the present invention remains in state one 1052 .
- the recognizer 324 is operational. The recognizer 324 recognizes the current video frame and initializes a link between paper and the input video frame. If recognition fails the process of the present invention returns to state one 1052 .
- state three 1056 if recognition succeeds the process of the present invention transitions to state three 1056 because a link between the paper and video frame has been established and the following frames will be processed by the paper tracker 322 .
- the paper tracker 322 tracks the camera motion with respect to the paper with the paper plane acting as a reference frame.
- the paper tracker 322 also operates a periodic timer. Once the periodic timer expires, the process of the present invention transitions from state three 1056 to state four 1058 .
- state four 1058 drift coordinator 326 provides drift correction of the paper tracker 322 . This ensures that the alignment between the image and the paper is maintained.
- the process of the present invention returns to state three 1056 and the following frames are processed by the paper tracker 322 .
- the method of the present invention remains in state three 1056 with periodic transitions to and from state four 1058 for drift correction.
- the method of the present invention returns to state one 1052 and the following frames are processed by the video tracker 320 .
- a feature point appears multiple times in a video clip.
- the present invention associates these examples of a same feature point, and uses them together to retrieve the kNN of such feature point.
- the present invention can perform super-resolution by integration over time, or even simply choose the instance with best resolution or viewing angle for retrieval.
- many of the feature points might already have their ID and therefore k-nearest neighbors from previous frames.
- the present invention is able to eliminate many kNN retrievals by simply tracking feature points across video frames.
- FIGS. 10D-10K graphical representations for an example user interface generated by the combined recognizing and tracking in accordance with the present invention are shown.
- the user interfaces provided are for a portable computing device such as a smart phone.
- this display format is only used by way of example and that the components of the graphical user interface may be used in a variety of other configurations.
- the particular color, format, shape and appearance of the visual representations are provided only by way of example and any number of other visual representations may be used.
- like reference numerals are used to indicate like components with the same or similar functionality.
- FIG. 10D illustrates an example graphic user interface (GUI) including a window 1050 having a label 1052 with a conventional format.
- GUI graphic user interface
- Display area 1054 for showing the recognition and tracking GUI.
- Display area 1054 includes a selectable start button 1056 and a button 1058 for closing the window of the display area 1054 .
- the start button 1056 has been selected, the GUI is updated to the display shown in FIG. 10E .
- FIG. 10E illustrates the GUI with the display area 1054 showing the first input image.
- the GUI displays the input query image that is being captured by the camera or input device 106 .
- the method transition to showing the GUI of FIG. 10F .
- FIG. 10F illustrates the GUI with the display area 1054 after recognition has succeeded. Once recognition has succeeded, the GUI shows the full pathname 1060 of the electronic document recognized by the system 108 in the display area 1054 .
- the button 056 in the GUI is also modified to be a stop button 1056 . After the GUI of FIG. 10F has been displayed the method transitions to showing the GUI of FIG. 10G .
- FIG. 10G illustrates the GUI showing a representation of the original electronic document in the display area 1054 . Overlaid on top of this representation is a camera viewing window 1062 and a plurality of camera-look-at points 1064 as recognized by the present invention. This is the interface shown during operation of the paper tracker 322 . This view is particularly advantageous because of the very specific alignment of a point on the paper with a point in the electronic document, and especially the visual feedback provided to the user indicating that relationship.
- FIG. 10H illustrates the GUI showing a blurry image that cannot usually be recognized in the display area 1054 .
- FIG. 10I illustrates the GUI including a representation of the electronic document shown in the display area 1054 overlaid with a viewing window 1062 and camera-look-at points 1064 .
- the camera-look-at points 1064 , 1066 in this embodiment are represented by green squares. The positions of these green squares 1064 , 1066 show the trajectory of the camera tracking results.
- the present invention advantageously makes an image that was initially unrecognizable, recognizable after a small matter of time because the number of frames with redundant data that are processed. For example, within a matter of a few seconds, and initially unrecognizable image can be transformed into the recognizable image and the GUI created with the visual representations of the present invention as shown in FIG. 10I .
- FIG. 10J illustrates the GUI showing another example of an initially unrecognizable image.
- This example illustrates how the recognition and tracking method of the present invention is particularly advantageous for images that contain both text content and graphical content.
- the initial unrecognizable image contains little text for performing feature extraction and feature retrieval.
- FIG. 10K a GUI shows the results of combined recognition and tracking. While the viewing region 1062 was not initially recognizable, once the image frame has been recognized, the tracking of the present invention allows the GUI to show the tracking of the look-at-camera points. The tracking resolves the ambiguity issue as the input image has many hits in the database but tracking provides context cue so that the final output is unique.
- the relative movement of the camera is captured images and provides modifications to the position of the viewing window 1062 as shown by the trajectory of the camera look at points 1064 , 1066 .
- the viewing window moves from a point including a significant amount of text to a previously unrecognizable point having little text as shown in FIG. 10K , the point with little text becomes recognizable. This is particularly advantageous for documents that mix text and graphic images.
- the feature extraction and recognition system 108 can be integrated into the multi-function peripheral 1100 along with other modules for security and annotation to provide a unique set of applications. While this integrated embodiment of the present invention will be described below in the context of a multi-function peripheral 1100 , those skilled in the art will recognize that this multifunction peripheral 1100 could have any number of configurations from a low end multi-function desktop printer to a high end, commercial grade, multifunction photocopier, or any configuration in between.
- the multifunction peripheral 1100 comprises: a scanner 1102 , a printer 1104 , a facsimile module 1106 , a processor/controller 1108 , an input/output device 1110 , a database 1112 , the feature extraction and recognition system 108 , a communication device 216 , a database updating unit 1114 , a security module 1116 , and an annotation module 1118 .
- a scanner 1102 a printer 1104
- a facsimile module 1106 e.g., a facsimile module 1106
- a processor/controller 1108 e.g., a facsimile module 1106
- an input/output device 1110 e.g., a printer 1104
- the feature extraction and recognition system 108 e.g., a communication device 216 , a communication device 216 , a database updating unit 1114 , a security module 1116 , and an annotation module 1118 .
- Like reference numerals have been used
- the scanner 1102 , the printer 1104 , the facsimile module 1106 , the processor/controller 1108 , the input/output device 1110 , the communication device 216 and the database 1112 are similar in functionality to conventional components of a high-end commercially available photocopier; and therefore will not be discussed extensively here.
- the scanner 1102 is any of a variety of conventional image capture devices.
- the printer 1104 is any one of a variety of conventional print engines and paper handling mechanisms.
- the facsimile module 1106 is conventional hardware or software and cooperates with the processor/controller 1108 and the communication device 216 to receive and send facsimile messages.
- the processor/controller 1108 is similar to the processor 202 described above with reference to FIG. 2 .
- the input/output device 1110 includes the functionality of devices 210 , 212 and 214 described above with reference to FIG. 2 , and in one embodiment, includes a keypad input and display output capabilities found in a photocopier.
- the communication device 216 has been described above with reverence to FIG. 2 .
- the communication device 216 couples the multifunction peripheral 1100 to a network (not shown) via signal line 1052 .
- FIG. 11 illustrates two alternate embodiments for the database 110 , 1112 .
- the database 110 , 1112 represents non-volatile data storage, and preferably stores data including electronic documents, feature indexes and feature descriptors as has been previously described above.
- the database 110 , 1112 stores mixed-media reality (MMR) documents as described in co-pending U.S. patent application Ser. No. 11/461,017, filed Jul. 31, 2006, entitled “System and Methods for Creation and Use of a Mixed Media Environment” which is herein incorporated in its entirety by reference.
- MMR mixed-media reality
- the database 1112 is included within the multifunction peripheral 1100 .
- the database 110 is coupled to the multifunction peripheral 1100 by signal line 1052 and a communication device 216 and is external to the multifunction peripheral 1100 .
- the database 110 , 1112 also stores a variety of other information generated by the feature extraction and recognition system 108 , the database updating unit 1114 , the security module 1116 and the annotation module 1118 .
- the invention will be described as using database 1112 , however it should be understood that either database 1112 or database 110 can be utilized for those applications.
- the feature extraction and recognition system 108 as the functionality that has been described above.
- the feature extraction recognition system 108 is also adapted to be controlled by the processor/controller 1108 and cooperate with the database updating unit 1114 , the security module 1116 and the annotation module 1118 as will be described below with reference to FIGS. 12 , 13 and 14 .
- the database updating unit 1114 is software and routines for operating on the processor/controller 1108 to provide the functionality that will be described below with reference to FIG. 12 .
- the database updating unit 1114 is adapted for communication with the feature extraction and recognition system 108 , the scanner 1102 , the input/output device 1110 , the database 1112 and a processor/controller 1108 .
- the database updating unit 1114 utilizes the capabilities of the feature extraction and recognition system 108 to identify scanned documents and update entries in the database 1112 .
- the database updating unit 1114 receives and stores information in the database 1112 , uses the feature extraction and recognition system 108 to determine if there is a corresponding electronic document stored in the database 1112 , and creates an association between the received information and the recognized electronic document in the database 112 .
- the security module 1116 is software and routines for operating on the processor/controller 1108 to provide document security. More specifically, the security module 1116 controls whether the functionality of the multifunction peripheral device 1100 can be used to copy or transmit a document that the user has scanned.
- the security module 1116 is adapted for communication with the scanner 1102 , the processor/controller 1108 , the input/output device 1110 , the database 1112 and the feature extraction recognition system 108 .
- the security module 1116 determines from an input image whether there is a corresponding electronic document stored in the database 1112 and whether that document has a security protocol associated with it. If so the security module 1116 executes routines on the processor/controller 1108 to implement the associated security protocol. The operation of the security module 1116 is described in more detail with reference to FIG. 13 below.
- the annotation module 1118 is software entertains for operating on the processor/controller 1108 to provide annotations to electronic document stored in the database 1112 . It should be understood that while the annotation module 1118 is described here in a context of being part of the multi-function peripheral 1100 , the annotation module 1118 could be operational on other devices such as a cellular telephone, a personal computer, or a server associated with the database 110 .
- the annotation module 1118 is adapted for communication with the scanner 1102 , the processor/controller 1108 , the input/output device 1110 , the feature extraction recognition system 108 and the database 1112 . The operation of the annotation module 1118 is described in more detail below with reference to FIG. 14 .
- the process begins by scanning 1202 an image of a document or a portion of the document.
- the multifunction peripheral 1100 performs feature extraction and retrieval 1204 using the feature extraction and recognition system 108 . As has been described above, this produces information about a recognized electronic document or an indication that recognition is not possible.
- the method determines 1206 whether feature recognition identified a matching electronic document. If the feature extraction and recognition system 108 was unable to identify a matching electronic document, the process continues and processes 1208 the input scanned image as it normally would.
- this normal processing could include making a copy, sending a facsimile message, e-mailing a scanned image or any one of the other conventional functions performed by the multifunction peripheral 1100 .
- this normal processing could include disabling the functions of the multifunction peripheral 1100 and displaying a message to the user on the input/output device 1110 indicating that the scanned image does not have an electronic equivalent and/or could not be recognized.
- the process continues by determining 1210 a security procedure associated with the scanned image.
- the database 1112 includes a security procedure for each image indexed and stored in the database 1112 . Once the associated security procedure has been determined 1210 , the security procedure is retrieved and the scanned image is processed 1212 using the retrieved security procedure.
- security procedures may be implemented by the multifunction peripheral 1100 in step 1212 . The following security procedures are provided by way of example and not by way of limitation as processes that the multifunction peripheral 1100 may implement. Furthermore, each of the security procedures may be combined with others for increased security and functionality.
- the security module 1116 causes the input/output device 1110 to output a request for a user's password.
- the input output device 1110 receives the password and confirms it against that corresponding to the user's ID such as may have been input to access the functionality of the multifunction peripheral 1100 . If the input password matches information stored in the database 1112 the processor/controller 1108 causes a printout of the scanned image to be produced by the printer 1104 . On the other hand if the input password does not match, the multifunction peripheral 1100 will not output any documents. This ensures that only an individual with the password can make a copy of the scanned document. Alternatively, the user's access code may be used instead of a password. In such a case, only those users authorized to copy a document will be allowed to do so. The user's identity will be verified by the access code used to gain access to the multi-function peripheral 1100 .
- the security module 1116 stores a record in the database 1112 indicating that the recognized document was printed and outputs a copy. Under this security procedure, the security module 1116 may record the time, number of copies, method of output, whether the image or the electronic document was used to generate the output and a recipient that were used. If the multifunction peripheral 1100 requires that the user enter an access code, the access code could also be recorded in the database 1112 .
- the security module 1116 inhibits or presents a copy from being output by the multifunction peripheral 1100 .
- the document may be extremely restricted in access and may not be allowed to be copied.
- the security module 1116 automatically enforces that procedure.
- the security module 1116 In a fourth security procedure, the security module 1116 generates a notification and sends the notification such as by facsimile or e-mail to a manager or administrator.
- the notification could indicate that the recognized document has been copied. Alternatively, the notification could indicate that an attempt to copy the document was made.
- the security module cooperates with the processor/controller 1108 to generate the message and the facsimile module 1106 is sent by facsimile or the communication device to 16 is sent electronically such as by e-mail or instant messaging.
- the security module 1116 operates in a cooperative environment in which multiple users access and annotates a single document.
- the security module 1116 cooperates with the processor/controller 1108 to determine whether the input image corresponds to: 1) a document that has been updated since the input scanned image was produced, 2) a page or portion of a much longer document, 3) an original electronic document stored in the database 1112 or 4) a previously copied or printed version of the document.
- the security module 1116 can output a different document than the document scanned. For example, the updated version of the document could be output, the entire document to be output, a perfect copy generated from the old electronic original could be output or the previously copied or printed version could be output.
- the security procedures described above were primarily discussed in the context of copying, those skilled in the art will recognize that the security procedures may be applied to any function performed by the multifunction peripheral 1100 .
- the security procedures may also be applied to facsimile transmission of the input image by the multifunction peripheral 1100 or electronic transmission such as e-mail or instant messaging of the input image by the multifunction peripheral 1100 .
- FIG. 13 a method for using invisible junction recognition as a part of database updating and document processing in accordance with the present invention is described. While the present invention will be described with reference to mixed-media reality (MMR), those skilled in the art should recognize that the method described here is applicable to any type of updating to the database 1112 that maybe required or desirable.
- the method begins with steps similar to those described above with reference to FIG. 12 , namely scanning 1202 an image, performing feature extraction and retrieval 1204 and determining 1206 whether feature recognition identified a matching electronic document. If a matching electronic document is not recognized, the method proceeds to step 1208 in which the document is processed normally. This entails processing the input image as requested by the user using the input/output device 1110 .
- MMR mixed-media reality
- the method prompts 1302 the user for processing and updating instructions.
- the database updating unit 1114 causes the input/output device 1110 to display a request for processing and updating instructions. The user then responds using the input/output device 1110 to provide the processing and updating instructions.
- Example processing instructions include: 1) associating the scanned image with the matching electronic document; 2) adding metadata and associating it with the matching electronic document; 3) creating a new independent MMR document; 4) generating an index for the scanned image and storing it in the database 1112 ; and 5) receive/retrieve additional information and associating it with the matching electronic document.
- the method begins with steps similar to those described above with reference to FIG. 12 , namely scanning 1202 an image, performing feature extraction and retrieval 1204 and determining 1206 / 1402 whether feature recognition identified a matching electronic document.
- the invention is discussed in the context of an MMR document although it can be generalized to any electronic document.
- the determining step may be the identification of an MMR document that includes a representation of a portion of the printed document, an action or second media, an index or hotspot, and an electronic representation of the entire document. If a matching electronic document is not recognized, the method proceeds to step 1208 in which the document is processed normally. This entails processing the input image as requested by the user using the input/output device 1110 .
- the method proceeds to prompt 1404 the user for processing and updating instructions.
- the annotation module 1118 causes the input/output device 1110 to display a request for processing and updating instructions. The user then responds using the input/output device 1110 to provide the processing and updating instructions. Additionally, the user inputs any type of annotation such as text, sound, video or other information that may be captured by input/output devices 1110 coupled to the multifunction peripheral 1100 . For those embodiments where the annotation module 1118 is included as part of a cell phone, the type of annotations available would be limited to the input/output capabilities of the cell phone.
- the annotation module 1118 receives 1406 the processing instructions and an annotation from the user.
- the annotation module 1118 then updates 1408 the database 1112 according to the input user instructions.
- the annotation information is added to an MMR document.
- the present invention is particularly advantageous because in the process of performing recognition, the recognizer 324 and associated components generate information about the position of the camera imaging plane, and therefore the camera, with respect to the document or paper plane.
- This information includes camera pose (including camera position and orientation) with respective to the paper plane, which is derived from the estimated projective transformation between the camera imaging plane and the paper plane.
- the recognizer 324 estimates the camera pose with respect to the paper for every video plane.
- the camera intrinsic parameters can be estimated online, or one can pre-calibrate the camera for the intrinsic parameters.
- this information can be used as part of a user interface to provide a new way to input commands and instructions to a computing system by using both the printed document and the camera pose relative to that printed document.
- the use of the camera pose relative to the paper as a construct for performing operations and inputting commands can be part of any number of different graphical user interfaces and the two described below are provided merely by way of example. This unique construct of the present invention will now be described with reference to two different graphical user interfaces. These graphical user interface examples combine features of the electronic document identification and navigation of three-dimensional models.
- the first graphical user interface example provides electronic document identification from a paper document, and three-dimensional model retrieval with subsequent navigation semi-independent of the original document.
- the second graphical user interface example provides three-dimensional model retrieval and three-dimensional navigation of the model based upon the camera pose relative to a two-dimensional printed document.
- the three-dimensional user interface module 328 in cooperation with the other components of memory unit 204 operate to perform method shown in FIG. 15 .
- the process begins by receiving 1202 an image, performing 1204 feature extraction and retrieval and determining 1206 whether there is a matching electronic document based on comparisons of invisible junction descriptors as has been described above with reference to FIG. 12 . If no matching electronic document is found in the determining 1206 step, the method of the present invention generates a message indicating that there is no additional content associated with the input image and the method is complete and ends.
- step 1504 the method determines whether there is a three-dimensional model associated with the document page, viewing position or camera look at point. Since the feature extraction and retrieval step 1204 generates this information for the scanned image, the information can be compared to information in the database 110 for the existence of any associated three-dimensional model.
- Next method retrieves 1506 the three-dimensional model and associated data from the database 110 .
- the object model stored in database 110 can be represented by three-dimensional model, panoramic image of the object, or a sequence of pre-taken images of the object. Then the method displays 1508 the three-dimensional model and data in an initial view.
- the viewing angle of the three-dimensional object is defined by pose of the camera (e.g., cell phone) with respect to the paper. By tracking the camera pose with respect to the paper, the user can get a continuous view of the object model from different viewing angles.
- the object model could be specified by a three-dimensional geometric model plus texture maps, or a panoramic image of an object.
- FIG. 17A shows a graphic presentation of a paper document 1702 that has an example webpage with information about an automobile.
- a viewing region that is captured by the image capture device 106 is shown as a generally rectangular box and the noted as a 1704 is not present on the printed page 1702 but is shown in FIG. 17A to illustrate the relationship between the printed document 1702 and the captured image 1706 shown in FIG. 17B .
- FIG. 17B shows a graphic representation of the image 1706 captured by the image capture device 106 .
- FIG. 17B also shows the initial view 1708 of the three-dimensional model retrieved in step 1508 .
- FIG. 19A is a graphic representation of a map that has been printed on paper 1902 .
- a viewing region 1904 and a camera-looks-at point 1906 are shown to illustrate the present invention.
- FIG. 19B is a graphic representation of an image captured by image capture device 106 .
- the display of the image also includes the display of the camera-look-at point 1906 .
- FIG. 19B also shows a graphic representation of an initial image view 1910 .
- the initial image 1910 is a real world plan view of the geographic area corresponding to the geographic area represented by the portion of the map in the captured image. This initial view using the GUI of the present invention includes a camera look at point 1912 .
- the method of the present invention continues to determine whether the image capture device 106 is providing additional frames or images of data. If not, the method is complete and ends. However if additional frames or images of data are being provided, the method of the present continues in step 1512 where the next frame is processed. In step 512 , the user interface module 328 retrieves the next frame and determines the camera pose for that frame. Next method determines 1514 the change in camera position between the current frame and the next frame. Since for each frame the camera position and orientation are known, the difference between the current and next camera position can be calculated and represented as a vector.
- the display of the three-dimensional model and data are updated 1516 based on the change in position to determined in step 1514 .
- the process for modifying the display will be described in more detail below for the first example with reference to FIGS. 16 and 17 , for the second example with reference to FIGS. 18 and 19 .
- the method returns to step 1510 to determine whether additional frames or images are being received. If so, the method loops through steps 1512 , 1514 and 1516 for each frame. In an alternate embodiment, the loop may be performed for every nth frame where n is an integer greater than 2.
- step 1514 determines the change in camera pose.
- the determination of a change in camera position and orientation has six possible changes.
- the determined change in camera position may be one of the following states: 1) the camera may be rotated to the left, 2) the camera may be rotated to the right, 3) the camera may be rotated upward, 4) the camera may be rotated downward, 5) the camera may be rotated clockwise, 6) the camera may be rotated counterclockwise, and 7) no change.
- the method transitions from step 1514 to a corresponding step 1602 , 1604 , 1606 , 1608 , 1610 or 1612 .
- FIG. 17A shows the paper document 1702 as has been described above.
- FIG. 17B shows the image captured 1706 from the viewing region 1704 as well as the initial view 1708 of the three-dimensional model.
- FIG. 17C shows the paper document 1702 essentially the same as in FIG.
- the viewing region 1712 has a trapezoidal shape that reflects rotation of the camera to the right. This in turn yields the image captured 1714 in FIG. 17D and causes the display of the three-dimensional model to be modified to show a side view 1716 as shown in FIG. 17D .
- step 1602 to rotate the view of the three-dimensional model to the left.
- This causes image capture device 106 to produce the image 1718 of FIG. 17E and the user interface module 328 to display the three-dimensional model modified to show a front view 1720 shown in FIG. 17E .
- step 1606 rotate the view of the three-dimensional model to show the top of the model. This causes image capture device 106 to produce the image 1722 of FIG. 17F and the user interface module 328 to display the three-dimensional model modified to show a plan view 1724 shown in FIG. 17F .
- step 1608 rotates the view of the three-dimensional model to show a bottom perspective view of the model. This causes image capture device 106 to produce the image 1726 of FIG. 17G and the user interface module 328 to display the three-dimensional model modified to show a partial bottom view 1728 shown in FIG. 17G .
- other positional changes such as moving of the camera in a clockwise or counterclockwise direction may cause the graphical user interface to display additional data.
- the method of the present invention shows 1610 a first data set along with the current three-dimensional model being shown.
- the display could transition to the display of the first data set only.
- the method of the present invention shows 1612 its second data set along with the current three-dimensional model that was displayed. In particular, turning the camera to the right would cause price information about the vehicle to be displayed, while turning the device of laughed would cause a list of options or accessories for the vehicle to be displayed.
- the present invention may be used as an input device along with a piece of paper to provide searching, annotation, and modification commands to applications such as spreadsheets, presentation program, word processing program, and other applications that provide annotation of documents with audio, video and pictures.
- FIG. 17H shows a graphical representation of a piece of paper 1750 .
- the piece of paper 1750 includes an advertisement for an automobile.
- a viewing region 1752 and camera look at points 1754 are shown overlaid on a piece of paper in FIG. 17H .
- movement of the camera to the left, right, or up causes translation of the image displayed in the graphical user interface such as shown for three examples translate right 1756 , translate left 1758 and translate up in FIG. 17I .
- step 1514 determines the change in camera pose.
- the determination of a change in camera pose has six possible changes.
- the determined change in camera pose may be one of the following states: 1) the camera may moved up or down, 2) the camera may pan right, 3) the camera may pan left, 4) the camera may rotated clockwise or counterclockwise, 5) the camera may be moved north or south, 6) the camera may moved east or west, and 7) no change.
- the method transitions from step 1514 to a corresponding step 1802 , 1804 , 1806 , 1808 , 1810 or 1812 .
- Each of the steps 1802 , 1804 and 1806 different three-dimensional views of the terrain corresponding to the real world location represented by the map/paper. These steps pan left, pan right or show the front or rear views.
- the GUI advantageous also provides for plan views of the terrain and movement of the point of view.
- Each of the steps 1808 , 1810 and 1812 provides and adjustment of an aerial view such as rotating the aerial view, moving the aerial view up or down to cover a region or moving the aerial view left or right to cover a new region. Examples for some of these adjustments to the GUI are shown and described below with reference to FIGS. 19A-19H
- step 1802 modify the view of the three-dimensional model from a plan view to a front perspective view.
- the method continues in step 1802 to modify the view of the three-dimensional model from a plan view to a front perspective view.
- FIGS. 19A and 19B show the page, viewing region, captured image and three-dimensional model shown in FIGS. 19C and 19D .
- FIG. 19A shows the paper document 1902 as has been described above.
- FIG. 19B shows the image captured 1908 from the viewing region 1904 as well as the initial view 1910 of the three-dimensional model.
- the GUI can display a version of the map with the camera viewing region shown as a quadrangle.
- the camera look at point 1906 is shown also shown.
- the corresponding terrain data 1910 and camera look at point 1912 are also shown in FIG. 19B .
- FIG. 19C shows the paper document 1902 essentially the same as in FIG. 19A , except that the viewing region 1914 has a trapezoidal shape that reflects movement of the camera to the downward. This in turn yields the image captured 1916 in FIG. 19D and causes the display of the terrain data to be modified to show a three-dimensional model 1918 as shown in FIG. 19D .
- the three-dimensional building is visualized at different angles determined by the pose of the camera with respect to the 2D map.
- FIG. 19E also illustrates an example where the camera is moved, but the paper 1902 is also rotated.
- FIG. 19E shows the paper document 1902 rotated.
- the viewing region 1920 is overlaid on an image of the paper 1902 .
- FIG. 19F shows the image 1924 captured by the image capture device 106 for display in the GUI. By tilting down the camera and rotating the map, the user can get a view of the building from every side.
- FIG. 19F also shows a view 1926 of the three-dimensional building from the same perspective as the camera to the paper.
- FIG. 19G shows another image 1928 captured by the image capture device 106 for display in the GUI. By tilting to the left, a different view 1930 of the building is shown. Similarly, in FIG. 19H also shows the image captured 1932 and a view 1934 of the three-dimensional building again from the same perspective as the camera to the paper.
- modules, routines, features, attributes, methodologies and other aspects of the present invention can be implemented as software, hardware, firmware or any combination of the three.
- a component, an example of which is a module, of the present invention is implemented as software
- the component can be implemented as a standalone program, as part of a larger program, as a plurality of separate programs, as a statically or dynamically linked library, as a kernel loadable module, as a device driver, and/or in every and any other way known now or in the future to those of ordinary skill in the art of computer programming.
- the present invention is in no way limited to implementation in any specific programming language, or for any specific operating system or environment. Accordingly, the disclosure of the present invention is intended to be illustrative, but not limiting, of the scope of the present invention, which is set forth in the following claims.
Abstract
Description
-
- Location: the coordinate (x, y) of the feature point, with the top-left corner of the document page as the origin;
- Orientation: for setting up a local coordinate system for each feature point when we compute the description of each feature point using its neighborhood pixels;
- Scale: the size of the feature;
- Description: the content of the feature.
H 1:2,1:2 =R(α)(λ1/λ2)R(φ) (1)
In accordance with the present invention, the
-
- Hovering: the camera is mostly steady and looking at the paper document;
- Scanning: the camera is scanning the document by translating and/or rotating;
- Zooming: the camera is moving to or from the document page; and
- Abrupt change: the camera is moving to another document page.
Claims (35)
Priority Applications (24)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/827,530 US8276088B2 (en) | 2007-07-11 | 2007-07-11 | User interface for three-dimensional navigation |
JP2008180793A JP5181888B2 (en) | 2007-07-11 | 2008-07-10 | Method and system for generating a graphical user interface |
EP08252377.0A EP2015227B8 (en) | 2007-07-11 | 2008-07-11 | User interface for three-dimensional navigation |
US12/210,519 US8510283B2 (en) | 2006-07-31 | 2008-09-15 | Automatic adaption of an image recognition system to image capture devices |
US12/210,511 US8825682B2 (en) | 2006-07-31 | 2008-09-15 | Architecture for mixed media reality retrieval of locations and registration of images |
US12/210,532 US8868555B2 (en) | 2006-07-31 | 2008-09-15 | Computation of a recongnizability score (quality predictor) for image retrieval |
US12/210,540 US8856108B2 (en) | 2006-07-31 | 2008-09-15 | Combining results of image retrieval processes |
US12/240,596 US8676810B2 (en) | 2006-07-31 | 2008-09-29 | Multiple index mixed media reality recognition using unequal priority indexes |
US12/240,590 US8369655B2 (en) | 2006-07-31 | 2008-09-29 | Mixed media reality recognition using multiple specialized indexes |
US12/247,205 US9063952B2 (en) | 2006-07-31 | 2008-10-07 | Mixed media reality recognition with image tracking |
US12/247,202 US8073263B2 (en) | 2006-07-31 | 2008-10-07 | Multi-classifier selection and monitoring for MMR-based image recognition |
US12/253,715 US9176984B2 (en) | 2006-07-31 | 2008-10-17 | Mixed media reality retrieval of differentially-weighted links |
US12/253,815 US8201076B2 (en) | 2006-07-31 | 2008-10-17 | Capturing symbolic information from documents upon printing |
US12/265,502 US8489987B2 (en) | 2006-07-31 | 2008-11-05 | Monitoring and analyzing creation and usage of visual content using image and hotspot interaction |
US12/340,124 US9020966B2 (en) | 2006-07-31 | 2008-12-19 | Client device for interacting with a mixed media reality recognition system |
US12/342,330 US8156116B2 (en) | 2006-07-31 | 2008-12-23 | Dynamic presentation of targeted information in a mixed media reality recognition system |
US13/415,756 US8452780B2 (en) | 2006-01-06 | 2012-03-08 | Dynamic presentation of targeted information in a mixed media reality recognition system |
US13/494,008 US9311336B2 (en) | 2006-07-31 | 2012-06-11 | Generating and storing a printed representation of a document on a local computer upon printing |
US13/729,458 US8965145B2 (en) | 2006-07-31 | 2012-12-28 | Mixed media reality recognition using multiple specialized indexes |
US13/789,669 US9087104B2 (en) | 2006-01-06 | 2013-03-07 | Dynamic presentation of targeted information in a mixed media reality recognition system |
US13/914,417 US9870388B2 (en) | 2006-07-31 | 2013-06-10 | Analyzing usage of visual content to determine relationships indicating unsuccessful attempts to retrieve the visual content |
US14/604,619 US9495385B2 (en) | 2004-10-01 | 2015-01-23 | Mixed media reality recognition using multiple specialized indexes |
US14/746,669 US9972108B2 (en) | 2006-07-31 | 2015-06-22 | Mixed media reality recognition with image tracking |
US14/804,241 US10007928B2 (en) | 2004-10-01 | 2015-07-20 | Dynamic presentation of targeted information in a mixed media reality recognition system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/827,530 US8276088B2 (en) | 2007-07-11 | 2007-07-11 | User interface for three-dimensional navigation |
Related Parent Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/776,530 Continuation-In-Part US8184155B2 (en) | 2004-10-01 | 2007-07-11 | Recognition and tracking using invisible junctions |
US12/060,194 Continuation-In-Part US8156115B1 (en) | 2004-10-01 | 2008-03-31 | Document-based networking with mixed media reality |
US12/121,275 Continuation-In-Part US8385589B2 (en) | 2004-10-01 | 2008-05-15 | Web-based content detection in images, extraction and recognition |
Related Child Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/461,109 Continuation-In-Part US9384619B2 (en) | 2004-10-01 | 2006-07-31 | Searching media content for objects specified using identifiers |
US11/776,530 Continuation-In-Part US8184155B2 (en) | 2004-10-01 | 2007-07-11 | Recognition and tracking using invisible junctions |
US11/777,142 Continuation-In-Part US8176054B2 (en) | 2004-10-01 | 2007-07-12 | Retrieving electronic documents by converting them to synthetic text |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090019402A1 US20090019402A1 (en) | 2009-01-15 |
US8276088B2 true US8276088B2 (en) | 2012-09-25 |
Family
ID=39869729
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/827,530 Active 2029-02-08 US8276088B2 (en) | 2004-10-01 | 2007-07-11 | User interface for three-dimensional navigation |
Country Status (3)
Country | Link |
---|---|
US (1) | US8276088B2 (en) |
EP (1) | EP2015227B8 (en) |
JP (1) | JP5181888B2 (en) |
Cited By (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100306681A1 (en) * | 2003-09-30 | 2010-12-02 | Dirtt Environmental Solutions Ltd. | Creation and modification of valid functional design layouts |
US20110255782A1 (en) * | 2010-01-15 | 2011-10-20 | Copanion, Inc. | Systems and methods for automatically processing electronic documents using multiple image transformation algorithms |
US20120250987A1 (en) * | 2011-03-31 | 2012-10-04 | Sony Corporation | System and method for effectively performing an image identification procedure |
US8369655B2 (en) | 2006-07-31 | 2013-02-05 | Ricoh Co., Ltd. | Mixed media reality recognition using multiple specialized indexes |
US8385660B2 (en) | 2009-06-24 | 2013-02-26 | Ricoh Co., Ltd. | Mixed media reality indexing and retrieval for repeated content |
US8385589B2 (en) | 2008-05-15 | 2013-02-26 | Berna Erol | Web-based content detection in images, extraction and recognition |
US8452780B2 (en) | 2006-01-06 | 2013-05-28 | Ricoh Co., Ltd. | Dynamic presentation of targeted information in a mixed media reality recognition system |
US8478761B2 (en) | 2007-07-12 | 2013-07-02 | Ricoh Co., Ltd. | Retrieving electronic documents by converting them to synthetic text |
US8489987B2 (en) | 2006-07-31 | 2013-07-16 | Ricoh Co., Ltd. | Monitoring and analyzing creation and usage of visual content using image and hotspot interaction |
US8510283B2 (en) | 2006-07-31 | 2013-08-13 | Ricoh Co., Ltd. | Automatic adaption of an image recognition system to image capture devices |
US8521737B2 (en) | 2004-10-01 | 2013-08-27 | Ricoh Co., Ltd. | Method and system for multi-tier image matching in a mixed media environment |
US8600989B2 (en) | 2004-10-01 | 2013-12-03 | Ricoh Co., Ltd. | Method and system for image matching in a mixed media environment |
US8612475B2 (en) | 2011-07-27 | 2013-12-17 | Ricoh Co., Ltd. | Generating a discussion group in a social network based on metadata |
US20140043438A1 (en) * | 2009-05-01 | 2014-02-13 | Microsoft Corporation | Systems and Methods for Detecting a Tilt Angle from a Depth Image |
US8676810B2 (en) | 2006-07-31 | 2014-03-18 | Ricoh Co., Ltd. | Multiple index mixed media reality recognition using unequal priority indexes |
US8706711B2 (en) | 2011-06-22 | 2014-04-22 | Qualcomm Incorporated | Descriptor storage and searches of k-dimensional trees |
US8825682B2 (en) | 2006-07-31 | 2014-09-02 | Ricoh Co., Ltd. | Architecture for mixed media reality retrieval of locations and registration of images |
US8838591B2 (en) | 2005-08-23 | 2014-09-16 | Ricoh Co., Ltd. | Embedding hot spots in electronic documents |
US8856108B2 (en) | 2006-07-31 | 2014-10-07 | Ricoh Co., Ltd. | Combining results of image retrieval processes |
US8868555B2 (en) | 2006-07-31 | 2014-10-21 | Ricoh Co., Ltd. | Computation of a recongnizability score (quality predictor) for image retrieval |
US8949287B2 (en) | 2005-08-23 | 2015-02-03 | Ricoh Co., Ltd. | Embedding hot spots in imaged documents |
US8989431B1 (en) | 2007-07-11 | 2015-03-24 | Ricoh Co., Ltd. | Ad hoc paper-based networking with mixed media reality |
US9020966B2 (en) | 2006-07-31 | 2015-04-28 | Ricoh Co., Ltd. | Client device for interacting with a mixed media reality recognition system |
US9019568B2 (en) | 2012-03-22 | 2015-04-28 | Intuit Inc. | Accurately extracting information based on orthogonal images |
US9036925B2 (en) | 2011-04-14 | 2015-05-19 | Qualcomm Incorporated | Robust feature matching for visual search |
US9063953B2 (en) | 2004-10-01 | 2015-06-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment |
US9063952B2 (en) | 2006-07-31 | 2015-06-23 | Ricoh Co., Ltd. | Mixed media reality recognition with image tracking |
US9164653B2 (en) | 2013-03-15 | 2015-10-20 | Inspace Technologies Limited | Three-dimensional space for navigating objects connected in hierarchy |
US9171202B2 (en) | 2005-08-23 | 2015-10-27 | Ricoh Co., Ltd. | Data organization and access for mixed media document system |
US9176984B2 (en) | 2006-07-31 | 2015-11-03 | Ricoh Co., Ltd | Mixed media reality retrieval of differentially-weighted links |
US9189571B2 (en) | 2011-06-11 | 2015-11-17 | Ice Edge Business Solutions, Ltd. | Automated re-use of structural components |
US9245064B2 (en) | 2009-11-24 | 2016-01-26 | Ice Edge Business Solutions | Securely sharing design renderings over a network |
US20160026628A1 (en) * | 2014-07-22 | 2016-01-28 | Verizon Patent And Licensing Inc. | Providing content based on image item |
US9311336B2 (en) | 2006-07-31 | 2016-04-12 | Ricoh Co., Ltd. | Generating and storing a printed representation of a document on a local computer upon printing |
US9357098B2 (en) | 2005-08-23 | 2016-05-31 | Ricoh Co., Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US9355197B2 (en) | 2013-01-25 | 2016-05-31 | Dirtt Environmental Solutions, Ltd | Real-time depth of field effects within design software |
US9373029B2 (en) | 2007-07-11 | 2016-06-21 | Ricoh Co., Ltd. | Invisible junction feature recognition for document security or annotation |
US9384619B2 (en) | 2006-07-31 | 2016-07-05 | Ricoh Co., Ltd. | Searching media content for objects specified using identifiers |
US9405751B2 (en) | 2005-08-23 | 2016-08-02 | Ricoh Co., Ltd. | Database for mixed media document system |
US9519407B2 (en) | 2008-03-11 | 2016-12-13 | Ice Edge Business Solutions, Ltd. | Automatically creating and modifying furniture layouts in design software |
US9530050B1 (en) | 2007-07-11 | 2016-12-27 | Ricoh Co., Ltd. | Document annotation sharing |
US9536340B2 (en) | 2004-08-17 | 2017-01-03 | Dirtt Environmental Solutions, Ltd. | Software incorporating efficient 3-D rendering |
US9619920B2 (en) | 2013-01-31 | 2017-04-11 | Ice Edge Business Solutions, Ltd. | Method and system for efficient modeling of specular reflection |
US9747680B2 (en) | 2013-11-27 | 2017-08-29 | Industrial Technology Research Institute | Inspection apparatus, method, and computer program product for machine vision inspection |
US10922450B2 (en) | 2014-06-09 | 2021-02-16 | Dirtt Environmental Solutions, Ltd. | Associating computer-executable objects with timber frames within an architectural design environment |
US11103787B1 (en) | 2010-06-24 | 2021-08-31 | Gregory S. Rabin | System and method for generating a synthetic video stream |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8144921B2 (en) * | 2007-07-11 | 2012-03-27 | Ricoh Co., Ltd. | Information retrieval using invisible junctions and geometric constraints |
US9395190B1 (en) | 2007-05-31 | 2016-07-19 | Trx Systems, Inc. | Crowd sourced mapping with robust structural features |
US8751151B2 (en) * | 2012-06-12 | 2014-06-10 | Trx Systems, Inc. | System and method for localizing a trackee at a location and mapping the location using inertial sensor information |
US20090083669A1 (en) * | 2007-09-26 | 2009-03-26 | Autodesk, Inc. | Navigation system for a 3d virtual scene |
JP4645652B2 (en) * | 2008-01-08 | 2011-03-09 | コニカミノルタビジネステクノロジーズ株式会社 | Image forming apparatus, image forming method, and image forming program |
US8543929B1 (en) * | 2008-05-14 | 2013-09-24 | Adobe Systems Incorporated | User ratings allowing access to features for modifying content |
TW201011259A (en) * | 2008-09-12 | 2010-03-16 | Wistron Corp | Method capable of generating real-time 3D map images and navigation system thereof |
US8180766B2 (en) * | 2008-09-22 | 2012-05-15 | Microsoft Corporation | Bayesian video search reranking |
US8606774B1 (en) * | 2009-05-18 | 2013-12-10 | Google Inc. | Methods and systems for 3D shape retrieval |
US20110013014A1 (en) * | 2009-07-17 | 2011-01-20 | Sony Ericsson Mobile Communication Ab | Methods and arrangements for ascertaining a target position |
US20120271864A1 (en) * | 2009-10-29 | 2012-10-25 | Clayton Richard Morlock | Method for assisted road extrapolation from imagery |
WO2011062581A1 (en) * | 2009-11-18 | 2011-05-26 | Telenav, Inc. | Navigation system with relative ranking mechanism and method of operation thereof |
US8898008B2 (en) * | 2009-11-18 | 2014-11-25 | Telenav, Inc. | Navigation system with relative ranking mechanism and method of operation thereof |
US8639034B2 (en) * | 2010-11-19 | 2014-01-28 | Ricoh Co., Ltd. | Multimedia information retrieval system with progressive feature selection and submission |
WO2013079098A1 (en) * | 2011-11-29 | 2013-06-06 | Layar B.V. | Dynamically configuring an image processing function |
US9330323B2 (en) * | 2012-04-29 | 2016-05-03 | Hewlett-Packard Development Company, L.P. | Redigitization system and service |
US11156464B2 (en) | 2013-03-14 | 2021-10-26 | Trx Systems, Inc. | Crowd sourced mapping with robust structural features |
US11268818B2 (en) | 2013-03-14 | 2022-03-08 | Trx Systems, Inc. | Crowd sourced mapping with robust structural features |
USD823312S1 (en) * | 2014-08-11 | 2018-07-17 | Sony Corporation | Display panel or screen with graphical user interface |
US9759561B2 (en) | 2015-01-06 | 2017-09-12 | Trx Systems, Inc. | Heading constraints in a particle filter |
JP6586824B2 (en) | 2015-08-27 | 2019-10-09 | 富士通株式会社 | Image processing apparatus, image processing method, and image processing program |
JP6798379B2 (en) | 2017-03-21 | 2020-12-09 | 株式会社リコー | Information processing system and information processing method |
EP3410353A1 (en) | 2017-06-01 | 2018-12-05 | eyecandylab Corp. | Method for estimating a timestamp in a video stream and method of augmenting a video stream with information |
JP7231529B2 (en) * | 2019-11-20 | 2023-03-01 | Kddi株式会社 | Information terminal device, server and program |
US11082679B1 (en) | 2021-01-12 | 2021-08-03 | Iamchillpill Llc. | Synchronizing secondary audiovisual content based on frame transitions in streaming content |
Citations (317)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1915993A (en) | 1931-04-27 | 1933-06-27 | Gen Electric | Statistical machine |
US4759075A (en) | 1983-03-14 | 1988-07-19 | Ana Tech Corporation | Method and apparatus for vectorizing documents and symbol recognition |
US5027421A (en) | 1987-11-30 | 1991-06-25 | Kabushiki Kaisha Toshiba | Data processing apparatus having function for registering abstract image data |
US5035302A (en) | 1989-03-03 | 1991-07-30 | Otis Elevator Company | "Artificial Intelligence" based learning system predicting "Peak-Period" times for elevator dispatching |
US5077805A (en) | 1990-05-07 | 1991-12-31 | Eastman Kodak Company | Hybrid feature-based and template matching optical character recognition system |
US5109439A (en) | 1990-06-12 | 1992-04-28 | Horst Froessl | Mass document storage and retrieval system |
US5263100A (en) | 1991-07-19 | 1993-11-16 | Samsung Electronics Co., Ltd. | Image coding method and apparatus |
US5392447A (en) | 1992-01-10 | 1995-02-21 | Eastman Kodak Compay | Image-based electronic pocket organizer with integral scanning unit |
US5416892A (en) | 1990-03-07 | 1995-05-16 | Fujitsu Limited | Best first search considering difference between scores |
US5432864A (en) | 1992-10-05 | 1995-07-11 | Daozheng Lu | Identification card verification system |
US5465353A (en) | 1994-04-01 | 1995-11-07 | Ricoh Company, Ltd. | Image matching and retrieval by multi-access redundant hashing |
US5493689A (en) | 1993-03-01 | 1996-02-20 | International Business Machines Corporation | System for configuring an event driven interface including control blocks defining good loop locations in a memory which represent detection of a characteristic pattern |
US5546502A (en) | 1993-03-19 | 1996-08-13 | Ricoh Company, Ltd. | Automatic invocation of computational resources without user intervention |
US5553217A (en) | 1993-09-23 | 1996-09-03 | Ricoh Company, Ltd. | Document layout using tiling |
US5555556A (en) | 1994-09-30 | 1996-09-10 | Xerox Corporation | Method and apparatus for document segmentation by background analysis |
US5579471A (en) | 1992-11-09 | 1996-11-26 | International Business Machines Corporation | Image query system and method |
US5706097A (en) | 1995-11-13 | 1998-01-06 | Eastman Kodak Company | Index print with a digital recording medium containing still images, motion sequences, and sound sequences |
US5752055A (en) | 1994-12-27 | 1998-05-12 | International Business Machine Corp. | Systems and method for automatically linking parts within compound documents |
US5761344A (en) | 1994-04-15 | 1998-06-02 | Canon Kabushiki Kaisha | Image pre-processor for character recognition system |
US5764277A (en) | 1995-11-08 | 1998-06-09 | Bell Communications Research, Inc. | Group-of-block based video signal combining for multipoint continuous presence video conferencing |
US5806005A (en) | 1996-05-10 | 1998-09-08 | Ricoh Company, Ltd. | Wireless image transfer from a digital still video camera to a networked computer |
US5832474A (en) | 1996-02-26 | 1998-11-03 | Matsushita Electric Industrial Co., Ltd. | Document search and retrieval system with partial match searching of user-drawn annotations |
US5832530A (en) | 1994-09-12 | 1998-11-03 | Adobe Systems Incorporated | Method and apparatus for identifying words described in a portable electronic document |
US5842194A (en) | 1995-07-28 | 1998-11-24 | Mitsubishi Denki Kabushiki Kaisha | Method of recognizing images of faces or general images using fuzzy combination of multiple resolutions |
WO1999005658A1 (en) | 1997-07-25 | 1999-02-04 | Checkpoint Systems, Inc. | Electrically, physically or virtually reactivating rfid tags |
US5873077A (en) | 1995-01-13 | 1999-02-16 | Ricoh Corporation | Method and apparatus for searching for and retrieving documents using a facsimile machine |
US5892843A (en) | 1997-01-21 | 1999-04-06 | Matsushita Electric Industrial Co., Ltd. | Title, caption and photo extraction from scanned document images |
US5899999A (en) | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
US5905502A (en) | 1995-08-04 | 1999-05-18 | Sun Microsystems, Inc. | Compression of three-dimensional graphics data using a generalized triangle mesh format utilizing a mesh buffer |
US5918012A (en) | 1996-03-29 | 1999-06-29 | British Telecommunications Public Limited Company | Hyperlinking time-based data files |
US5933525A (en) | 1996-04-10 | 1999-08-03 | Bbn Corporation | Language-independent and segmentation-free optical character recognition system and method |
US5956468A (en) | 1996-07-12 | 1999-09-21 | Seiko Epson Corporation | Document segmentation system |
US5968175A (en) | 1994-03-29 | 1999-10-19 | Kabushiki Kaisha Toshiba | Software use method control system |
US5999664A (en) | 1997-11-14 | 1999-12-07 | Xerox Corporation | System for searching a corpus of document images by user specified document layout components |
US5999915A (en) | 1994-02-16 | 1999-12-07 | Honicorp, Inc. | Computerized, multimedia, network, real time, interactive marketing and transactional system |
US6035055A (en) | 1997-11-03 | 2000-03-07 | Hewlett-Packard Company | Digital image management system in a distributed data access network system |
JP2000165645A (en) | 1998-11-25 | 2000-06-16 | Canon Inc | Peripheral device, its control method and storage medium |
US6104834A (en) | 1996-08-01 | 2000-08-15 | Ricoh Company Limited | Matching CCITT compressed document images |
EP0706283B1 (en) | 1993-06-22 | 2000-09-06 | Omron Corporation | Image processor including copying machine |
US6121969A (en) | 1997-07-29 | 2000-09-19 | The Regents Of The University Of California | Visual navigation in perceptual databases |
US6138129A (en) | 1997-12-16 | 2000-10-24 | World One Telecom, Ltd. | Method and apparatus for providing automated searching and linking of electronic documents |
US6192157B1 (en) | 1998-10-27 | 2001-02-20 | Hewlett-Packard Company | Modifications of postscript adaptive data compression (ADC) for 3 plane, 8 bit color images, JPEG lossy compression, and variable Q factors |
US6208771B1 (en) | 1996-12-20 | 2001-03-27 | Xerox Parc | Methods and apparatus for robust decoding of glyph address carpets |
US6223171B1 (en) | 1998-08-25 | 2001-04-24 | Microsoft Corporation | What-if index analysis utility for database systems |
US6253201B1 (en) | 1998-06-23 | 2001-06-26 | Philips Electronics North America Corporation | Scalable solution for image retrieval |
US20010011276A1 (en) | 1997-05-07 | 2001-08-02 | Robert T. Durst Jr. | Scanner enhanced remote control unit and system for automatically linking to on-line resources |
JP2001211359A (en) | 2000-01-25 | 2001-08-03 | Olympus Optical Co Ltd | Electronic camera |
US20010013546A1 (en) | 1996-01-09 | 2001-08-16 | Ross William Leslie | Identification system |
JP2001230916A (en) | 2000-02-18 | 2001-08-24 | Ricoh Co Ltd | File system |
US20010024514A1 (en) | 2000-03-22 | 2001-09-27 | Shinichi Matsunaga | Image processing device, singular spot detection method, and recording medium upon which singular spot detection program is recorded |
US6301386B1 (en) | 1998-12-09 | 2001-10-09 | Ncr Corporation | Methods and apparatus for gray image based text identification |
US20010042030A1 (en) | 2000-02-15 | 2001-11-15 | Sadao Ito | Information providing system for providing information about parts |
US20010042085A1 (en) | 1998-09-30 | 2001-11-15 | Mark Peairs | Automatic document classification using text and images |
US20010043741A1 (en) | 1998-10-13 | 2001-11-22 | Xerox Corporation | Automatic extraction of text regions and region borders for an electronic work surface |
US20010049700A1 (en) | 2000-05-26 | 2001-12-06 | Shinobu Ichikura | Information processing apparatus, information processing method and storage medium |
US6332039B1 (en) | 1997-08-07 | 2001-12-18 | Matsushita Graphic Communication Systems, Inc. | Structured document preparation apparatus and structured document preparation method |
US6345274B1 (en) | 1998-06-29 | 2002-02-05 | Eastman Kodak Company | Method and computer program product for subjective image content similarity-based retrieval |
US6353822B1 (en) | 1996-08-22 | 2002-03-05 | Massachusetts Institute Of Technology | Program-listing appendix |
US6363381B1 (en) | 1998-11-03 | 2002-03-26 | Ricoh Co., Ltd. | Compressed document matching |
US20020038430A1 (en) | 2000-09-13 | 2002-03-28 | Charles Edwards | System and method of data collection, processing, analysis, and annotation for monitoring cyber-threats and the notification thereof to subscribers |
US20020052872A1 (en) | 2000-10-27 | 2002-05-02 | Fuji Xerox Co., Ltd | Electronic information management server, electronic information management client, electronic information management method and recording medium for recording electronic information management program |
JP2002513480A (en) | 1996-06-04 | 2002-05-08 | ジェイ. マンシー,グラント | Techniques for creating and modifying 3D models and correlating such models with 2D pictures |
US20020054059A1 (en) | 2000-02-18 | 2002-05-09 | B.A. Schneiderman | Methods for the electronic annotation, retrieval, and use of electronic images |
US6393142B1 (en) | 1998-04-22 | 2002-05-21 | At&T Corp. | Method and apparatus for adaptive stripe based patch matching for depth estimation |
US6397213B1 (en) | 1999-05-12 | 2002-05-28 | Ricoh Company Ltd. | Search and retrieval using document decomposition |
US20020063709A1 (en) | 1998-05-13 | 2002-05-30 | Scott Gilbert | Panoramic movie which utilizes a series of captured panoramic images to display movement as observed by a viewer looking in a selected direction |
US6405172B1 (en) | 2000-09-09 | 2002-06-11 | Mailcode Inc. | Voice-enabled directory look-up based on recognized spoken initial characters |
US20020073236A1 (en) | 2000-01-14 | 2002-06-13 | Helgeson Christopher S. | Method and apparatus for managing data exchange among systems in a network |
US6408257B1 (en) | 1999-08-31 | 2002-06-18 | Xerox Corporation | Augmented-reality display method and system |
US6411953B1 (en) | 1999-01-25 | 2002-06-25 | Lucent Technologies Inc. | Retrieval and matching of color patterns based on a predetermined vocabulary and grammar |
US20020093538A1 (en) | 2000-08-22 | 2002-07-18 | Bruce Carlin | Network-linked interactive three-dimensional composition and display of saleable objects in situ in viewer-selected scenes for purposes of object promotion and procurement, and generation of object advertisements |
US20020102966A1 (en) | 2000-11-06 | 2002-08-01 | Lev Tsvi H. | Object identification method for portable devices |
EP1229496A2 (en) | 2001-02-02 | 2002-08-07 | Matsushita Electric Industrial Co., Ltd. | Information terminal apparatus and authenticating system |
US20020118379A1 (en) | 2000-12-18 | 2002-08-29 | Amit Chakraborty | System and user interface supporting user navigation of multimedia data file content |
US6448979B1 (en) | 1999-01-25 | 2002-09-10 | Airclic, Inc. | Printed medium activated interactive communication of multimedia information, including advertising |
US6457026B1 (en) | 1997-12-22 | 2002-09-24 | Ricoh Company, Ltd. | System to facilitate reading a document |
US6460036B1 (en) | 1994-11-29 | 2002-10-01 | Pinpoint Incorporated | System and method for providing customized electronic newspapers and target advertisements |
US20020146176A1 (en) | 2001-04-09 | 2002-10-10 | Meyers Gary Elliott | Method for representing and comparing digital images |
US6470264B2 (en) | 1997-06-03 | 2002-10-22 | Stephen Bide | Portable information-providing apparatus |
US20020154148A1 (en) | 2001-04-20 | 2002-10-24 | Aiichi Inoue | Image displaying method |
US20020157028A1 (en) | 2000-12-19 | 2002-10-24 | Fuji Xerox Co., Ltd. | Internet facsimile and control method thereof and communication instruction terminal |
US20020159640A1 (en) | 1999-07-02 | 2002-10-31 | Philips Electronics North America Corporation | Meta-descriptor for multimedia information |
US20020194264A1 (en) | 2001-05-31 | 2002-12-19 | Pioneer Corporation | Digital information distribution apparatus and method |
US20020191848A1 (en) | 2001-03-29 | 2002-12-19 | The Boeing Company | Method, computer program product, and system for performing automated text recognition and text search within a graphic file |
US20020191003A1 (en) | 2000-08-09 | 2002-12-19 | Hobgood Andrew W. | Method for using a motorized camera mount for tracking in augmented reality |
US6504571B1 (en) | 1998-05-18 | 2003-01-07 | International Business Machines Corporation | System and methods for querying digital image archives using recorded parameters |
US20030025714A1 (en) | 2001-07-16 | 2003-02-06 | Ebersole John Franklin | Method to view unseen atmospheric phenomenon using augmented reality |
US20030030835A1 (en) | 2001-08-07 | 2003-02-13 | Konica Corporation | Printing control apparatus and a method of charging for printing |
US20030030828A1 (en) * | 2001-07-13 | 2003-02-13 | Pau Soler | Spot color application in printer device |
US6537324B1 (en) | 1997-02-17 | 2003-03-25 | Ricoh Company, Ltd. | Generating and storing a link correlation table in hypertext documents at the time of storage |
US6567799B2 (en) | 1997-11-14 | 2003-05-20 | Adobe Systems Incorporated | Retrieving documents transitively linked to an initial document |
US20030098877A1 (en) | 2001-11-29 | 2003-05-29 | International Business Machines Corporation | Method and system for appending information to graphical files stored in specific graphical file formats |
US6574644B2 (en) | 1997-11-26 | 2003-06-03 | Siemens Corporate Research, Inc | Automatic capturing of hyperlink specifications for multimedia documents |
US6574375B1 (en) | 1992-04-06 | 2003-06-03 | Ricoh Company, Ltd. | Method for detecting inverted text images on a digital scanning device |
US20030110216A1 (en) | 2001-12-06 | 2003-06-12 | Mathias Althin | Mobile guide communications system |
US20030110130A1 (en) | 2001-07-20 | 2003-06-12 | International Business Machines Corporation | Method and system for delivering encrypted content with associated geographical-based advertisements |
US20030112930A1 (en) | 2001-12-18 | 2003-06-19 | Bosik Barry S. | Call management system responsive to network presence |
US6584223B1 (en) | 1998-04-02 | 2003-06-24 | Canon Kabushiki Kaisha | Image search apparatus and method |
US20030126147A1 (en) | 2001-10-12 | 2003-07-03 | Hassane Essafi | Method and a system for managing multimedia databases |
US20030122922A1 (en) | 2001-11-26 | 2003-07-03 | Saffer Kevin D. | Video e-mail system and associated method |
US20030128375A1 (en) | 2002-01-07 | 2003-07-10 | Ruhl Jan Matthias | Systems and methods for authenticating and verifying documents |
US20030142106A1 (en) | 2002-01-25 | 2003-07-31 | Xerox Corporation | Method and apparatus to convert bitmapped images for use in a structured text/graphics editor |
US20030151674A1 (en) | 2002-02-12 | 2003-08-14 | Qian Lin | Method and system for assessing the photo quality of a captured image in a digital still camera |
US20030152293A1 (en) | 2002-01-24 | 2003-08-14 | Joel Bresler | Method and system for locating position in printed texts and delivering multimedia information |
US6611862B2 (en) | 1994-05-31 | 2003-08-26 | Richard R. Reisman | User station software that controls transport and presentation of content from a remote source |
US20030187886A1 (en) | 2000-09-01 | 2003-10-02 | Hull Jonathan J. | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
US20030190094A1 (en) | 2001-11-29 | 2003-10-09 | Kazuaki Yokota | Document identification device, document definition method and document identification method |
US20030193530A1 (en) | 2002-04-11 | 2003-10-16 | Blackman Jeffrey R. | Barcode printing module |
US20030212585A1 (en) | 2002-05-09 | 2003-11-13 | Kabushiki Kaisha Toshiba | Idea drawing support method and program product therefor |
US20030229857A1 (en) | 2002-06-05 | 2003-12-11 | Fuji Xerox Co., Ltd. | Apparatus, method, and computer program product for document manipulation which embeds information in document data |
US20040017482A1 (en) | 2000-11-17 | 2004-01-29 | Jacob Weitman | Application for a mobile digital camera, that distinguish between text-, and image-information in an image |
US20040027604A1 (en) | 1999-08-05 | 2004-02-12 | Jeran Paul L. | Methods of document management, methods of automated document tracking, document tracking methods, and document tracking systems |
US20040036679A1 (en) | 2002-05-02 | 2004-02-26 | Emerson Harry E. | Computer system providing a visual indication when typing in caps lock mode |
US20040042667A1 (en) | 1999-04-08 | 2004-03-04 | Dar-Shyang Lee | Extracting information from symbolically compressed document images |
US6732915B1 (en) | 1999-09-10 | 2004-05-11 | Fuji Xerox Co., Ltd. | Systems and methods for controlling a presentation using physical objects |
US20040102898A1 (en) | 2002-11-23 | 2004-05-27 | Tatsuo Yokota | Navigation method and system for indicating area-specific traffic information |
US6751343B1 (en) | 1999-09-20 | 2004-06-15 | Ut-Battelle, Llc | Method for indexing and retrieving manufacturing-specific digital imagery based on image content |
US20040122811A1 (en) | 1997-01-10 | 2004-06-24 | Google, Inc. | Method for searching media |
US20040133582A1 (en) | 2002-10-11 | 2004-07-08 | Howard James V. | Systems and methods for recognition of individuals using multiple biometric searches |
US20040139391A1 (en) | 2003-01-15 | 2004-07-15 | Xerox Corporation | Integration of handwritten annotations into an electronic original |
US6766363B1 (en) | 2000-02-28 | 2004-07-20 | Barpoint.Com, Inc. | System and method of linking items in audio, visual, and printed media to related information stored on an electronic network using a mobile device |
US20040143644A1 (en) | 2003-01-21 | 2004-07-22 | Nec Laboratories America, Inc. | Meta-search engine architecture |
WO2004072897A2 (en) | 2003-02-06 | 2004-08-26 | Centerframe, L.L.C. | Obtaining person-specific images in a public venue |
US6791605B1 (en) | 1999-05-07 | 2004-09-14 | Eastman Kodak Company | Image capture and printing device |
US6799201B1 (en) | 2000-09-19 | 2004-09-28 | Motorola, Inc. | Remotely configurable multimedia entertainment and information system for vehicles |
US20040199531A1 (en) | 1999-12-01 | 2004-10-07 | Konan Technology Inc. | Content-based image retrieval system and method for retrieving image using the same |
US20040198396A1 (en) | 2000-05-08 | 2004-10-07 | Owest Communications International Inc. | Location based messaging method and system |
US6804332B1 (en) | 1994-09-30 | 2004-10-12 | Wildfire Communications, Inc. | Network based knowledgeable assistant |
US6804659B1 (en) | 2000-01-14 | 2004-10-12 | Ricoh Company Ltd. | Content based web advertising |
US20040201706A1 (en) | 2001-10-26 | 2004-10-14 | Katsutoshi Shimizu | Corrected image generating apparatus and corrected image generating program storage medium |
US20040205347A1 (en) | 2003-04-11 | 2004-10-14 | Ricoh Company, Ltd. | Techniques for accessing information captured during a presentation using a paper document handout for the presentation |
US20040215689A1 (en) | 2003-01-09 | 2004-10-28 | Dooley Michael J. | Computer and vision-based augmented interaction in the use of printed media |
US6813381B2 (en) | 2000-03-30 | 2004-11-02 | Glory Ltd. | Method and apparatus for identification of documents, and computer product |
US20040221244A1 (en) | 2000-12-20 | 2004-11-04 | Eastman Kodak Company | Method and apparatus for producing digital images with embedded image capture location icons |
US20040233235A1 (en) | 1999-12-07 | 2004-11-25 | Microsoft Corporation | Computer user interface architecture that saves a user's non-linear navigation history and intelligently maintains that history |
US6824057B2 (en) | 1994-05-25 | 2004-11-30 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US20040238621A1 (en) | 2001-07-10 | 2004-12-02 | American Express Travel Related Services Company, Inc. | Method and system for fingerprint biometrics on a fob |
US20040243514A1 (en) | 2003-01-23 | 2004-12-02 | John Wankmueller | System and method for secure telephone and computer transactions using voice authentication |
US20040260625A1 (en) | 2003-06-20 | 2004-12-23 | Yoshinori Usami | Digital photo print order acceptance method and digital image management method |
US20040260680A1 (en) | 2003-06-19 | 2004-12-23 | International Business Machines Corporation | Personalized indexing and searching for information in a distributed data processing system |
US6842755B2 (en) | 2000-09-25 | 2005-01-11 | Divine Technology Ventures | System and method for automatic retrieval of structured online documents |
US20050012960A1 (en) | 2003-07-18 | 2005-01-20 | Sharp Laboratories Of America, Inc. | Print content system and method for providing document control |
US6859909B1 (en) | 2000-03-07 | 2005-02-22 | Microsoft Corporation | System and method for annotating web-based documents |
US6865302B2 (en) | 2000-03-16 | 2005-03-08 | The Regents Of The University Of California | Perception-based image retrieval |
US20050080693A1 (en) | 2003-10-14 | 2005-04-14 | Foss Sheldon H. | Point-of-sale customer identification system |
US20050086224A1 (en) | 2003-10-15 | 2005-04-21 | Xerox Corporation | System and method for computing a measure of similarity between documents |
US20050086188A1 (en) | 2001-04-11 | 2005-04-21 | Hillis Daniel W. | Knowledge web |
US20050089246A1 (en) | 2003-10-27 | 2005-04-28 | Huitao Luo | Assessing image quality |
US20050100219A1 (en) | 2003-11-10 | 2005-05-12 | Kathrin Berkner | Features for retrieval and similarity matching of documents from the JPEG 2000-compressed domain |
WO2005043270A2 (en) | 2003-11-04 | 2005-05-12 | Nokia Corporation | System and method for registering attendance of entities associated with content creation |
US20050114325A1 (en) | 2000-10-30 | 2005-05-26 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20050125390A1 (en) | 2003-12-03 | 2005-06-09 | Oliver Hurst-Hiller | Automated satisfaction measurement for web search |
US20050129293A1 (en) | 2003-12-11 | 2005-06-16 | Tinku Acharya | Fingerprint minutiae matching using scoring techniques |
US20050135483A1 (en) | 2003-12-23 | 2005-06-23 | Genesis Microchip Inc. | Temporal motion vector filtering |
EP1555626A2 (en) | 2004-01-15 | 2005-07-20 | Microsoft Corporation | Image-based indexing and retrieval of text documents |
US20050160115A1 (en) | 2001-05-22 | 2005-07-21 | Microsoft Corporation | Document imaging and indexing system |
US20050160258A1 (en) | 2003-12-11 | 2005-07-21 | Bioobservation Systems Limited | Detecting objectionable content in displayed images |
US6922699B2 (en) | 1999-01-26 | 2005-07-26 | Xerox Corporation | System and method for quantitatively representing data objects in vector space |
US20050165784A1 (en) | 2004-01-23 | 2005-07-28 | Garrison Gomez | System and method to store and retrieve identifier associated information content |
US20050169520A1 (en) | 2003-12-29 | 2005-08-04 | Canon Kabushiki Kaisha | Detecting human faces and detecting red eyes |
US20050182773A1 (en) | 2004-02-18 | 2005-08-18 | Feinsmith Jason B. | Machine-implemented activity management system using asynchronously shared activity data objects and journal data items |
US20050185060A1 (en) | 2004-02-20 | 2005-08-25 | Neven Hartmut Sr. | Image base inquiry system for search engines for mobile telephones with integrated camera |
US20050185225A1 (en) | 2003-12-12 | 2005-08-25 | Brawn Dennis E. | Methods and apparatus for imaging documents |
US20050190972A1 (en) | 2004-02-11 | 2005-09-01 | Thomas Graham A. | System and method for position determination |
US20050190273A1 (en) | 2001-10-09 | 2005-09-01 | Microsoft Corporation | System and method for exchanging images |
US6940491B2 (en) | 2000-10-27 | 2005-09-06 | International Business Machines Corporation | Method and system for generating hyperlinked physical copies of hyperlinked electronic documents |
US20050198095A1 (en) | 2003-12-31 | 2005-09-08 | Kavin Du | System and method for obtaining information relating to an item of commerce using a portable imaging device |
US20050216257A1 (en) | 2004-03-18 | 2005-09-29 | Pioneer Corporation | Sound information reproducing apparatus and method of preparing keywords of music data |
JP2005286395A (en) | 2004-03-26 | 2005-10-13 | Fuji Photo Film Co Ltd | Image processing system and method therefor |
US20050234851A1 (en) | 2004-02-15 | 2005-10-20 | King Martin T | Automatic modification of web pages |
US20050240381A1 (en) | 2004-04-26 | 2005-10-27 | Pjm Interconnection, Llc. | Systems and methods for analysis of a commodity transmission network |
US6963358B2 (en) | 2000-11-10 | 2005-11-08 | Dropfire, Inc. | Wireless digital camera adapter and systems and methods related thereto and for use with such an adapter |
US6964374B1 (en) | 1998-10-02 | 2005-11-15 | Lucent Technologies Inc. | Retrieval and manipulation of electronically stored information via pointers embedded in the associated printed material |
US20050261990A1 (en) | 2004-04-16 | 2005-11-24 | Russell Gocht | Mobile query system and method based on visual cues |
US20050273812A1 (en) | 2004-06-02 | 2005-12-08 | Kabushiki Kaisha Toshiba | User profile editing apparatus, method and program |
US6981224B1 (en) | 2000-05-05 | 2005-12-27 | Point Roll, Inc. | System and method for conditionally associating a plurality of enabled images with a predetermined image |
US6980962B1 (en) | 1999-03-02 | 2005-12-27 | Quixtar Investments, Inc. | Electronic commerce transactions within a marketing system that may contain a membership buying opportunity |
US20050288911A1 (en) | 2004-06-28 | 2005-12-29 | Porikli Fatih M | Hidden markov model based object tracking and similarity metrics |
US20050289447A1 (en) | 2004-06-29 | 2005-12-29 | The Boeing Company | Systems and methods for generating and storing referential links in a database |
US20050289182A1 (en) | 2004-06-15 | 2005-12-29 | Sand Hill Systems Inc. | Document management system with enhanced intelligent document recognition capabilities |
US20050288859A1 (en) | 2004-03-23 | 2005-12-29 | Golding Andrew R | Visually-oriented driving directions in digital mapping system |
US20060002607A1 (en) | 2000-11-06 | 2006-01-05 | Evryx Technologies, Inc. | Use of image-derived information as search criteria for internet and other search engines |
US20060012677A1 (en) | 2004-02-20 | 2006-01-19 | Neven Hartmut Sr | Image-based search engine for mobile phones with camera |
US20060020630A1 (en) | 2004-07-23 | 2006-01-26 | Stager Reed R | Facial database methods and systems |
US6993573B2 (en) | 2003-06-06 | 2006-01-31 | Neomedia Technologies, Inc. | Automatic access of internet content with a camera-enabled cell phone |
US20060041605A1 (en) | 2004-04-01 | 2006-02-23 | King Martin T | Determining actions involving captured information and electronic content associated with rendered documents |
JP2006053568A (en) | 2004-08-13 | 2006-02-23 | Fujitsu Ltd | Deteriorated character image generation program, method and apparatus |
US20060043188A1 (en) | 2004-08-27 | 2006-03-02 | Gregg Kricorissian | Imaging method and apparatus for object identification |
US20060048059A1 (en) | 2004-08-26 | 2006-03-02 | Henry Etkin | System and method for dynamically generating, maintaining, and growing an online social network |
JP2006059351A (en) | 2004-08-18 | 2006-03-02 | Fujitsu Ltd | Deterioration dictionary generation program, method and device |
US20060053097A1 (en) | 2004-04-01 | 2006-03-09 | King Martin T | Searching and accessing documents on private networks for use with captures from rendered documents |
CN1245935C (en) | 1999-02-05 | 2006-03-22 | 宝洁公司 | Photostable UV protection compositions |
US20060070120A1 (en) | 2004-09-02 | 2006-03-30 | Brother Kogyo Kabushiki Kaisha | File transmitting device and multi function device |
US20060074828A1 (en) | 2004-09-14 | 2006-04-06 | Heumann John M | Methods and apparatus for detecting temporal process variation and for managing and predicting performance of automatic classifiers |
US7031965B1 (en) | 2000-03-23 | 2006-04-18 | Mitsubishi Denki Kabushiki Kaisha | Image retrieving and delivering system and image retrieving and delivering method |
US20060082438A1 (en) | 2003-09-05 | 2006-04-20 | Bazakos Michael E | Distributed stand-off verification and face recognition systems (FRS) |
US20060085735A1 (en) | 2003-09-18 | 2006-04-20 | Fujitsu Limited | Annotation management system, annotation managing method, document transformation server, document transformation program, and electronic document attachment program |
US20060085477A1 (en) | 2004-10-01 | 2006-04-20 | Ricoh Company, Ltd. | Techniques for retrieving documents using an image capture device |
US20060104515A1 (en) | 2004-07-19 | 2006-05-18 | King Martin T | Automatic modification of WEB pages |
US7051086B2 (en) | 1995-07-27 | 2006-05-23 | Digimarc Corporation | Method of linking on-line data to printed documents |
US20060112092A1 (en) | 2002-08-09 | 2006-05-25 | Bell Canada | Content-based image retrieval method |
EP1662064A1 (en) | 2000-04-24 | 2006-05-31 | Hunter Douglas Inc. | Compressible structural panel |
US20060116555A1 (en) | 2004-11-18 | 2006-06-01 | Honeywell International Inc. | Controlled environment thermal image detection system and methods regarding same |
US20060119880A1 (en) | 2004-12-02 | 2006-06-08 | Dandekar Shree A | Local photo printing |
US20060123347A1 (en) | 2004-12-06 | 2006-06-08 | Joe Hewitt | Managing and collaborating with digital content using a dynamic user interface |
US20060122884A1 (en) | 1997-12-22 | 2006-06-08 | Ricoh Company, Ltd. | Method, system and computer code for content based web advertising |
US20060122983A1 (en) | 2004-12-03 | 2006-06-08 | King Martin T | Locating electronic instances of documents based on rendered instances, document fragment digest generation, and digest based document fragment determination |
US7062722B1 (en) * | 2000-08-22 | 2006-06-13 | Bruce Carlin | Network-linked interactive three-dimensional composition and display of saleable objects in situ in viewer-selected scenes for purposes of promotion and procurement |
US20060140614A1 (en) | 2004-12-28 | 2006-06-29 | Samsung Electronic Co., Ltd. | Apparatus, medium, and method for photographing based on face detection |
US20060143176A1 (en) | 2002-04-15 | 2006-06-29 | International Business Machines Corporation | System and method for measuring image similarity based on semantic meaning |
US20060140475A1 (en) | 2004-12-15 | 2006-06-29 | Chin Frances M | Network support for corrupting images captured at a prohibited location |
US20060150079A1 (en) | 2004-12-17 | 2006-07-06 | International Business Machines Corporation | Method for associating annotations with document families |
US7089487B2 (en) | 1992-02-25 | 2006-08-08 | Mary Y. Y. Tsai | Method and apparatus for linking designated portions of a received document image with an electronic address |
US7092953B1 (en) | 2000-12-28 | 2006-08-15 | Rightlsline, Inc. | Apparatus and methods for intellectual property database navigation |
JP2006215756A (en) | 2005-02-02 | 2006-08-17 | Dainippon Ink & Chem Inc | Image processing apparatus, image processing method, and program for the same |
US20060190812A1 (en) | 2005-02-22 | 2006-08-24 | Geovector Corporation | Imaging systems including hyperlink associations |
US20060200480A1 (en) | 2005-03-01 | 2006-09-07 | Harris David N | System and method for using product identifiers |
US20060206335A1 (en) | 2003-03-17 | 2006-09-14 | Eric Thelen | Method for remote control of an audio device |
US20060218225A1 (en) | 2005-03-28 | 2006-09-28 | Hee Voon George H | Device for sharing social network information among users over a network |
US20060227992A1 (en) | 2005-04-08 | 2006-10-12 | Rathus Spencer A | System and method for accessing electronic data via an image search engine |
US20060240862A1 (en) | 2004-02-20 | 2006-10-26 | Hartmut Neven | Mobile image-based information retrieval system |
US7134095B1 (en) | 1999-10-20 | 2006-11-07 | Gateway, Inc. | Simulated three-dimensional navigational menu system |
US20060251292A1 (en) | 2005-05-09 | 2006-11-09 | Salih Burak Gokturk | System and method for recognizing objects from images and identifying relevancy amongst images and information |
US20060253491A1 (en) | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling search and retrieval from image files based on recognized information |
US20060251339A1 (en) | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling the use of captured images through recognition |
US20060253439A1 (en) | 2005-05-09 | 2006-11-09 | Liwei Ren | Matching engine for querying relevant documents |
US7136093B1 (en) * | 1999-03-23 | 2006-11-14 | Canon Kabushiki Kaisha | Information presenting apparatus, operation processing method therefor, storage medium storing program for executing operation processing |
US20060262352A1 (en) | 2004-10-01 | 2006-11-23 | Hull Jonathan J | Method and system for image matching in a mixed media environment |
US20060262962A1 (en) | 2004-10-01 | 2006-11-23 | Hull Jonathan J | Method And System For Position-Based Image Matching In A Mixed Media Environment |
US20060262976A1 (en) | 2004-10-01 | 2006-11-23 | Hart Peter E | Method and System for Multi-Tier Image Matching in a Mixed Media Environment |
US7150399B2 (en) | 2004-06-09 | 2006-12-19 | Ricoh Co., Ltd. | Embedding barcode data in an auxiliary field of an image file |
US20060286951A1 (en) | 2003-08-25 | 2006-12-21 | Fujitsu Ten Limited | Music information,updating system, music information broadcasting apparatus, terminal apparatus having music information updating function, music information updating method, music information broadcasting method, and music information updating method of terminal apparatus |
US20060285755A1 (en) * | 2005-06-16 | 2006-12-21 | Strider Labs, Inc. | System and method for recognition in 2D images using 3D class models |
US20060285772A1 (en) | 2004-10-01 | 2006-12-21 | Hull Jonathan J | System and methods for creation and use of a mixed media environment |
US20060285172A1 (en) | 2004-10-01 | 2006-12-21 | Hull Jonathan J | Method And System For Document Fingerprint Matching In A Mixed Media Environment |
US20060294049A1 (en) | 2005-06-27 | 2006-12-28 | Microsoft Corporation | Back-off mechanism for search |
US20070003147A1 (en) | 2005-07-01 | 2007-01-04 | Microsoft Corporation | Grammatical parsing of document visual structures |
US20070003166A1 (en) | 2005-06-30 | 2007-01-04 | Kathrin Berkner | White space graphs and trees for content-adaptive scaling of document images |
US7167574B2 (en) | 2002-03-14 | 2007-01-23 | Seiko Epson Corporation | Method and apparatus for content-based image copy detection |
US20070019261A1 (en) | 2005-07-21 | 2007-01-25 | Yu-Chung Chu | Method and system for distributing consumer photograph-based advertisement and/or information |
US7174031B2 (en) | 1999-05-19 | 2007-02-06 | Digimarc Corporation | Methods for using wireless phones having optical capabilities |
US20070036469A1 (en) | 2005-06-20 | 2007-02-15 | Samsung Electronics Co., Ltd. | Method and system for providing image-related information to user, and mobile terminal therefor |
US20070041668A1 (en) | 2005-07-28 | 2007-02-22 | Canon Kabushiki Kaisha | Search apparatus and search method |
US7185274B1 (en) | 1999-12-07 | 2007-02-27 | Microsoft Corporation | Computer user interface architecture wherein users interact with both content and user interface by activating links |
US20070047819A1 (en) | 2005-08-23 | 2007-03-01 | Hull Jonathan J | Data organization and access for mixed media document system |
US20070053513A1 (en) | 1999-10-05 | 2007-03-08 | Hoffberg Steven M | Intelligent electronic appliance system and method |
US20070052997A1 (en) * | 2005-08-23 | 2007-03-08 | Hull Jonathan J | System and methods for portable device for mixed media system |
US20070076922A1 (en) | 2005-09-30 | 2007-04-05 | Sony United Kingdom Limited | Object detection |
US7206820B1 (en) | 2000-03-18 | 2007-04-17 | Digimarc Corporation | System for linking from object to remote resource |
EP1783681A1 (en) | 2005-06-30 | 2007-05-09 | Olympus Corporation | Retrieval system and retrieval method |
US20070106721A1 (en) | 2005-11-04 | 2007-05-10 | Philipp Schloter | Scalable visual search system simplifying access to network and device functionality |
US20070115373A1 (en) | 2005-11-22 | 2007-05-24 | Eastman Kodak Company | Location based image classification with map segmentation |
US7232057B2 (en) | 1994-05-25 | 2007-06-19 | Rathus Spencer A | Method and apparatus for accessing electronic data via a familiar printed medium |
US7236632B2 (en) | 2003-04-11 | 2007-06-26 | Ricoh Company, Ltd. | Automated techniques for comparing contents of images |
US20070150466A1 (en) | 2004-12-29 | 2007-06-28 | Scott Brave | Method and apparatus for suggesting/disambiguation query terms based upon usage patterns observed |
WO2007073347A1 (en) | 2005-12-19 | 2007-06-28 | Agency For Science, Technology And Research | Annotation of video footage and personalised video generation |
US7240279B1 (en) | 2002-06-19 | 2007-07-03 | Microsoft Corporation | XML patterns language |
US20070165904A1 (en) * | 2005-08-23 | 2007-07-19 | Nudd Geoffrey H | System and Method for Using Individualized Mixed Document |
US7249123B2 (en) | 2002-10-31 | 2007-07-24 | International Business Machines Corporation | System and method for building social networks based on activity around shared virtual objects |
US20070174269A1 (en) | 2006-01-23 | 2007-07-26 | Microsoft Corporation | Generating clusters of images for search results |
US7251689B2 (en) | 2002-03-27 | 2007-07-31 | International Business Machines Corporation | Managing storage resources in decentralized networks |
US20070175998A1 (en) | 2005-09-01 | 2007-08-02 | Lev Zvi H | System and method for reliable content access using a cellular/wireless device with imaging capabilities |
US7263205B2 (en) | 2004-12-06 | 2007-08-28 | Dspv, Ltd. | System and method of generic symbol recognition and user authentication using a communication device with imaging capabilities |
US20070233613A1 (en) | 2006-03-31 | 2007-10-04 | Ricoh Company, Ltd. | Techniques for using media keys |
US20070276845A1 (en) | 2006-05-12 | 2007-11-29 | Tele Atlas North America, Inc. | Locality indexes and method for indexing localities |
US7305435B2 (en) | 2003-08-25 | 2007-12-04 | Nokia Corporation | Internet access via smartphone camera |
US7310779B2 (en) | 2003-06-26 | 2007-12-18 | International Business Machines Corporation | Method for creating and selecting active regions on physical documents |
US7310769B1 (en) | 2003-03-12 | 2007-12-18 | Adobe Systems Incorporated | Text encoding using dummy font |
US20080010605A1 (en) | 2006-06-12 | 2008-01-10 | Metacarta, Inc. | Systems and methods for generating and correcting location references extracted from text |
US7337175B2 (en) | 2000-08-14 | 2008-02-26 | Alcatel | Method of storing data in a multimedia file using relative timebases |
US20080071767A1 (en) | 2006-08-25 | 2008-03-20 | Semdirector, Inc. | System and method for measuring the effectiveness of an on-line advertisement campaign |
US20080078836A1 (en) | 2006-09-29 | 2008-04-03 | Hiroshi Tomita | Barcode for two-way verification of a document |
US7359094B1 (en) | 1999-12-15 | 2008-04-15 | Fuji Xerox Co., Ltd. | Image processing apparatus and image forming medium |
US7366979B2 (en) | 2001-03-09 | 2008-04-29 | Copernicus Investments, Llc | Method and apparatus for annotating a document |
US20080106594A1 (en) | 2006-11-07 | 2008-05-08 | The Board Of Trustees Of The Leland Stanford Jr. University | System and method for tagging objects in a panoramic video and associating functions and indexing panoramic images with same |
US7386789B2 (en) | 2004-02-27 | 2008-06-10 | Hewlett-Packard Development Company, L.P. | Method for determining logical components of a document |
US7392287B2 (en) | 2001-03-27 | 2008-06-24 | Hemisphere Ii Investment Lp | Method and apparatus for sharing information using a handheld device |
US7406214B2 (en) | 1999-05-19 | 2008-07-29 | Digimarc Corporation | Methods and devices employing optical sensors and/or steganography |
US7421153B1 (en) | 2002-04-05 | 2008-09-02 | Bank Of America Corporation | Image and data processing system |
US7424541B2 (en) | 2004-02-09 | 2008-09-09 | Proxpro, Inc. | Method and computer system for matching mobile device users for business and social networking |
US20080229192A1 (en) | 2007-03-15 | 2008-09-18 | Microsoft Corporation | Interactive image tagging |
WO2008129373A2 (en) | 2007-04-24 | 2008-10-30 | Nokia Corporation | Method, device and computer program product for integrating code-based and optical character recognition technologies into a mobile visual search |
US20080275881A1 (en) | 2006-09-05 | 2008-11-06 | Gloto Corporation | Real time collaborative on-line multimedia albums |
US7450760B2 (en) | 2005-05-18 | 2008-11-11 | Scanr, Inc. | System and method for capturing and processing business data |
US20080288476A1 (en) | 2007-05-17 | 2008-11-20 | Sang-Heun Kim | Method and system for desktop tagging of a web page |
US7458014B1 (en) | 1999-12-07 | 2008-11-25 | Microsoft Corporation | Computer user interface architecture wherein both content and user interface are composed of documents with links |
US20080296362A1 (en) | 2001-11-30 | 2008-12-04 | International Barcode Corporation | Method for identifying and authenticating goods using codes, bar codes and radio frequency identification |
US7463270B2 (en) * | 2006-02-10 | 2008-12-09 | Microsoft Corporation | Physical-virtual interpolation |
US20080310717A1 (en) | 2006-02-01 | 2008-12-18 | Motorola, Inc. | Apparatus and Method for Image Labeling |
US7489415B2 (en) | 2003-07-16 | 2009-02-10 | Ricoh Company, Ltd. | Document processing system |
US20090067726A1 (en) | 2006-07-31 | 2009-03-12 | Berna Erol | Computation of a recognizability score (quality predictor) for image retrieval |
US7509386B2 (en) | 2001-12-11 | 2009-03-24 | Sony Corporation | Chat system displaying a link arrow directed from a hyperlink to content of an associated attachment file |
US7546524B1 (en) | 2005-03-30 | 2009-06-09 | Amazon Technologies, Inc. | Electronic input device, system, and method using human-comprehensible content to automatically correlate an annotation of a paper document with a digital version of the document |
US20090152357A1 (en) | 2007-12-12 | 2009-06-18 | 3M Innovative Properties Company | Document verification using dynamic document identification framework |
US7567262B1 (en) | 2005-02-25 | 2009-07-28 | IDV Solutions LLC | Hybrid graphics for interactive reporting |
US7587681B2 (en) * | 2003-06-16 | 2009-09-08 | Sony Computer Entertainment Inc. | Method and apparatus for presenting information |
US7585224B2 (en) * | 2005-04-28 | 2009-09-08 | Nintendo Co., Ltd. | Storage medium having game program stored therein and game apparatus |
US7593961B2 (en) | 2003-04-30 | 2009-09-22 | Canon Kabushiki Kaisha | Information processing apparatus for retrieving image data similar to an entered image |
US7613686B2 (en) | 1999-09-13 | 2009-11-03 | Microsoft Corporation | Image retrieval based on relevance feedback |
US7644078B2 (en) | 2005-02-28 | 2010-01-05 | Gm Global Technology Operations, Inc. | System and method for mining of temporal data |
US20100013615A1 (en) | 2004-03-31 | 2010-01-21 | Carnegie Mellon University | Obstacle detection having enhanced classification |
US7653238B2 (en) | 2003-12-05 | 2010-01-26 | British Telecommunications Plc | Image filtering based on comparison of pixel groups |
US7668405B2 (en) | 2006-04-07 | 2010-02-23 | Eastman Kodak Company | Forming connections between image collections |
US7676767B2 (en) * | 2005-06-15 | 2010-03-09 | Microsoft Corporation | Peel back user interface to show hidden functions |
US20100063961A1 (en) | 2008-09-05 | 2010-03-11 | Fotonauts, Inc. | Reverse Tagging of Images in System for Managing and Sharing Digital Images |
US7680850B2 (en) | 2006-03-31 | 2010-03-16 | Fujitsu Limited | Computer-readable recording medium storing information search program, information search method, and information search system |
US7683933B2 (en) * | 2000-05-19 | 2010-03-23 | Canon Kabushiki Kaisha | Image display apparatus, image display system, and image display method |
US7746376B2 (en) | 2004-06-16 | 2010-06-29 | Felipe Mendoza | Method and apparatus for accessing multi-dimensional mapping and information |
US7752534B2 (en) | 2006-09-19 | 2010-07-06 | International Business Machines Corporation | Method and apparatus for customizing the display of multidimensional data |
US20100174783A1 (en) | 2007-10-12 | 2010-07-08 | Rony Zarom | System and method for coordinating simultaneous edits of shared digital data |
US7761436B2 (en) | 2006-01-03 | 2010-07-20 | Yahoo! Inc. | Apparatus and method for controlling content access based on shared annotations for annotated users in a folksonomy scheme |
US7812986B2 (en) | 2005-08-23 | 2010-10-12 | Ricoh Co. Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US7872669B2 (en) | 2004-01-22 | 2011-01-18 | Massachusetts Institute Of Technology | Photo-based mobile deixis system and related techniques |
US7882113B2 (en) | 2003-03-28 | 2011-02-01 | International Business Machines Corporation | Method, apparatus, and system for formatting time data to improve processing in a sort utility |
US7930292B2 (en) | 2005-01-28 | 2011-04-19 | Canon Kabushiki Kaisha | Information processing apparatus and control method thereof |
US7946491B2 (en) | 2006-08-03 | 2011-05-24 | Nokia Corporation | Method, apparatus, and computer program product for providing a camera barcode reader |
US20110121069A1 (en) | 2007-01-18 | 2011-05-26 | Target Brands, Inc. | Barcodes with Graphical Elements |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5867597A (en) * | 1995-09-05 | 1999-02-02 | Ricoh Corporation | High-speed retrieval by example |
WO2007023994A1 (en) * | 2005-08-23 | 2007-03-01 | Ricoh Company, Ltd. | System and methods for creation and use of a mixed media environment |
-
2007
- 2007-07-11 US US11/827,530 patent/US8276088B2/en active Active
-
2008
- 2008-07-10 JP JP2008180793A patent/JP5181888B2/en active Active
- 2008-07-11 EP EP08252377.0A patent/EP2015227B8/en active Active
Patent Citations (342)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1915993A (en) | 1931-04-27 | 1933-06-27 | Gen Electric | Statistical machine |
US4759075A (en) | 1983-03-14 | 1988-07-19 | Ana Tech Corporation | Method and apparatus for vectorizing documents and symbol recognition |
US5027421A (en) | 1987-11-30 | 1991-06-25 | Kabushiki Kaisha Toshiba | Data processing apparatus having function for registering abstract image data |
US5035302A (en) | 1989-03-03 | 1991-07-30 | Otis Elevator Company | "Artificial Intelligence" based learning system predicting "Peak-Period" times for elevator dispatching |
US5416892A (en) | 1990-03-07 | 1995-05-16 | Fujitsu Limited | Best first search considering difference between scores |
US5077805A (en) | 1990-05-07 | 1991-12-31 | Eastman Kodak Company | Hybrid feature-based and template matching optical character recognition system |
US5109439A (en) | 1990-06-12 | 1992-04-28 | Horst Froessl | Mass document storage and retrieval system |
US5263100A (en) | 1991-07-19 | 1993-11-16 | Samsung Electronics Co., Ltd. | Image coding method and apparatus |
US5392447A (en) | 1992-01-10 | 1995-02-21 | Eastman Kodak Compay | Image-based electronic pocket organizer with integral scanning unit |
US7089487B2 (en) | 1992-02-25 | 2006-08-08 | Mary Y. Y. Tsai | Method and apparatus for linking designated portions of a received document image with an electronic address |
US6574375B1 (en) | 1992-04-06 | 2003-06-03 | Ricoh Company, Ltd. | Method for detecting inverted text images on a digital scanning device |
US5432864A (en) | 1992-10-05 | 1995-07-11 | Daozheng Lu | Identification card verification system |
US5579471A (en) | 1992-11-09 | 1996-11-26 | International Business Machines Corporation | Image query system and method |
US5493689A (en) | 1993-03-01 | 1996-02-20 | International Business Machines Corporation | System for configuring an event driven interface including control blocks defining good loop locations in a memory which represent detection of a characteristic pattern |
US5546502A (en) | 1993-03-19 | 1996-08-13 | Ricoh Company, Ltd. | Automatic invocation of computational resources without user intervention |
EP0706283B1 (en) | 1993-06-22 | 2000-09-06 | Omron Corporation | Image processor including copying machine |
US5553217A (en) | 1993-09-23 | 1996-09-03 | Ricoh Company, Ltd. | Document layout using tiling |
US5999915A (en) | 1994-02-16 | 1999-12-07 | Honicorp, Inc. | Computerized, multimedia, network, real time, interactive marketing and transactional system |
US5968175A (en) | 1994-03-29 | 1999-10-19 | Kabushiki Kaisha Toshiba | Software use method control system |
US5465353A (en) | 1994-04-01 | 1995-11-07 | Ricoh Company, Ltd. | Image matching and retrieval by multi-access redundant hashing |
US5761344A (en) | 1994-04-15 | 1998-06-02 | Canon Kabushiki Kaisha | Image pre-processor for character recognition system |
US6929182B2 (en) | 1994-05-25 | 2005-08-16 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6866196B1 (en) | 1994-05-25 | 2005-03-15 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6824057B2 (en) | 1994-05-25 | 2004-11-30 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US7232057B2 (en) | 1994-05-25 | 2007-06-19 | Rathus Spencer A | Method and apparatus for accessing electronic data via a familiar printed medium |
US6827267B2 (en) | 1994-05-25 | 2004-12-07 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6830187B2 (en) | 1994-05-25 | 2004-12-14 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6834804B2 (en) | 1994-05-25 | 2004-12-28 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6843411B2 (en) | 1994-05-25 | 2005-01-18 | Spencer A. Rathus | Method and apparatus for accessing electronic data via a familiar printed medium |
US6611862B2 (en) | 1994-05-31 | 2003-08-26 | Richard R. Reisman | User station software that controls transport and presentation of content from a remote source |
US5832530A (en) | 1994-09-12 | 1998-11-03 | Adobe Systems Incorporated | Method and apparatus for identifying words described in a portable electronic document |
US6804332B1 (en) | 1994-09-30 | 2004-10-12 | Wildfire Communications, Inc. | Network based knowledgeable assistant |
US5555556A (en) | 1994-09-30 | 1996-09-10 | Xerox Corporation | Method and apparatus for document segmentation by background analysis |
US6460036B1 (en) | 1994-11-29 | 2002-10-01 | Pinpoint Incorporated | System and method for providing customized electronic newspapers and target advertisements |
US5752055A (en) | 1994-12-27 | 1998-05-12 | International Business Machine Corp. | Systems and method for automatically linking parts within compound documents |
US5873077A (en) | 1995-01-13 | 1999-02-16 | Ricoh Corporation | Method and apparatus for searching for and retrieving documents using a facsimile machine |
US7051086B2 (en) | 1995-07-27 | 2006-05-23 | Digimarc Corporation | Method of linking on-line data to printed documents |
US5842194A (en) | 1995-07-28 | 1998-11-24 | Mitsubishi Denki Kabushiki Kaisha | Method of recognizing images of faces or general images using fuzzy combination of multiple resolutions |
US5905502A (en) | 1995-08-04 | 1999-05-18 | Sun Microsystems, Inc. | Compression of three-dimensional graphics data using a generalized triangle mesh format utilizing a mesh buffer |
US5764277A (en) | 1995-11-08 | 1998-06-09 | Bell Communications Research, Inc. | Group-of-block based video signal combining for multipoint continuous presence video conferencing |
US5706097A (en) | 1995-11-13 | 1998-01-06 | Eastman Kodak Company | Index print with a digital recording medium containing still images, motion sequences, and sound sequences |
US20010013546A1 (en) | 1996-01-09 | 2001-08-16 | Ross William Leslie | Identification system |
US5832474A (en) | 1996-02-26 | 1998-11-03 | Matsushita Electric Industrial Co., Ltd. | Document search and retrieval system with partial match searching of user-drawn annotations |
US5918012A (en) | 1996-03-29 | 1999-06-29 | British Telecommunications Public Limited Company | Hyperlinking time-based data files |
US5933525A (en) | 1996-04-10 | 1999-08-03 | Bbn Corporation | Language-independent and segmentation-free optical character recognition system and method |
US5806005A (en) | 1996-05-10 | 1998-09-08 | Ricoh Company, Ltd. | Wireless image transfer from a digital still video camera to a networked computer |
JP2002513480A (en) | 1996-06-04 | 2002-05-08 | ジェイ. マンシー,グラント | Techniques for creating and modifying 3D models and correlating such models with 2D pictures |
US5956468A (en) | 1996-07-12 | 1999-09-21 | Seiko Epson Corporation | Document segmentation system |
US6104834A (en) | 1996-08-01 | 2000-08-15 | Ricoh Company Limited | Matching CCITT compressed document images |
US6353822B1 (en) | 1996-08-22 | 2002-03-05 | Massachusetts Institute Of Technology | Program-listing appendix |
US5899999A (en) | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
US6208771B1 (en) | 1996-12-20 | 2001-03-27 | Xerox Parc | Methods and apparatus for robust decoding of glyph address carpets |
US20040122811A1 (en) | 1997-01-10 | 2004-06-24 | Google, Inc. | Method for searching media |
US5892843A (en) | 1997-01-21 | 1999-04-06 | Matsushita Electric Industrial Co., Ltd. | Title, caption and photo extraction from scanned document images |
US6537324B1 (en) | 1997-02-17 | 2003-03-25 | Ricoh Company, Ltd. | Generating and storing a link correlation table in hypertext documents at the time of storage |
US7363580B2 (en) | 1997-02-17 | 2008-04-22 | Ricoh Company, Ltd. | Intermediate address table for converting position information to an address of a link structure correlated information file on a hypertext document |
US20030121006A1 (en) | 1997-02-17 | 2003-06-26 | Yasuhiro Tabata | Document information management system |
US20010011276A1 (en) | 1997-05-07 | 2001-08-02 | Robert T. Durst Jr. | Scanner enhanced remote control unit and system for automatically linking to on-line resources |
US6470264B2 (en) | 1997-06-03 | 2002-10-22 | Stephen Bide | Portable information-providing apparatus |
WO1999005658A1 (en) | 1997-07-25 | 1999-02-04 | Checkpoint Systems, Inc. | Electrically, physically or virtually reactivating rfid tags |
US6121969A (en) | 1997-07-29 | 2000-09-19 | The Regents Of The University Of California | Visual navigation in perceptual databases |
US6332039B1 (en) | 1997-08-07 | 2001-12-18 | Matsushita Graphic Communication Systems, Inc. | Structured document preparation apparatus and structured document preparation method |
US6035055A (en) | 1997-11-03 | 2000-03-07 | Hewlett-Packard Company | Digital image management system in a distributed data access network system |
US5999664A (en) | 1997-11-14 | 1999-12-07 | Xerox Corporation | System for searching a corpus of document images by user specified document layout components |
US6567799B2 (en) | 1997-11-14 | 2003-05-20 | Adobe Systems Incorporated | Retrieving documents transitively linked to an initial document |
US6574644B2 (en) | 1997-11-26 | 2003-06-03 | Siemens Corporate Research, Inc | Automatic capturing of hyperlink specifications for multimedia documents |
US6138129A (en) | 1997-12-16 | 2000-10-24 | World One Telecom, Ltd. | Method and apparatus for providing automated searching and linking of electronic documents |
US6457026B1 (en) | 1997-12-22 | 2002-09-24 | Ricoh Company, Ltd. | System to facilitate reading a document |
US20060122884A1 (en) | 1997-12-22 | 2006-06-08 | Ricoh Company, Ltd. | Method, system and computer code for content based web advertising |
US6584223B1 (en) | 1998-04-02 | 2003-06-24 | Canon Kabushiki Kaisha | Image search apparatus and method |
US6393142B1 (en) | 1998-04-22 | 2002-05-21 | At&T Corp. | Method and apparatus for adaptive stripe based patch matching for depth estimation |
US20020063709A1 (en) | 1998-05-13 | 2002-05-30 | Scott Gilbert | Panoramic movie which utilizes a series of captured panoramic images to display movement as observed by a viewer looking in a selected direction |
US6504571B1 (en) | 1998-05-18 | 2003-01-07 | International Business Machines Corporation | System and methods for querying digital image archives using recorded parameters |
US6253201B1 (en) | 1998-06-23 | 2001-06-26 | Philips Electronics North America Corporation | Scalable solution for image retrieval |
US6345274B1 (en) | 1998-06-29 | 2002-02-05 | Eastman Kodak Company | Method and computer program product for subjective image content similarity-based retrieval |
US6223171B1 (en) | 1998-08-25 | 2001-04-24 | Microsoft Corporation | What-if index analysis utility for database systems |
US20010042085A1 (en) | 1998-09-30 | 2001-11-15 | Mark Peairs | Automatic document classification using text and images |
US6964374B1 (en) | 1998-10-02 | 2005-11-15 | Lucent Technologies Inc. | Retrieval and manipulation of electronically stored information via pointers embedded in the associated printed material |
US20010043741A1 (en) | 1998-10-13 | 2001-11-22 | Xerox Corporation | Automatic extraction of text regions and region borders for an electronic work surface |
US6192157B1 (en) | 1998-10-27 | 2001-02-20 | Hewlett-Packard Company | Modifications of postscript adaptive data compression (ADC) for 3 plane, 8 bit color images, JPEG lossy compression, and variable Q factors |
US6363381B1 (en) | 1998-11-03 | 2002-03-26 | Ricoh Co., Ltd. | Compressed document matching |
JP2000165645A (en) | 1998-11-25 | 2000-06-16 | Canon Inc | Peripheral device, its control method and storage medium |
US6301386B1 (en) | 1998-12-09 | 2001-10-09 | Ncr Corporation | Methods and apparatus for gray image based text identification |
US6753883B2 (en) | 1999-01-25 | 2004-06-22 | Airclic Inc. | Printed medium activated interactive communication of multimedia information, including advertising |
US6448979B1 (en) | 1999-01-25 | 2002-09-10 | Airclic, Inc. | Printed medium activated interactive communication of multimedia information, including advertising |
US6411953B1 (en) | 1999-01-25 | 2002-06-25 | Lucent Technologies Inc. | Retrieval and matching of color patterns based on a predetermined vocabulary and grammar |
US6922699B2 (en) | 1999-01-26 | 2005-07-26 | Xerox Corporation | System and method for quantitatively representing data objects in vector space |
CN1245935C (en) | 1999-02-05 | 2006-03-22 | 宝洁公司 | Photostable UV protection compositions |
US6980962B1 (en) | 1999-03-02 | 2005-12-27 | Quixtar Investments, Inc. | Electronic commerce transactions within a marketing system that may contain a membership buying opportunity |
US7136093B1 (en) * | 1999-03-23 | 2006-11-14 | Canon Kabushiki Kaisha | Information presenting apparatus, operation processing method therefor, storage medium storing program for executing operation processing |
US20040042667A1 (en) | 1999-04-08 | 2004-03-04 | Dar-Shyang Lee | Extracting information from symbolically compressed document images |
US6791605B1 (en) | 1999-05-07 | 2004-09-14 | Eastman Kodak Company | Image capture and printing device |
US6397213B1 (en) | 1999-05-12 | 2002-05-28 | Ricoh Company Ltd. | Search and retrieval using document decomposition |
US7406214B2 (en) | 1999-05-19 | 2008-07-29 | Digimarc Corporation | Methods and devices employing optical sensors and/or steganography |
US7174031B2 (en) | 1999-05-19 | 2007-02-06 | Digimarc Corporation | Methods for using wireless phones having optical capabilities |
US20020159640A1 (en) | 1999-07-02 | 2002-10-31 | Philips Electronics North America Corporation | Meta-descriptor for multimedia information |
US20040027604A1 (en) | 1999-08-05 | 2004-02-12 | Jeran Paul L. | Methods of document management, methods of automated document tracking, document tracking methods, and document tracking systems |
US6408257B1 (en) | 1999-08-31 | 2002-06-18 | Xerox Corporation | Augmented-reality display method and system |
US6732915B1 (en) | 1999-09-10 | 2004-05-11 | Fuji Xerox Co., Ltd. | Systems and methods for controlling a presentation using physical objects |
US7613686B2 (en) | 1999-09-13 | 2009-11-03 | Microsoft Corporation | Image retrieval based on relevance feedback |
US6751343B1 (en) | 1999-09-20 | 2004-06-15 | Ut-Battelle, Llc | Method for indexing and retrieving manufacturing-specific digital imagery based on image content |
US20070053513A1 (en) | 1999-10-05 | 2007-03-08 | Hoffberg Steven M | Intelligent electronic appliance system and method |
US7134095B1 (en) | 1999-10-20 | 2006-11-07 | Gateway, Inc. | Simulated three-dimensional navigational menu system |
US20040199531A1 (en) | 1999-12-01 | 2004-10-07 | Konan Technology Inc. | Content-based image retrieval system and method for retrieving image using the same |
US20040233235A1 (en) | 1999-12-07 | 2004-11-25 | Microsoft Corporation | Computer user interface architecture that saves a user's non-linear navigation history and intelligently maintains that history |
US7458014B1 (en) | 1999-12-07 | 2008-11-25 | Microsoft Corporation | Computer user interface architecture wherein both content and user interface are composed of documents with links |
US7185274B1 (en) | 1999-12-07 | 2007-02-27 | Microsoft Corporation | Computer user interface architecture wherein users interact with both content and user interface by activating links |
US7359094B1 (en) | 1999-12-15 | 2008-04-15 | Fuji Xerox Co., Ltd. | Image processing apparatus and image forming medium |
US6804659B1 (en) | 2000-01-14 | 2004-10-12 | Ricoh Company Ltd. | Content based web advertising |
US20020073236A1 (en) | 2000-01-14 | 2002-06-13 | Helgeson Christopher S. | Method and apparatus for managing data exchange among systems in a network |
JP2001211359A (en) | 2000-01-25 | 2001-08-03 | Olympus Optical Co Ltd | Electronic camera |
US20010042030A1 (en) | 2000-02-15 | 2001-11-15 | Sadao Ito | Information providing system for providing information about parts |
US20020054059A1 (en) | 2000-02-18 | 2002-05-09 | B.A. Schneiderman | Methods for the electronic annotation, retrieval, and use of electronic images |
JP2001230916A (en) | 2000-02-18 | 2001-08-24 | Ricoh Co Ltd | File system |
US6766363B1 (en) | 2000-02-28 | 2004-07-20 | Barpoint.Com, Inc. | System and method of linking items in audio, visual, and printed media to related information stored on an electronic network using a mobile device |
US6859909B1 (en) | 2000-03-07 | 2005-02-22 | Microsoft Corporation | System and method for annotating web-based documents |
US6865302B2 (en) | 2000-03-16 | 2005-03-08 | The Regents Of The University Of California | Perception-based image retrieval |
US7206820B1 (en) | 2000-03-18 | 2007-04-17 | Digimarc Corporation | System for linking from object to remote resource |
US20010024514A1 (en) | 2000-03-22 | 2001-09-27 | Shinichi Matsunaga | Image processing device, singular spot detection method, and recording medium upon which singular spot detection program is recorded |
US7031965B1 (en) | 2000-03-23 | 2006-04-18 | Mitsubishi Denki Kabushiki Kaisha | Image retrieving and delivering system and image retrieving and delivering method |
US6813381B2 (en) | 2000-03-30 | 2004-11-02 | Glory Ltd. | Method and apparatus for identification of documents, and computer product |
EP1662064A1 (en) | 2000-04-24 | 2006-05-31 | Hunter Douglas Inc. | Compressible structural panel |
US6981224B1 (en) | 2000-05-05 | 2005-12-27 | Point Roll, Inc. | System and method for conditionally associating a plurality of enabled images with a predetermined image |
US20040198396A1 (en) | 2000-05-08 | 2004-10-07 | Owest Communications International Inc. | Location based messaging method and system |
US7683933B2 (en) * | 2000-05-19 | 2010-03-23 | Canon Kabushiki Kaisha | Image display apparatus, image display system, and image display method |
US20010049700A1 (en) | 2000-05-26 | 2001-12-06 | Shinobu Ichikura | Information processing apparatus, information processing method and storage medium |
US20020191003A1 (en) | 2000-08-09 | 2002-12-19 | Hobgood Andrew W. | Method for using a motorized camera mount for tracking in augmented reality |
US7337175B2 (en) | 2000-08-14 | 2008-02-26 | Alcatel | Method of storing data in a multimedia file using relative timebases |
US20020093538A1 (en) | 2000-08-22 | 2002-07-18 | Bruce Carlin | Network-linked interactive three-dimensional composition and display of saleable objects in situ in viewer-selected scenes for purposes of object promotion and procurement, and generation of object advertisements |
US7062722B1 (en) * | 2000-08-22 | 2006-06-13 | Bruce Carlin | Network-linked interactive three-dimensional composition and display of saleable objects in situ in viewer-selected scenes for purposes of promotion and procurement |
US20030187886A1 (en) | 2000-09-01 | 2003-10-02 | Hull Jonathan J. | Method and apparatus for simultaneous highlighting of a physical version of a document and an electronic version of a document |
US6405172B1 (en) | 2000-09-09 | 2002-06-11 | Mailcode Inc. | Voice-enabled directory look-up based on recognized spoken initial characters |
US20020038430A1 (en) | 2000-09-13 | 2002-03-28 | Charles Edwards | System and method of data collection, processing, analysis, and annotation for monitoring cyber-threats and the notification thereof to subscribers |
US6799201B1 (en) | 2000-09-19 | 2004-09-28 | Motorola, Inc. | Remotely configurable multimedia entertainment and information system for vehicles |
US6842755B2 (en) | 2000-09-25 | 2005-01-11 | Divine Technology Ventures | System and method for automatic retrieval of structured online documents |
US20020052872A1 (en) | 2000-10-27 | 2002-05-02 | Fuji Xerox Co., Ltd | Electronic information management server, electronic information management client, electronic information management method and recording medium for recording electronic information management program |
US6940491B2 (en) | 2000-10-27 | 2005-09-06 | International Business Machines Corporation | Method and system for generating hyperlinked physical copies of hyperlinked electronic documents |
US20050114325A1 (en) | 2000-10-30 | 2005-05-26 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20020102966A1 (en) | 2000-11-06 | 2002-08-01 | Lev Tsvi H. | Object identification method for portable devices |
US20060002607A1 (en) | 2000-11-06 | 2006-01-05 | Evryx Technologies, Inc. | Use of image-derived information as search criteria for internet and other search engines |
US6963358B2 (en) | 2000-11-10 | 2005-11-08 | Dropfire, Inc. | Wireless digital camera adapter and systems and methods related thereto and for use with such an adapter |
US20040017482A1 (en) | 2000-11-17 | 2004-01-29 | Jacob Weitman | Application for a mobile digital camera, that distinguish between text-, and image-information in an image |
US20020118379A1 (en) | 2000-12-18 | 2002-08-29 | Amit Chakraborty | System and user interface supporting user navigation of multimedia data file content |
US7013309B2 (en) | 2000-12-18 | 2006-03-14 | Siemens Corporate Research | Method and apparatus for extracting anchorable information units from complex PDF documents |
US20020157028A1 (en) | 2000-12-19 | 2002-10-24 | Fuji Xerox Co., Ltd. | Internet facsimile and control method thereof and communication instruction terminal |
US20040221244A1 (en) | 2000-12-20 | 2004-11-04 | Eastman Kodak Company | Method and apparatus for producing digital images with embedded image capture location icons |
US7092953B1 (en) | 2000-12-28 | 2006-08-15 | Rightlsline, Inc. | Apparatus and methods for intellectual property database navigation |
EP1229496A2 (en) | 2001-02-02 | 2002-08-07 | Matsushita Electric Industrial Co., Ltd. | Information terminal apparatus and authenticating system |
US7366979B2 (en) | 2001-03-09 | 2008-04-29 | Copernicus Investments, Llc | Method and apparatus for annotating a document |
US7392287B2 (en) | 2001-03-27 | 2008-06-24 | Hemisphere Ii Investment Lp | Method and apparatus for sharing information using a handheld device |
US20020191848A1 (en) | 2001-03-29 | 2002-12-19 | The Boeing Company | Method, computer program product, and system for performing automated text recognition and text search within a graphic file |
US20020146176A1 (en) | 2001-04-09 | 2002-10-10 | Meyers Gary Elliott | Method for representing and comparing digital images |
US20050086188A1 (en) | 2001-04-11 | 2005-04-21 | Hillis Daniel W. | Knowledge web |
US20020154148A1 (en) | 2001-04-20 | 2002-10-24 | Aiichi Inoue | Image displaying method |
US20050160115A1 (en) | 2001-05-22 | 2005-07-21 | Microsoft Corporation | Document imaging and indexing system |
US20020194264A1 (en) | 2001-05-31 | 2002-12-19 | Pioneer Corporation | Digital information distribution apparatus and method |
US20040238621A1 (en) | 2001-07-10 | 2004-12-02 | American Express Travel Related Services Company, Inc. | Method and system for fingerprint biometrics on a fob |
US20030030828A1 (en) * | 2001-07-13 | 2003-02-13 | Pau Soler | Spot color application in printer device |
US7239402B2 (en) * | 2001-07-13 | 2007-07-03 | Hewlett-Packard Development Company, L.P. | Spot color application in printer device |
US20030025714A1 (en) | 2001-07-16 | 2003-02-06 | Ebersole John Franklin | Method to view unseen atmospheric phenomenon using augmented reality |
US20030110130A1 (en) | 2001-07-20 | 2003-06-12 | International Business Machines Corporation | Method and system for delivering encrypted content with associated geographical-based advertisements |
US20030030835A1 (en) | 2001-08-07 | 2003-02-13 | Konica Corporation | Printing control apparatus and a method of charging for printing |
US20050190273A1 (en) | 2001-10-09 | 2005-09-01 | Microsoft Corporation | System and method for exchanging images |
US20030126147A1 (en) | 2001-10-12 | 2003-07-03 | Hassane Essafi | Method and a system for managing multimedia databases |
US20040201706A1 (en) | 2001-10-26 | 2004-10-14 | Katsutoshi Shimizu | Corrected image generating apparatus and corrected image generating program storage medium |
US20030122922A1 (en) | 2001-11-26 | 2003-07-03 | Saffer Kevin D. | Video e-mail system and associated method |
US20030098877A1 (en) | 2001-11-29 | 2003-05-29 | International Business Machines Corporation | Method and system for appending information to graphical files stored in specific graphical file formats |
US20030190094A1 (en) | 2001-11-29 | 2003-10-09 | Kazuaki Yokota | Document identification device, document definition method and document identification method |
US20080296362A1 (en) | 2001-11-30 | 2008-12-04 | International Barcode Corporation | Method for identifying and authenticating goods using codes, bar codes and radio frequency identification |
US20030110216A1 (en) | 2001-12-06 | 2003-06-12 | Mathias Althin | Mobile guide communications system |
US7509386B2 (en) | 2001-12-11 | 2009-03-24 | Sony Corporation | Chat system displaying a link arrow directed from a hyperlink to content of an associated attachment file |
US20030112930A1 (en) | 2001-12-18 | 2003-06-19 | Bosik Barry S. | Call management system responsive to network presence |
US20030128375A1 (en) | 2002-01-07 | 2003-07-10 | Ruhl Jan Matthias | Systems and methods for authenticating and verifying documents |
US20030152293A1 (en) | 2002-01-24 | 2003-08-14 | Joel Bresler | Method and system for locating position in printed texts and delivering multimedia information |
US20030142106A1 (en) | 2002-01-25 | 2003-07-31 | Xerox Corporation | Method and apparatus to convert bitmapped images for use in a structured text/graphics editor |
US20030151674A1 (en) | 2002-02-12 | 2003-08-14 | Qian Lin | Method and system for assessing the photo quality of a captured image in a digital still camera |
US7167574B2 (en) | 2002-03-14 | 2007-01-23 | Seiko Epson Corporation | Method and apparatus for content-based image copy detection |
US7251689B2 (en) | 2002-03-27 | 2007-07-31 | International Business Machines Corporation | Managing storage resources in decentralized networks |
US7421153B1 (en) | 2002-04-05 | 2008-09-02 | Bank Of America Corporation | Image and data processing system |
US20030193530A1 (en) | 2002-04-11 | 2003-10-16 | Blackman Jeffrey R. | Barcode printing module |
US20060143176A1 (en) | 2002-04-15 | 2006-06-29 | International Business Machines Corporation | System and method for measuring image similarity based on semantic meaning |
US20040036679A1 (en) | 2002-05-02 | 2004-02-26 | Emerson Harry E. | Computer system providing a visual indication when typing in caps lock mode |
US20030212585A1 (en) | 2002-05-09 | 2003-11-13 | Kabushiki Kaisha Toshiba | Idea drawing support method and program product therefor |
US20030229857A1 (en) | 2002-06-05 | 2003-12-11 | Fuji Xerox Co., Ltd. | Apparatus, method, and computer program product for document manipulation which embeds information in document data |
US7240279B1 (en) | 2002-06-19 | 2007-07-03 | Microsoft Corporation | XML patterns language |
US20060112092A1 (en) | 2002-08-09 | 2006-05-25 | Bell Canada | Content-based image retrieval method |
US20040133582A1 (en) | 2002-10-11 | 2004-07-08 | Howard James V. | Systems and methods for recognition of individuals using multiple biometric searches |
US7249123B2 (en) | 2002-10-31 | 2007-07-24 | International Business Machines Corporation | System and method for building social networks based on activity around shared virtual objects |
US20040102898A1 (en) | 2002-11-23 | 2004-05-27 | Tatsuo Yokota | Navigation method and system for indicating area-specific traffic information |
US20040215689A1 (en) | 2003-01-09 | 2004-10-28 | Dooley Michael J. | Computer and vision-based augmented interaction in the use of printed media |
US20040139391A1 (en) | 2003-01-15 | 2004-07-15 | Xerox Corporation | Integration of handwritten annotations into an electronic original |
US20040143644A1 (en) | 2003-01-21 | 2004-07-22 | Nec Laboratories America, Inc. | Meta-search engine architecture |
US20040243514A1 (en) | 2003-01-23 | 2004-12-02 | John Wankmueller | System and method for secure telephone and computer transactions using voice authentication |
WO2004072897A2 (en) | 2003-02-06 | 2004-08-26 | Centerframe, L.L.C. | Obtaining person-specific images in a public venue |
US7310769B1 (en) | 2003-03-12 | 2007-12-18 | Adobe Systems Incorporated | Text encoding using dummy font |
US20060206335A1 (en) | 2003-03-17 | 2006-09-14 | Eric Thelen | Method for remote control of an audio device |
US7882113B2 (en) | 2003-03-28 | 2011-02-01 | International Business Machines Corporation | Method, apparatus, and system for formatting time data to improve processing in a sort utility |
US20040205347A1 (en) | 2003-04-11 | 2004-10-14 | Ricoh Company, Ltd. | Techniques for accessing information captured during a presentation using a paper document handout for the presentation |
US7236632B2 (en) | 2003-04-11 | 2007-06-26 | Ricoh Company, Ltd. | Automated techniques for comparing contents of images |
US7593961B2 (en) | 2003-04-30 | 2009-09-22 | Canon Kabushiki Kaisha | Information processing apparatus for retrieving image data similar to an entered image |
US6993573B2 (en) | 2003-06-06 | 2006-01-31 | Neomedia Technologies, Inc. | Automatic access of internet content with a camera-enabled cell phone |
US7587681B2 (en) * | 2003-06-16 | 2009-09-08 | Sony Computer Entertainment Inc. | Method and apparatus for presenting information |
US20040260680A1 (en) | 2003-06-19 | 2004-12-23 | International Business Machines Corporation | Personalized indexing and searching for information in a distributed data processing system |
US20070271247A1 (en) | 2003-06-19 | 2007-11-22 | Best Steven F | Personalized Indexing And Searching For Information In A Distributed Data Processing System |
US20040260625A1 (en) | 2003-06-20 | 2004-12-23 | Yoshinori Usami | Digital photo print order acceptance method and digital image management method |
US7310779B2 (en) | 2003-06-26 | 2007-12-18 | International Business Machines Corporation | Method for creating and selecting active regions on physical documents |
US7489415B2 (en) | 2003-07-16 | 2009-02-10 | Ricoh Company, Ltd. | Document processing system |
US20050012960A1 (en) | 2003-07-18 | 2005-01-20 | Sharp Laboratories Of America, Inc. | Print content system and method for providing document control |
US20060286951A1 (en) | 2003-08-25 | 2006-12-21 | Fujitsu Ten Limited | Music information,updating system, music information broadcasting apparatus, terminal apparatus having music information updating function, music information updating method, music information broadcasting method, and music information updating method of terminal apparatus |
US7305435B2 (en) | 2003-08-25 | 2007-12-04 | Nokia Corporation | Internet access via smartphone camera |
US20060082438A1 (en) | 2003-09-05 | 2006-04-20 | Bazakos Michael E | Distributed stand-off verification and face recognition systems (FRS) |
US20060085735A1 (en) | 2003-09-18 | 2006-04-20 | Fujitsu Limited | Annotation management system, annotation managing method, document transformation server, document transformation program, and electronic document attachment program |
US20050080693A1 (en) | 2003-10-14 | 2005-04-14 | Foss Sheldon H. | Point-of-sale customer identification system |
US20050086224A1 (en) | 2003-10-15 | 2005-04-21 | Xerox Corporation | System and method for computing a measure of similarity between documents |
US20050089246A1 (en) | 2003-10-27 | 2005-04-28 | Huitao Luo | Assessing image quality |
WO2005043270A2 (en) | 2003-11-04 | 2005-05-12 | Nokia Corporation | System and method for registering attendance of entities associated with content creation |
US20050100219A1 (en) | 2003-11-10 | 2005-05-12 | Kathrin Berkner | Features for retrieval and similarity matching of documents from the JPEG 2000-compressed domain |
US20050125390A1 (en) | 2003-12-03 | 2005-06-09 | Oliver Hurst-Hiller | Automated satisfaction measurement for web search |
US7653238B2 (en) | 2003-12-05 | 2010-01-26 | British Telecommunications Plc | Image filtering based on comparison of pixel groups |
US20050129293A1 (en) | 2003-12-11 | 2005-06-16 | Tinku Acharya | Fingerprint minutiae matching using scoring techniques |
US20050160258A1 (en) | 2003-12-11 | 2005-07-21 | Bioobservation Systems Limited | Detecting objectionable content in displayed images |
US20050185225A1 (en) | 2003-12-12 | 2005-08-25 | Brawn Dennis E. | Methods and apparatus for imaging documents |
US20050135483A1 (en) | 2003-12-23 | 2005-06-23 | Genesis Microchip Inc. | Temporal motion vector filtering |
US20050169520A1 (en) | 2003-12-29 | 2005-08-04 | Canon Kabushiki Kaisha | Detecting human faces and detecting red eyes |
US20050198095A1 (en) | 2003-12-31 | 2005-09-08 | Kavin Du | System and method for obtaining information relating to an item of commerce using a portable imaging device |
EP1555626A2 (en) | 2004-01-15 | 2005-07-20 | Microsoft Corporation | Image-based indexing and retrieval of text documents |
US20050165747A1 (en) | 2004-01-15 | 2005-07-28 | Bargeron David M. | Image-based document indexing and retrieval |
US7872669B2 (en) | 2004-01-22 | 2011-01-18 | Massachusetts Institute Of Technology | Photo-based mobile deixis system and related techniques |
US20050165784A1 (en) | 2004-01-23 | 2005-07-28 | Garrison Gomez | System and method to store and retrieve identifier associated information content |
US7424541B2 (en) | 2004-02-09 | 2008-09-09 | Proxpro, Inc. | Method and computer system for matching mobile device users for business and social networking |
US20050190972A1 (en) | 2004-02-11 | 2005-09-01 | Thomas Graham A. | System and method for position determination |
US7593605B2 (en) | 2004-02-15 | 2009-09-22 | Exbiblio B.V. | Data capture from rendered documents using handheld device |
US7707039B2 (en) | 2004-02-15 | 2010-04-27 | Exbiblio B.V. | Automatic modification of web pages |
US20060023945A1 (en) | 2004-02-15 | 2006-02-02 | King Martin T | Search engines and systems with handheld document data capture devices |
US20060026140A1 (en) | 2004-02-15 | 2006-02-02 | King Martin T | Content access with handheld document data capture devices |
US7437023B2 (en) | 2004-02-15 | 2008-10-14 | Exbiblio B.V. | Methods, systems and computer program products for data gathering in a digital and hard copy document environment |
US20050234851A1 (en) | 2004-02-15 | 2005-10-20 | King Martin T | Automatic modification of web pages |
US7421155B2 (en) | 2004-02-15 | 2008-09-02 | Exbiblio B.V. | Archive of text captures from rendered documents |
US20060061806A1 (en) | 2004-02-15 | 2006-03-23 | King Martin T | Information gathering system and method |
US20050182773A1 (en) | 2004-02-18 | 2005-08-18 | Feinsmith Jason B. | Machine-implemented activity management system using asynchronously shared activity data objects and journal data items |
US20060012677A1 (en) | 2004-02-20 | 2006-01-19 | Neven Hartmut Sr | Image-based search engine for mobile phones with camera |
US20060240862A1 (en) | 2004-02-20 | 2006-10-26 | Hartmut Neven | Mobile image-based information retrieval system |
US20050185060A1 (en) | 2004-02-20 | 2005-08-25 | Neven Hartmut Sr. | Image base inquiry system for search engines for mobile telephones with integrated camera |
US7386789B2 (en) | 2004-02-27 | 2008-06-10 | Hewlett-Packard Development Company, L.P. | Method for determining logical components of a document |
US20050216257A1 (en) | 2004-03-18 | 2005-09-29 | Pioneer Corporation | Sound information reproducing apparatus and method of preparing keywords of music data |
US20050288859A1 (en) | 2004-03-23 | 2005-12-29 | Golding Andrew R | Visually-oriented driving directions in digital mapping system |
JP2005286395A (en) | 2004-03-26 | 2005-10-13 | Fuji Photo Film Co Ltd | Image processing system and method therefor |
US20100013615A1 (en) | 2004-03-31 | 2010-01-21 | Carnegie Mellon University | Obstacle detection having enhanced classification |
US20060053097A1 (en) | 2004-04-01 | 2006-03-09 | King Martin T | Searching and accessing documents on private networks for use with captures from rendered documents |
US20060041605A1 (en) | 2004-04-01 | 2006-02-23 | King Martin T | Determining actions involving captured information and electronic content associated with rendered documents |
US20050261990A1 (en) | 2004-04-16 | 2005-11-24 | Russell Gocht | Mobile query system and method based on visual cues |
US20050240381A1 (en) | 2004-04-26 | 2005-10-27 | Pjm Interconnection, Llc. | Systems and methods for analysis of a commodity transmission network |
US20050273812A1 (en) | 2004-06-02 | 2005-12-08 | Kabushiki Kaisha Toshiba | User profile editing apparatus, method and program |
US7150399B2 (en) | 2004-06-09 | 2006-12-19 | Ricoh Co., Ltd. | Embedding barcode data in an auxiliary field of an image file |
US20050289182A1 (en) | 2004-06-15 | 2005-12-29 | Sand Hill Systems Inc. | Document management system with enhanced intelligent document recognition capabilities |
US7746376B2 (en) | 2004-06-16 | 2010-06-29 | Felipe Mendoza | Method and apparatus for accessing multi-dimensional mapping and information |
US20050288911A1 (en) | 2004-06-28 | 2005-12-29 | Porikli Fatih M | Hidden markov model based object tracking and similarity metrics |
US20050289447A1 (en) | 2004-06-29 | 2005-12-29 | The Boeing Company | Systems and methods for generating and storing referential links in a database |
US20060104515A1 (en) | 2004-07-19 | 2006-05-18 | King Martin T | Automatic modification of WEB pages |
US20060020630A1 (en) | 2004-07-23 | 2006-01-26 | Stager Reed R | Facial database methods and systems |
US20060056697A1 (en) | 2004-08-13 | 2006-03-16 | Fujitsu Limited | Degraded character image generation method and apparatus |
JP2006053568A (en) | 2004-08-13 | 2006-02-23 | Fujitsu Ltd | Deteriorated character image generation program, method and apparatus |
JP2006059351A (en) | 2004-08-18 | 2006-03-02 | Fujitsu Ltd | Deterioration dictionary generation program, method and device |
US20060056696A1 (en) | 2004-08-18 | 2006-03-16 | Fujitsu Limited | Degraded dictionary generation method and apparatus |
US20060048059A1 (en) | 2004-08-26 | 2006-03-02 | Henry Etkin | System and method for dynamically generating, maintaining, and growing an online social network |
US20060043188A1 (en) | 2004-08-27 | 2006-03-02 | Gregg Kricorissian | Imaging method and apparatus for object identification |
US20060070120A1 (en) | 2004-09-02 | 2006-03-30 | Brother Kogyo Kabushiki Kaisha | File transmitting device and multi function device |
US20060074828A1 (en) | 2004-09-14 | 2006-04-06 | Heumann John M | Methods and apparatus for detecting temporal process variation and for managing and predicting performance of automatic classifiers |
US20060262976A1 (en) | 2004-10-01 | 2006-11-23 | Hart Peter E | Method and System for Multi-Tier Image Matching in a Mixed Media Environment |
US20060262352A1 (en) | 2004-10-01 | 2006-11-23 | Hull Jonathan J | Method and system for image matching in a mixed media environment |
US20060262962A1 (en) | 2004-10-01 | 2006-11-23 | Hull Jonathan J | Method And System For Position-Based Image Matching In A Mixed Media Environment |
US20060085477A1 (en) | 2004-10-01 | 2006-04-20 | Ricoh Company, Ltd. | Techniques for retrieving documents using an image capture device |
US20060285172A1 (en) | 2004-10-01 | 2006-12-21 | Hull Jonathan J | Method And System For Document Fingerprint Matching In A Mixed Media Environment |
US20060285772A1 (en) | 2004-10-01 | 2006-12-21 | Hull Jonathan J | System and methods for creation and use of a mixed media environment |
US20060116555A1 (en) | 2004-11-18 | 2006-06-01 | Honeywell International Inc. | Controlled environment thermal image detection system and methods regarding same |
US20060119880A1 (en) | 2004-12-02 | 2006-06-08 | Dandekar Shree A | Local photo printing |
US20060122983A1 (en) | 2004-12-03 | 2006-06-08 | King Martin T | Locating electronic instances of documents based on rendered instances, document fragment digest generation, and digest based document fragment determination |
US7263205B2 (en) | 2004-12-06 | 2007-08-28 | Dspv, Ltd. | System and method of generic symbol recognition and user authentication using a communication device with imaging capabilities |
US20060123347A1 (en) | 2004-12-06 | 2006-06-08 | Joe Hewitt | Managing and collaborating with digital content using a dynamic user interface |
US20060140475A1 (en) | 2004-12-15 | 2006-06-29 | Chin Frances M | Network support for corrupting images captured at a prohibited location |
US20060150079A1 (en) | 2004-12-17 | 2006-07-06 | International Business Machines Corporation | Method for associating annotations with document families |
US20060140614A1 (en) | 2004-12-28 | 2006-06-29 | Samsung Electronic Co., Ltd. | Apparatus, medium, and method for photographing based on face detection |
US20070150466A1 (en) | 2004-12-29 | 2007-06-28 | Scott Brave | Method and apparatus for suggesting/disambiguation query terms based upon usage patterns observed |
US7930292B2 (en) | 2005-01-28 | 2011-04-19 | Canon Kabushiki Kaisha | Information processing apparatus and control method thereof |
JP2006215756A (en) | 2005-02-02 | 2006-08-17 | Dainippon Ink & Chem Inc | Image processing apparatus, image processing method, and program for the same |
US20060190812A1 (en) | 2005-02-22 | 2006-08-24 | Geovector Corporation | Imaging systems including hyperlink associations |
US7567262B1 (en) | 2005-02-25 | 2009-07-28 | IDV Solutions LLC | Hybrid graphics for interactive reporting |
US7644078B2 (en) | 2005-02-28 | 2010-01-05 | Gm Global Technology Operations, Inc. | System and method for mining of temporal data |
US20060200480A1 (en) | 2005-03-01 | 2006-09-07 | Harris David N | System and method for using product identifiers |
US20060218225A1 (en) | 2005-03-28 | 2006-09-28 | Hee Voon George H | Device for sharing social network information among users over a network |
US7546524B1 (en) | 2005-03-30 | 2009-06-09 | Amazon Technologies, Inc. | Electronic input device, system, and method using human-comprehensible content to automatically correlate an annotation of a paper document with a digital version of the document |
US7765231B2 (en) | 2005-04-08 | 2010-07-27 | Rathus Spencer A | System and method for accessing electronic data via an image search engine |
US20060227992A1 (en) | 2005-04-08 | 2006-10-12 | Rathus Spencer A | System and method for accessing electronic data via an image search engine |
US7585224B2 (en) * | 2005-04-28 | 2009-09-08 | Nintendo Co., Ltd. | Storage medium having game program stored therein and game apparatus |
US20060251292A1 (en) | 2005-05-09 | 2006-11-09 | Salih Burak Gokturk | System and method for recognizing objects from images and identifying relevancy amongst images and information |
US20060253439A1 (en) | 2005-05-09 | 2006-11-09 | Liwei Ren | Matching engine for querying relevant documents |
US20060253491A1 (en) | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling search and retrieval from image files based on recognized information |
US20060251339A1 (en) | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling the use of captured images through recognition |
US7450760B2 (en) | 2005-05-18 | 2008-11-11 | Scanr, Inc. | System and method for capturing and processing business data |
US7676767B2 (en) * | 2005-06-15 | 2010-03-09 | Microsoft Corporation | Peel back user interface to show hidden functions |
US20060285755A1 (en) * | 2005-06-16 | 2006-12-21 | Strider Labs, Inc. | System and method for recognition in 2D images using 3D class models |
US20070036469A1 (en) | 2005-06-20 | 2007-02-15 | Samsung Electronics Co., Ltd. | Method and system for providing image-related information to user, and mobile terminal therefor |
US20060294049A1 (en) | 2005-06-27 | 2006-12-28 | Microsoft Corporation | Back-off mechanism for search |
US20070003166A1 (en) | 2005-06-30 | 2007-01-04 | Kathrin Berkner | White space graphs and trees for content-adaptive scaling of document images |
EP1783681A1 (en) | 2005-06-30 | 2007-05-09 | Olympus Corporation | Retrieval system and retrieval method |
US20070003147A1 (en) | 2005-07-01 | 2007-01-04 | Microsoft Corporation | Grammatical parsing of document visual structures |
US20070019261A1 (en) | 2005-07-21 | 2007-01-25 | Yu-Chung Chu | Method and system for distributing consumer photograph-based advertisement and/or information |
US20070041668A1 (en) | 2005-07-28 | 2007-02-22 | Canon Kabushiki Kaisha | Search apparatus and search method |
US20070047819A1 (en) | 2005-08-23 | 2007-03-01 | Hull Jonathan J | Data organization and access for mixed media document system |
US20070052997A1 (en) * | 2005-08-23 | 2007-03-08 | Hull Jonathan J | System and methods for portable device for mixed media system |
US7812986B2 (en) | 2005-08-23 | 2010-10-12 | Ricoh Co. Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US20070165904A1 (en) * | 2005-08-23 | 2007-07-19 | Nudd Geoffrey H | System and Method for Using Individualized Mixed Document |
US20070175998A1 (en) | 2005-09-01 | 2007-08-02 | Lev Zvi H | System and method for reliable content access using a cellular/wireless device with imaging capabilities |
US20070076922A1 (en) | 2005-09-30 | 2007-04-05 | Sony United Kingdom Limited | Object detection |
US20070106721A1 (en) | 2005-11-04 | 2007-05-10 | Philipp Schloter | Scalable visual search system simplifying access to network and device functionality |
US20070115373A1 (en) | 2005-11-22 | 2007-05-24 | Eastman Kodak Company | Location based image classification with map segmentation |
WO2007073347A1 (en) | 2005-12-19 | 2007-06-28 | Agency For Science, Technology And Research | Annotation of video footage and personalised video generation |
US7761436B2 (en) | 2006-01-03 | 2010-07-20 | Yahoo! Inc. | Apparatus and method for controlling content access based on shared annotations for annotated users in a folksonomy scheme |
US20070174269A1 (en) | 2006-01-23 | 2007-07-26 | Microsoft Corporation | Generating clusters of images for search results |
US20080310717A1 (en) | 2006-02-01 | 2008-12-18 | Motorola, Inc. | Apparatus and Method for Image Labeling |
US7463270B2 (en) * | 2006-02-10 | 2008-12-09 | Microsoft Corporation | Physical-virtual interpolation |
US7680850B2 (en) | 2006-03-31 | 2010-03-16 | Fujitsu Limited | Computer-readable recording medium storing information search program, information search method, and information search system |
US20070233613A1 (en) | 2006-03-31 | 2007-10-04 | Ricoh Company, Ltd. | Techniques for using media keys |
US7668405B2 (en) | 2006-04-07 | 2010-02-23 | Eastman Kodak Company | Forming connections between image collections |
US20070276845A1 (en) | 2006-05-12 | 2007-11-29 | Tele Atlas North America, Inc. | Locality indexes and method for indexing localities |
US20080010605A1 (en) | 2006-06-12 | 2008-01-10 | Metacarta, Inc. | Systems and methods for generating and correcting location references extracted from text |
US20090067726A1 (en) | 2006-07-31 | 2009-03-12 | Berna Erol | Computation of a recognizability score (quality predictor) for image retrieval |
US7946491B2 (en) | 2006-08-03 | 2011-05-24 | Nokia Corporation | Method, apparatus, and computer program product for providing a camera barcode reader |
US20080071767A1 (en) | 2006-08-25 | 2008-03-20 | Semdirector, Inc. | System and method for measuring the effectiveness of an on-line advertisement campaign |
US20080275881A1 (en) | 2006-09-05 | 2008-11-06 | Gloto Corporation | Real time collaborative on-line multimedia albums |
US7752534B2 (en) | 2006-09-19 | 2010-07-06 | International Business Machines Corporation | Method and apparatus for customizing the display of multidimensional data |
US20080078836A1 (en) | 2006-09-29 | 2008-04-03 | Hiroshi Tomita | Barcode for two-way verification of a document |
US20080106594A1 (en) | 2006-11-07 | 2008-05-08 | The Board Of Trustees Of The Leland Stanford Jr. University | System and method for tagging objects in a panoramic video and associating functions and indexing panoramic images with same |
US20110121069A1 (en) | 2007-01-18 | 2011-05-26 | Target Brands, Inc. | Barcodes with Graphical Elements |
US20080229192A1 (en) | 2007-03-15 | 2008-09-18 | Microsoft Corporation | Interactive image tagging |
US20080267504A1 (en) | 2007-04-24 | 2008-10-30 | Nokia Corporation | Method, device and computer program product for integrating code-based and optical character recognition technologies into a mobile visual search |
WO2008129373A2 (en) | 2007-04-24 | 2008-10-30 | Nokia Corporation | Method, device and computer program product for integrating code-based and optical character recognition technologies into a mobile visual search |
US20090235187A1 (en) | 2007-05-17 | 2009-09-17 | Research In Motion Limited | System and method for content navigation |
US20080288476A1 (en) | 2007-05-17 | 2008-11-20 | Sang-Heun Kim | Method and system for desktop tagging of a web page |
US20100174783A1 (en) | 2007-10-12 | 2010-07-08 | Rony Zarom | System and method for coordinating simultaneous edits of shared digital data |
US20090152357A1 (en) | 2007-12-12 | 2009-06-18 | 3M Innovative Properties Company | Document verification using dynamic document identification framework |
US20100063961A1 (en) | 2008-09-05 | 2010-03-11 | Fotonauts, Inc. | Reverse Tagging of Images in System for Managing and Sharing Digital Images |
Non-Patent Citations (232)
Title |
---|
"Call for Papers: ICAT 2007," 17th International Conference on Artificial Reality and Telexistence, 2007, [Online] [Retrieved on Nov. 4, 2008] Retrieved from the Internet. |
"Call for Papers: ICAT 2007," 17th International Conference on Artificial Reality and Telexistence, 2007, [Online] [Retrieved on Nov. 4, 2008] Retrieved from the Internet<URL:http://www.idemployee.id.tue.nl/g.w.m.rauterberg/conferences/ICAT2007-CfP.pdf>. |
"Mobile Search Engines," Sonera MediaLab, Nov. 15, 2002, pp. 1-12. |
"Mobile Video Managed Service," Vidiator, 2008, [online] [Retrieved on Aug. 29, 2008] Retrieved from the Internet. |
"Mobile Video Managed Service," Vidiator, 2008, [online] [Retrieved on Aug. 29, 2008] Retrieved from the Internet<URL:http://www.vidiator.com/services/managed—mobile—video.aspx>. |
Adobe Acrobat Advanced Elements (for both PC and Mac Computers), 2002, pp. 1-19. |
Aggarwal, M et al, "On Cosine-fourth and Vignetting Effects in Real Lenses," ICCV Proceedings, IEEE, 2001, vol. 1, pp. 472-479, [online] Retrieved from the Internet. |
Aggarwal, M et al, "On Cosine-fourth and Vignetting Effects in Real Lenses," ICCV Proceedings, IEEE, 2001, vol. 1, pp. 472-479, [online] Retrieved from the Internet<URL: http://www.metaverselab.org/classis/635/reading/aggarwal-iccv.pdf>. |
Akenine-Moller, T. et al., "Real-Time Rendering," A.K. Peters, Natick, MA, 2nd Edition, 2002, pp. 70-84. |
Antonacopoulos et al., "Flexible Page Segmentation Using the Background", Proceedings of the IAPR International Conference on Pattern Recognition, Jerusalem, Oct. 9-13, 1994. Conference B: Pattern Recognition and Neural Networks; [Proceedings of the IAPR International Conference on Pattern Recognition], Los Alamitos, IEE, vol. 2, Oct. 9, 1994, pp. 339-344, XP000509905, ISBN: 978-0-8186-6272-0. |
Archive of "Barcodepedia.com-the online barcode database," [online] [Archived by http://archive.org on Jul. 9, 2006; Retrieved on Aug. 18, 2008] Retrieved from the Internet. |
Archive of "Barcodepedia.com—the online barcode database," [online] [Archived by http://archive.org on Jul. 9, 2006; Retrieved on Aug. 18, 2008] Retrieved from the Internet<http://web.archive.org/web/20060709101455/http://en.barcodepedia.com/>. |
Archive of Scanbuy Solutions | Optical Intelligence for your Mobile Devices, Scanbuy® Inc., www.scanbuy.com/website/solutions-summary.htm, [Online] [Archived by http://archive.org on Jun. 19, 2006; Retrieved on Mar. 3, 2009] Retrieved from the Internet. |
Archive of Scanbuy Solutions | Optical Intelligence for your Mobile Devices, Scanbuy® Inc., www.scanbuy.com/website/solutions—summary.htm, [Online] [Archived by http://archive.org on Jun. 19, 2006; Retrieved on Mar. 3, 2009] Retrieved from the Internet<URL:http://web.archive.org/web/20060619172549/http://www.scanbuy.com/website/solutions—su...>. |
Baba, M. et al., "Shadow Removal from a Real Image Based on Shadow Density," Poster at SIGGRAPH2004, Updated Aug. 16, 2004, 4 pages, [online] Retrieved from the Internet. |
Baba, M. et al., "Shadow Removal from a Real Image Based on Shadow Density," Poster at SIGGRAPH2004, Updated Aug. 16, 2004, 4 pages, [online] Retrieved from the Internet<URL:http://www.cv.its.hiroshima-cu.ac.jp/baba/Shadow/poster04-02.pdf>. |
Baird, H., "Document Image Defect Models," In Proc. of IAPR Workshop on Syntactic and Structural Pattern Recognition, Murray Hill, NJ, Jun. 1990, Structured Document Image Analysis, Springer-Verlag, pp. 546-556. |
Baird, H., "The State of the Art of Document Image Degradation Modeling," In Proc. of the 4th IAPR International Workshop on Document Analysis Systems, Rio de Janeiro, Brazil, 2000, pp. 1-16, [online] Retrieved from the Internet. |
Baird, H., "The State of the Art of Document Image Degradation Modeling," In Proc. of the 4th IAPR International Workshop on Document Analysis Systems, Rio de Janeiro, Brazil, 2000, pp. 1-16, [online] Retrieved from the Internet<URL:http://www2.parc.xerox.com/istl/members/baird/das00.pas.gz>. |
Baird, H.S., "Document Image Defect Models and Their Uses," Proc., IAPR 2nd International Conference on Document Analysis and Recognition, Tsukuba Science City, Japan, Oct. 20-22, 1993, 7 pages. |
Barney Smith, E.H. et al., "Text Degradations and OCR Training," International Conference on Document Analysis and Recognition 2005, Seoul, Korea, Aug. 2005, 5 pages, [online] Retrieved from the Internet. |
Barney Smith, E.H. et al., "Text Degradations and OCR Training," International Conference on Document Analysis and Recognition 2005, Seoul, Korea, Aug. 2005, 5 pages, [online] Retrieved from the Internet<URL:http://coen.boisestate.edu/EBarneySmith/Papers/ICDAR05—submit.pdf>. |
Bouget, J., "Camera Calibration Toolbox for Matlab," Online Source, Updated Jul. 24, 2006, 6 pages, [online] Retrieved from the Internet. |
Bouget, J., "Camera Calibration Toolbox for Matlab," Online Source, Updated Jul. 24, 2006, 6 pages, [online] Retrieved from the Internet<URL:http:www.vision.caltech.edu/bougetj/calib—doc/index.html#ref>. |
Boukraa, M. et al., "Tag-Based Vision: Assisting 3D Scene Analysis with Radio-Frequency Tags," Jul. 8, 2002, Proceedings of the Fifth International Conference on Information Fusion, Piscataway, N.J., IEEE, Jul. 8-11, 2002, pp. 412-418. |
Boyd, S., "EE263: Introduction to Linear Dynamical Systems," Online Lecture Notes, Stanford University, Spring Quarter, 2006-2007, Accessed on Sep. 11, 2006, 4 pages, [online] Retrieved from the Internet. |
Boyd, S., "EE263: Introduction to Linear Dynamical Systems," Online Lecture Notes, Stanford University, Spring Quarter, 2006-2007, Accessed on Sep. 11, 2006, 4 pages, [online] Retrieved from the Internet<URL:http://www.standford/edu/class/ee263/#lectures>. |
Brassil, J. et al., "Hiding Information in Document Images," Proc. Conf. Information Sciences and Systems (CISS-95), Mar. 1995, Johns Hopkins University, Baltimore, MD, pp. 482-489. |
Canny, J., "A Computational Approach to Edge Detection," IEEE Transactions on Pattern Analysis and Machine Intelligence, Nov. 1986, pp. 679-714, vol. PAMI-8, No. 6. |
China Patent Office, Office Action for Chinese Patent Application CN200680039376.7, Apr. 28, 2012, 11 pages. |
Constantini, R. et al., "Virtual Sensor Design," Proceedings of the SPIE, vol. 5301, 2004, pp. 408-419, Retrieved from the Internet. |
Constantini, R. et al., "Virtual Sensor Design," Proceedings of the SPIE, vol. 5301, 2004, pp. 408-419, Retrieved from the Internet<URL:http://ivrgwww.epfl.ch/publications/cs04.pdf>. |
Cover, T.M. et al., "Nearest Neighbor Pattern Classification," IEEE Transactions on Information Theory, Jan. 1967, pp. 21-27, vol. IT-13, No. 1. |
Davis, M. et al., "Towards Context-Aware Face Recognition," Proceedings of the13th Annual ACM International Conference on Multimedia, Nov. 6-11, 2005, pp. 483-486, vol. 13. |
Di Stefano, L. et al., "A Simple and Efficient Connected Components Labeling Algorithm,"International Conference on Image Analysis and Processing, 1999, pp. 322-327. |
Doermann, D. et al., "Progress in Camera-Based Document Image Analysis," Proceedings of the Seventh International Conference on Document Analysis and Recognition, ICDAR 2003, 11 pages, [online] Retrieved from the Internet<URL:http://www.cse.salford.ac.uk/prima/ICDAR2003/Papers/0111-keynote-III-doermann-d.pdf>. |
Duda, R. O. et al., "Use of the Hough Transformation to Detect Lines and Curves in Pictures," Communications of the ACM, Jan. 1972, pp. 11-15, vol. 15, No. 1. |
EPO Summons for Oral Proceedings, European Patent Application No. 07015093.3, Sep. 16, 2011, 4 pages. |
Erol, B. et al., "Linking Multimedia Presentations with Their Symbolic Source Documents: Algorithm and Applications," Nov. 2-8, 2003, pp. 498-507, [Online] [Retreived on Oct. 15, 2008] Retrieved from the Internet. |
Erol, B. et al., "Linking Multimedia Presentations with Their Symbolic Source Documents: Algorithm and Applications," Nov. 2-8, 2003, pp. 498-507, [Online] [Retreived on Oct. 15, 2008] Retrieved from the Internet<URL:http://rii.ricoh.com/{hull/pubs/p225—erol.pdf>. |
Erol, B. et al., "Linking Presentation Documents Using Image Analysis," IEEE, Nov. 9-12, 2003, pp. 97-101, vol. 1. |
Erol, B. et al., "Prescient Paper: Multimedia Document Creation with Document Image Matching," 17th International Conference on Pattern Recognition, Aug. 23-26, 2004, Cambridge, UK. |
Erol, B. et al., "Retrieval of Presentation Recordings with Digital Camera Images," IEEE Conference on Computer Vision and Pattern Recognition, Jun. 27-Jul. 2, 2004. |
Esposito, F. et al., "Machine Learning Methods for Automatically Processing Historical Documents: from Paper Acquiestion to XML Transformation," Proceedings of the First International Workshop on Document Image Analysis for Libraries (DIAL '04), IEEE, 2004, pp. 1-8. |
European Partial Search Report, European Application No. EP07015093.3, Dec. 17, 2007, 7 pages. |
European Search Report, European Application No. 06796844.6, Oct. 30, 2008, 12 pages. |
European Search Report, European Application No. 06796845.3, Oct. 30, 2008, 12 pages. |
European Search Report, European Application No. 06796846.1, Nov. 5, 2008, 11 pages. |
European Search Report, European Application No. 06796848.7, Oct. 31, 2008, 12 pages. |
European Search Report, European Application No. 07252397, Oct. 15, 2007, 7 pages. |
European Search Report, European Application No. 08159971.4, Nov. 14, 2008, 6 pages. |
European Search Report, European Application No. 08160112.2, Nov. 10, 2008, 7 pages. |
European Search Report, European Application No. 08160115.5, Nov. 12, 2008, 6 pages. |
European Search Report, European Application No. 08160125.4, Oct. 13, 2008, 5 pages. |
European Search Report, European Application No. 08160130.4, Nov. 12, 2008, 7 pages. |
European Search Report, European Application No. 09156089.6, Jun. 19, 2009, 8 pages. |
European Search Report, European Application No. 09170045.0, Nov. 24, 2009, 4 pages. |
Extended European Search Report, Application No. 09178280.5-2201/2202646, Aug. 31, 2010, 6 pages. |
Extended European Search Report, European Patent Application No. 082523770, May 2, 2011, 6 pages. |
Ezaki, N. et al., "Text Detection from Natural Scene Images: Towards a System for Visually Impaired Persons," Proc. of 17th Int. Conf. on Pattern Recognition (ICPR 2004), IEEE Computer Society, Aug. 23-26, 2004, Cambridge, UK, pp. 683-686, vol. II. |
Fadoua, D. et al., "Restoring Ink Bleed-Through Degraded Document Images Using a Recursive Unsupervised Classification Technique," Lecture Notes in Computer Science 3872, Document Analysis Systems VII, 7th International Workshop, DAS 2006, Feb. 13-15, 2006, Nelson, New Zealand, Bunke, H. et al. (eds.), pp. 38-49. |
Freund, Y. et al., "A Short Introduction to Boosting," Journal of Japanese Society for Artificial Intelligence, Sep. 1999, pp. 771-780, vol. 14, No. 5. |
Hjelmas, E. et al., "Face Detection: A Survey," Computer Vision and Image Understanding, 2001, pp. 236-274, vol. 83. |
Ho, T.K. et al., "Decision Combination in Multiple Classifier Systems," IEEE Transactions on Pattern Analysis and Machine Intelligence, Jan. 1994, pp. 66-75, vol. 16, No. 1. |
Ho, T.K. et al., "Evaluation of OCT Accuracy Using Synthetic Data," Proceedings of the 4th Annual Symposium on Document Analysis and Information Retrieval, Apr. 24-26, 1995, pp. 413-422. [online] Retrieved from the Internet<URL:http://citeseer.ist.psu.edu/cache/papers/cs/2303/http:zSzzSzcm.bell-labs.comzSzcmzSzcszSzwhozSzhsbzSzeoasd.pdf/ho95evaluation.pdf>. |
Hull, J. J., "Document Image Similarity and Equivalence Detection," International Journal on Document Analysis and Recognition, 1998, pp. 37-42, Springer-Verlag. |
Hull, J., "Document Image Skew Detection: Survey and Annotated Bibliography," Document Analysis Systems II, World Scientific, 1998, pp. 40-64. |
Hull, J.J. et al., "Document Image Matching Techniques," Apr. 30, 1997, pp. 31-35, [Online] [Retrieved on May 2, 1997] Retrieved from the Internet. |
Hull, J.J. et al., "Document Image Matching Techniques," Apr. 30, 1997, pp. 31-35, [Online] [Retrieved on May 2, 1997] Retrieved from the Internet<URL:http://rii.ricoch.com/hull/pubs/hull—sdiut97.pdf>. |
Hull, J.J. et al., "Paper-Based Augmented Reality," 17th International Conference on Artificial Reality and Telexistence, Nov. 1, 2007, pp. 205-209. |
Hull, J.J. et al., "Visualizing Multimedia Content on Paper Documents: Components of Key Frame Selection for Video Paper," Proceedings of the Seventh International Conference on Document Analysis and Recognition (ICDAR'03), IEEE, 2003, 4 pages. |
Hull, J.J., "Document Image Matching and Retrieval with Multiple Distortion-Invariant Descriptors," International Association for Pattern Recognition Workshop on Document Analysis Systems, Jan. 1, 1995, pp. 375-396. |
Hull, J.J., "Document Image Matching on CCITT Group 4 Compressed Images," SPIE Conference on Document Recognition IV, Feb. 8, 1997, pp. 82-87. |
Jagannathan, L. et al., Perspective Correction Methods for Camera Based Document Analysis, Proc. First Int. Workshop on Camera-based Document Analysis and Recognition, 2005, pp. 148-154. |
Jain, A.K. et al., "An Introduction to Biometric Recognition," IEEE Transactions on Circuits and Systems for Video Technology, Jan. 2004, pp. 4-20, vol. 14, No. 1. |
Japanese Office Action, Japanese Application No. 2004-293962, Aug. 24, 2010, 3 pages. |
Japanese Office Action, Japanese Patent Application No. 2008-008112, Oct. 25, 2011, 3 pages. |
Japanese Office Action, Japanese Patent Application No. 200910138044.X, Jan. 26, 2011, 6 pages. |
Kanungo, T. et al., "A Downhill Simplex Algorithm for Estimating Morphological Degradation Model Parameters," University of Maryland Technical Report, LAMP-RT-066, Feb. 2001, 15 pages, [online] Retrieved from the Internet. |
Kanungo, T. et al., "A Downhill Simplex Algorithm for Estimating Morphological Degradation Model Parameters," University of Maryland Technical Report, LAMP-RT-066, Feb. 2001, 15 pages, [online] Retrieved from the Internet<URL:http://lampsrv01. umiacs.umd.edu/pubs/TechReports/LAMP—066/LAMP—066.pdf>. |
Kanungo, T. et al., "Global and Local Document Degradation Models," Document Analysis and Recognition, 1993, Proceedings of the Second International Conference on Volume, Oct. 20-22, 1993, pp. 730-734. |
Khoubyari, S. et al., "Font and Funct on Word Ident ficat on n Document Recogn t on," Computer Vision and Image Understanding, Jan. 1996, pp. 66-74, vol. 63, No. 1. |
Khoubyari, S. et al., "Keyword Location and Noisy Document Images," Second Annual Symposium on Document Analysis and Information Retrieval, Las Vegas, NV, Apr. 26-28, 1993, pp. 217-231. |
Kopec, G.E. et al., "Document Image Decoding Using Markov Source Models," IEEE Transactions on Pattern Analysis and Machine Intelligence, Jun. 1994, pp. 602-617, vol. 16, No. 6. |
Li, Y. et al., "Validation of Image Defect Models for Optical Character Recognition," IEEE Trans. Pattern Anal. Mach. Intell. 18, Feb. 2, 1996, pp. 99-108, [online] Retrieved from the Internet. |
Li, Y. et al., "Validation of Image Defect Models for Optical Character Recognition," IEEE Trans. Pattern Anal. Mach. Intell. 18, Feb. 2, 1996, pp. 99-108, [online] Retrieved from the Internet<URL:http://www.cs.cmu.edu/afs/cs/usr/andrewt/papers/Validate/journal.ps.gz>. |
Liang, J. et al., "Flattening Curved Documents in Images," In Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2005, 8 pages, [online] Retrieved from the Internet<URL: http://www.cfar.umd.edu/~daniel/daniel-papersfordownload/liang-j-cpvr2005.pdf>. |
Liang, J. et al., "Flattening Curved Documents in Images," In Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2005, 8 pages, [online] Retrieved from the Internet<URL: http://www.cfar.umd.edu/˜daniel/daniel—papersfordownload/liang-j—cpvr2005.pdf>. |
Liu, T. et al., "A Fast Image Segmentation Algorithm for Interactive Video Hotspot Retrieval," IEEE, 2001, pp. 3-8. |
Liu, Y. et al., "Automatic Texture Segmentation for Texture-Based Image Retrieval," IEEE, Jan. 5-7, 2004, pp. 285-288. |
Lu, Y. et al., "Document Retrieval from Compressed Images," Pattern Recognition, 2003, pp. 987-996, vol. 36. |
Mae et al., "Object Recognition Using Appearance Models Accumulated into Environment", Proc. 15-th Intl. Conf. on Pattern Recognition, 2000, vol. 4, pp. 845-848. |
Marques, O. et al., "Content-Based Image and Video Retrieval, Video Content Representation, Indexing, and Retrieval, a Survey of Content-Based Image Retrieval Systems, CBVQ (Content-Based Visual Query)," Content-Based Image and Video Retrieval [Multimedia Systems and Applications Series], Apr. 1, 2002, pp. 15-117, vol. 21, Kluwer Academic Publishers Group, Boston, USA. |
McDonald, G., "Third Voice: Invisible Web Graffiti," PC World, May 18, 1999, [online] [Retrieved on Nov. 14, 2006] Retrieved from the Internet. |
McDonald, G., "Third Voice: Invisible Web Graffiti," PC World, May 18, 1999, [online] [Retrieved on Nov. 14, 2006] Retrieved from the Internet<URL: http://www.pcworld.com/news/article/0,aid,11016,00.asp>. |
Microsoft Computer Dictionary (5th ed.), 2002, "Hyperlink" Definition, pp. 260-261. |
Mukherjea, S. et al., "AMORE: A World Wide Web Image Retrieval Engine," C&C Research Laboratories, NEC USA Inc., Baltzer Science Publishers BV, World Wide Web 2, 1999, pp. 115-132. |
Pavlidis, T., "Effects of Distortions on the Recognition Rate of a Structural OCR System," In Pro. Conf. on Comp. Vision and Pattern Recog., IEEE, Washington, DC, 1983, pp. 303-309. |
PCT International Search Report and Written Opinion, PCT/JP2006/316810, Oct. 10, 2006, 9 pages. |
PCT International Search Report and Written Opinion, PCT/JP2006/316811, Oct. 10, 2006, 9 pages. |
PCT International Search Report and Written Opinion, PCT/JP2006/316812, Oct. 10, 2006, 9 pages. |
PCT International Search Report and Written Opinion, PCT/JP2006/316814, Oct. 10, 2006, 11 pages. |
Po, L-M. et al., "A Novel Four-Step Search Algorithm for Fast Block Motion Estimation," IEEE Transactions on Circuits and Systems for Video Technology, Jun. 1996, pp. 313-317, vol. 6, Issue 3. |
Rademacher, "View-Dependent Geometry", Computer Graphics Proceedings, Annual Conference Series, SIGGRAPH 99, Los Angeles, California Aug. 8-13, 1999, pp. 439-446, XP001024743. |
Rangarajan, K. et al. "Optimal Corner Detector," 1988, IEEE, pp. 90-94. |
Reniers et al., "Skeleton-based Hierarchical Shape Segmentation", IEEE International Conference on Shape Modeling and Applications. SMI'07, Jun. 1, 2007, Computer Society, pp. 179-188, XP031116745, ISBN: 978-0-7695-2815-1. |
Rosin, P.L. et al., "Image Difference Threshold Strategies and Shadow Detection," Proceedings of the 6th British Machine Vision Conference, 1995,10 pages. |
Roth, M.T. et al., "The Garlic Project," Proc. of the 1996 ACM SIGMOD International Conference on Management of Data, Montreal, Quebec, Canada, Jun. 4, 1996, pp. 557. |
Sato, T. et al., "High Resolution Video Mosaicing for Documents and Photos by Estimating Camera Motion," Proceedings of the SPIE 5299, 246, 2004, 8 pages, [online] Retrieved from the Internet. |
Sato, T. et al., "High Resolution Video Mosaicing for Documents and Photos by Estimating Camera Motion," Proceedings of the SPIE 5299, 246, 2004, 8 pages, [online] Retrieved from the Internet<URL: http://yokoya.naist.jp/paper/datas/711/spie2004.pdf>. |
Schalkoff, R.J., "Syntactic Pattern Recognition (SYNTPR) Overview," Pattern Recognition: Statistical, Structural and Neural Approaces, Jan. 1, 1992, pp. 127-150, vol. 3, Wiley. |
Sezgin, M. et al., "Survey Over Image Thresholding Techniques and Quantitative Performance Evaluation," Journal of Electronic Imaging, Jan. 2004, pp. 146-165, vol. 13, No. 1. |
Sivic, J. et al., "Video Google: A Text Retrieval Approach to Object Matching in Videos," Proceedings of the Ninth IEEE International Conference on ComputerVision (ICCV 2003), 2-Volume Set, 2003, IEEE, pp. 1-8. |
Stoyanov, D., "Camera Calibration Tools," Online Source, Updated Aug. 24, 2006, Accessed Aug. 31, 2006, 12 pages, [online] Retrieved from the Internet. |
Stoyanov, D., "Camera Calibration Tools," Online Source, Updated Aug. 24, 2006, Accessed Aug. 31, 2006, 12 pages, [online] Retrieved from the Internet<URL:http://ubimon.doc.ic.ac.uk/dvs/index.php?m=581>. |
Triantafyllidis, G.A. et al., "Detection of Blocking Artifacts of Compressed Still Images," Proceedings of the 11th International Conference on Image Analysis and Processing (ICIAP '01), IEEE, 2001, pp. 1-5. |
U.S. Appl. No. 10/696,735, filed Oct. 28, 2003, Erol, B. et al., "Techniques for Using a Captured Image for the Retrieval of Recorded Information," 58 pages. |
U.S. Appl. No. 10/813,901, filed Mar. 30, 2004, Erol et al. |
U.S. Notice of Allowance, U.S. Appl. No. 11/461,024, Mar. 16, 2011, 12 pages. |
U.S. Notice of Allowance, U.S. Appl. No. 11/461,143, Dec. 5, 2011, 15 pages. |
U.S. Notice of Allowance, U.S. Appl. No. 11/461,272, Feb. 23, 2011, 28 pages. |
U.S. Notice of Allowance, U.S. Appl. No. 11/624,466, Feb. 22, 2011, 12 pages. |
U.S. Notice of Allowance, U.S. Appl. No. 12/060,194, Nov. 28, 2011, 23 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,037, Mar. 30, 2011, 29 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,037, Nov. 23, 2011, 33 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,126, Apr. 12, 2011, 27 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,126, Dec. 1, 2011, 53 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,126, Dec. 23, 2010, 30 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,143, Feb. 4, 2011, 16 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,279, Jan. 7, 2011, 44 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,286, Jan. 21, 2011, 34 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,294, Apr. 12, 2011, 23 pages. |
U.S. Office Action, U.S. Appl. No. 11/461,300, Mar. 18, 2011, 25 pages. |
U.S. Office Action, U.S. Appl. No. 11/776,510, Dec. 10, 2010, 39 pages. |
U.S. Office Action, U.S. Appl. No. 11/776,520, Dec. 7, 2010, 43 pages. |
U.S. Office Action, U.S. Appl. No. 11/776,520, Oct. 17, 2011, 19 pages. |
U.S. Office Action, U.S. Appl. No. 11/777,142, Mar. 18, 2011, 21 pages. |
U.S. Office Action, U.S. Appl. No. 12/060,194, Feb. 25, 2011, 18 pages. |
U.S. Office Action, U.S. Appl. No. 12/121,275, Oct. 19, 2011, 24 pages. |
U.S. Office Action, U.S. Appl. No. 12/210,511, Apr. 4, 2011, 49 pages. |
U.S. Office Action, U.S. Appl. No. 12/210,519, Mar. 14, 2011, 38 pages. |
U.S. Office Action, U.S. Appl. No. 12/210,532, Oct. 31, 2011, 61 pages. |
U.S. Office Action, U.S. Appl. No. 12/210,540, Apr. 15, 2011, 45 pages. |
U.S. Office Action, U.S. Appl. No. 12/210,540, Oct. 14, 2011, 22 pages. |
U.S. Office Action, U.S. Appl. No. 12/240,596, Jan. 21, 2011, 21 pages. |
U.S. Office Action, U.S. Appl. No. 12/247,202, Apr. 6, 2011, 37 pages. |
U.S. Office Action, U.S. Appl. No. 12/265,502, Oct. 14, 2011, 61 pages. |
U.S. Office Action, U.S. Appl. No. 12/340,124, Apr. 15, 2011, 48 pages. |
U.S. Office Action, U.S. Appl. No. 12/340,124, Oct. 24, 2011, 31 pages. |
U.S. Office Action, U.S. Appl. No. 12/719,437, Dec. 9, 2010, 38 pages. |
U.S. Office Action, U.S. Appl. No. 12/879,933, Mar. 2, 2011, 7 pages. |
U.S. Office Action, U.S. Appl. No. 12/879,933, Oct. 28, 2011, 36 pages. |
U.S. Patent Office Action, U.S. Appl. No. 11/776,510, May 12, 2011, 20 pages. |
U.S. Patent Office Action, U.S. Appl. No. 11/776,520, Apr. 28, 2011, 10 pages. |
U.S. Patent Office Action, U.S. Appl. No. 12/060,194, Jun. 27, 2011, 18 pages. |
U.S. Patent Office Action, U.S. Appl. No. 12/121,275, Apr. 20, 2011, 44 pages. |
U.S. Patent Office Action, U.S. Appl. No. 12/210,519, Jun. 16, 2011, 22 pages. |
United States Final Office Action, U.S. Appl. No. 11/461,286, Apr. 16, 2012, 47 pages. |
United States Final Office Action, U.S. Appl. No. 11/461,294, Apr. 13, 2012, 23 pages. |
United States Final Office Action, U.S. Appl. No. 12/060,198, Apr. 12, 2012, 74 pages. |
United States Final Office Action, U.S. Appl. No. 12/060,200, Apr. 12, 2012, 65 pages. |
United States Final Office Action, U.S. Appl. No. 12/265,502, Apr. 5, 2012, 49 pages. |
United States Final Office Action, U.S. Appl. No. 12/719,437, Mar. 1, 2012, 21 pages. |
United States Non-Final Office Action, U.S. Appl. No. 11/461,279, Apr. 19, 2012, 59 pages. |
United States Non-Final Office Action, U.S. Appl. No. 12/121,275, May 18, 2012, 41 pages. |
United States Non-Final Office Action, U.S. Appl. No. 12/240,590, Apr. 4, 2012, 73 pages. |
United States Notice of Allowability, U.S. Appl. No. 11/461,091, Nov. 17, 2010, 22 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,024, Nov. 15, 2010, 10 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,049, Nov. 16, 2010, 10 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,091, Oct. 18, 2010, 31 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,095, Sep. 27, 2010, 29 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,126, Mar. 5, 2012, 19 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/461,143, Mar. 8, 2012, 9 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/776,510, Sep. 22, 2011, 11 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/776,520, Jan. 5, 2012, 28 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/776,530, Dec. 21, 2011, 17 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/776,530, Mar. 26, 2012, 5 pages. |
United States Notice of Allowance, U.S. Appl. No. 11/777,142, Jan. 20, 2012, 29 pages. |
United States Notice of Allowance, U.S. Appl. No. 12/247,202, Jul. 28, 2011, 16 pages. |
United States Notice of Allowance, U.S. Appl. No. 12/253,815, Jan. 26, 2012, 62 pages. |
United States Notice of Allowance, U.S. Appl. No. 12/342,330, Jan. 23, 2012, 21 pages. |
United States Notice of Allowance, U.S. Appl. No. 13/168,638, Apr. 4, 2012, 30 pages. |
United States Office Action, U.S. Appl. No. 11/461,024, Jul. 14, 2010, 29 pages. |
United States Office Action, U.S. Appl. No. 11/461,049, Jul. 28, 2010, 27 pages. |
United States Office Action, U.S. Appl. No. 11/461,126, Jul. 22, 2011, 35 pages. |
United States Office Action, U.S. Appl. No. 11/461,126, Sep. 3, 2010, 28 pages. |
United States Office Action, U.S. Appl. No. 11/461,143, Aug. 11, 2011, 20 pages. |
United States Office Action, U.S. Appl. No. 11/461,143, Aug. 18, 2010, 9 pages. |
United States Office Action, U.S. Appl. No. 11/461,272, Aug. 23, 2010, 31 pages. |
United States Office Action, U.S. Appl. No. 11/461,279, Aug. 5, 2010, 37 pages. |
United States Office Action, U.S. Appl. No. 11/461,279, Jul. 8, 2011, 46 pages. |
United States Office Action, U.S. Appl. No. 11/461,286, Aug. 5, 2010, 28 pages. |
United States Office Action, U.S. Appl. No. 11/461,286, Jan. 20, 2012, 27 pages. |
United States Office Action, U.S. Appl. No. 11/461,286, Jul. 15, 2011, 37 pages. |
United States Office Action, U.S. Appl. No. 11/461,294, Jan. 18, 2012, 38 pages. |
United States Office Action, U.S. Appl. No. 11/461,294, Jun. 11, 2010, 19 pages. |
United States Office Action, U.S. Appl. No. 11/461,294, Oct. 7, 2010, 17 pages. |
United States Office Action, U.S. Appl. No. 11/461,300, Feb. 23, 2012, 38 pages. |
United States Office Action, U.S. Appl. No. 11/461,300, Jun. 11, 2010, 20 pages. |
United States Office Action, U.S. Appl. No. 11/461,300, Oct. 6, 2010, 20 pages. |
United States Office Action, U.S. Appl. No. 11/624,466, Jun. 8, 2010, 29 pages. |
United States Office Action, U.S. Appl. No. 11/624,466, Oct. 14, 2010, 11 pages. |
United States Office Action, U.S. Appl. No. 11/776,530, Aug. 19, 2011, 54 pages. |
United States Office Action, U.S. Appl. No. 11/777,142, Nov. 10, 2010, 17 pages. |
United States Office Action, U.S. Appl. No. 11/777,142, Sep. 23, 2011, 22 pages. |
United States Office Action, U.S. Appl. No. 12/059,583, Jan. 26, 2012, 78 pages. |
United States Office Action, U.S. Appl. No. 12/060,194, Oct. 1, 2010, 29 pages. |
United States Office Action, U.S. Appl. No. 12/060,198, Sep. 1, 2011, 87 pages. |
United States Office Action, U.S. Appl. No. 12/060,200, Sep. 2, 2011, 65 pages. |
United States Office Action, U.S. Appl. No. 12/060,206, Dec. 15, 2011, 55 pages. |
United States Office Action, U.S. Appl. No. 12/210,511, Sep. 28, 2011, 25 pages. |
United States Office Action, U.S. Appl. No. 12/210,519, Jan. 5, 2012, 29 pages. |
United States Office Action, U.S. Appl. No. 12/240,596, Aug. 6, 2010, 32 pages. |
United States Office Action, U.S. Appl. No. 12/240,596, Feb. 2, 2012, 44 pages. |
United States Office Action, U.S. Appl. No. 12/247,205, Oct. 6, 2011, 56 pages. |
United States Office Action, U.S. Appl. No. 12/253,715, Aug. 31, 2011, 58 pages. |
United States Office Action, U.S. Appl. No. 12/342,330, Oct. 7, 2011, 47 pages. |
Veltkamp, R. et al., "Content-Based Image Retrieval Systems: A Survey," Department of Computing Science, Utrecht University, Oct. 28, 2002, pp. 1-62. |
Wikipedia Online Definition, "Optical Character Recognition," Sep. 14, 2008, pp. 1-7, [online] [Retrieved on Sep. 14, 2008] Retrieved from the Internet. |
Wikipedia Online Definition, "Optical Character Recognition," Sep. 14, 2008, pp. 1-7, [online] [Retrieved on Sep. 14, 2008] Retrieved from the Internet<URL:http://en.wikipedia.org/wiki/Optical—character—recognition>. |
Wikipedia Online Encyclopedia, "Image Scanner," Last Modified Feb. 9, 2010, pp. 1-9, [Online] [Retrieved on Feb. 13, 2010] Retrieved from the Internet. |
Wikipedia Online Encyclopedia, "Image Scanner," Last Modified Feb. 9, 2010, pp. 1-9, [Online] [Retrieved on Feb. 13, 2010] Retrieved from the Internet<URL:http://en.wikipedia.org/wiki/Image—scanner>. |
Wikipedia Online Encyclopedia, "Waypoint," Last Modified Feb. 13, 2010, pp. 1-4, [Online] Retrieved on Feb. 13, 2010] Retrieved from the Internet. |
Wikipedia Online Encyclopedia, "Waypoint," Last Modified Feb. 13, 2010, pp. 1-4, [Online] Retrieved on Feb. 13, 2010] Retrieved from the Internet<URL:http://en.wikipedia.org/wiki/Waypoint>. |
Wikipedia Online Encyclopedia,"Automatic Identification and Data Capture," Jul. 21, 2008, pp. 1-2, [Online] [Retrieved on Sep. 27, 2008] Retrieved from the Internet. |
Wikipedia Online Encyclopedia,"Automatic Identification and Data Capture," Jul. 21, 2008, pp. 1-2, [Online] [Retrieved on Sep. 27, 2008] Retrieved from the Internet<URL:http://en.wikipedia.org/wiki/Automatic—identification—and—data—capture>. |
Zanibbi, R. et al. "A Survey of Table Recognition," International Journal on Document Analysis and Recognition, 2004, pp. 1-33. |
Zhang, Z., "A Flexible New Technique for Camera Calibration," IEEE Transactions on Pattern Analysis and Machine Intelligence, Nov. 2000, pp. 1330-1334, vol. 22, No. 11. |
Zhao, W. et al., Face Recognition: A Literature Survey, ACM Computing Surveys (CSUR), 2003, pp. 399-458, vol. 35, No. 4. |
Zheng, Q.-F. et al., "Effective and Efficient Object-Based Image Retrieval Using Visual Phases," Proceedings of the 14th Annual ACM International Conference on Multimedia, MM'06, Oct. 23-27, 2006, Santa Barbara, CA, pp. 77-80. |
Zi, G., "Groundtruth Generation and Document Image Degradation," University of Maryland Language and Media Processing Laboratory Technical report (LAMP-TR-121), May 2005, 72 pages, [online] Retrieved from the Internet. |
Zi, G., "Groundtruth Generation and Document Image Degradation," University of Maryland Language and Media Processing Laboratory Technical report (LAMP-TR-121), May 2005, 72 pages, [online] Retrieved from the Internet<URL:http://lampsrv01.umiacs.umd.edu/pubs/TechReports/LAMP—121/LAMP—121.pdf>. |
Cited By (56)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100306681A1 (en) * | 2003-09-30 | 2010-12-02 | Dirtt Environmental Solutions Ltd. | Creation and modification of valid functional design layouts |
US8762877B2 (en) * | 2003-09-30 | 2014-06-24 | Ice Edge Business Solutions Ltd. | Creation and modification of valid functional design layouts |
US9536340B2 (en) | 2004-08-17 | 2017-01-03 | Dirtt Environmental Solutions, Ltd. | Software incorporating efficient 3-D rendering |
US8600989B2 (en) | 2004-10-01 | 2013-12-03 | Ricoh Co., Ltd. | Method and system for image matching in a mixed media environment |
US9063953B2 (en) | 2004-10-01 | 2015-06-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment |
US8521737B2 (en) | 2004-10-01 | 2013-08-27 | Ricoh Co., Ltd. | Method and system for multi-tier image matching in a mixed media environment |
US9405751B2 (en) | 2005-08-23 | 2016-08-02 | Ricoh Co., Ltd. | Database for mixed media document system |
US9171202B2 (en) | 2005-08-23 | 2015-10-27 | Ricoh Co., Ltd. | Data organization and access for mixed media document system |
US8949287B2 (en) | 2005-08-23 | 2015-02-03 | Ricoh Co., Ltd. | Embedding hot spots in imaged documents |
US8838591B2 (en) | 2005-08-23 | 2014-09-16 | Ricoh Co., Ltd. | Embedding hot spots in electronic documents |
US9357098B2 (en) | 2005-08-23 | 2016-05-31 | Ricoh Co., Ltd. | System and methods for use of voice mail and email in a mixed media environment |
US8452780B2 (en) | 2006-01-06 | 2013-05-28 | Ricoh Co., Ltd. | Dynamic presentation of targeted information in a mixed media reality recognition system |
US8510283B2 (en) | 2006-07-31 | 2013-08-13 | Ricoh Co., Ltd. | Automatic adaption of an image recognition system to image capture devices |
US8825682B2 (en) | 2006-07-31 | 2014-09-02 | Ricoh Co., Ltd. | Architecture for mixed media reality retrieval of locations and registration of images |
US9020966B2 (en) | 2006-07-31 | 2015-04-28 | Ricoh Co., Ltd. | Client device for interacting with a mixed media reality recognition system |
US9311336B2 (en) | 2006-07-31 | 2016-04-12 | Ricoh Co., Ltd. | Generating and storing a printed representation of a document on a local computer upon printing |
US8676810B2 (en) | 2006-07-31 | 2014-03-18 | Ricoh Co., Ltd. | Multiple index mixed media reality recognition using unequal priority indexes |
US8369655B2 (en) | 2006-07-31 | 2013-02-05 | Ricoh Co., Ltd. | Mixed media reality recognition using multiple specialized indexes |
US8489987B2 (en) | 2006-07-31 | 2013-07-16 | Ricoh Co., Ltd. | Monitoring and analyzing creation and usage of visual content using image and hotspot interaction |
US9063952B2 (en) | 2006-07-31 | 2015-06-23 | Ricoh Co., Ltd. | Mixed media reality recognition with image tracking |
US9384619B2 (en) | 2006-07-31 | 2016-07-05 | Ricoh Co., Ltd. | Searching media content for objects specified using identifiers |
US8856108B2 (en) | 2006-07-31 | 2014-10-07 | Ricoh Co., Ltd. | Combining results of image retrieval processes |
US8868555B2 (en) | 2006-07-31 | 2014-10-21 | Ricoh Co., Ltd. | Computation of a recongnizability score (quality predictor) for image retrieval |
US9176984B2 (en) | 2006-07-31 | 2015-11-03 | Ricoh Co., Ltd | Mixed media reality retrieval of differentially-weighted links |
US9530050B1 (en) | 2007-07-11 | 2016-12-27 | Ricoh Co., Ltd. | Document annotation sharing |
US10192279B1 (en) | 2007-07-11 | 2019-01-29 | Ricoh Co., Ltd. | Indexed document modification sharing with mixed media reality |
US8989431B1 (en) | 2007-07-11 | 2015-03-24 | Ricoh Co., Ltd. | Ad hoc paper-based networking with mixed media reality |
US9373029B2 (en) | 2007-07-11 | 2016-06-21 | Ricoh Co., Ltd. | Invisible junction feature recognition for document security or annotation |
US8478761B2 (en) | 2007-07-12 | 2013-07-02 | Ricoh Co., Ltd. | Retrieving electronic documents by converting them to synthetic text |
US9519407B2 (en) | 2008-03-11 | 2016-12-13 | Ice Edge Business Solutions, Ltd. | Automatically creating and modifying furniture layouts in design software |
US8385589B2 (en) | 2008-05-15 | 2013-02-26 | Berna Erol | Web-based content detection in images, extraction and recognition |
US9191570B2 (en) * | 2009-05-01 | 2015-11-17 | Microsoft Technology Licensing, Llc | Systems and methods for detecting a tilt angle from a depth image |
US9519970B2 (en) | 2009-05-01 | 2016-12-13 | Microsoft Technology Licensing, Llc | Systems and methods for detecting a tilt angle from a depth image |
US20140043438A1 (en) * | 2009-05-01 | 2014-02-13 | Microsoft Corporation | Systems and Methods for Detecting a Tilt Angle from a Depth Image |
US8385660B2 (en) | 2009-06-24 | 2013-02-26 | Ricoh Co., Ltd. | Mixed media reality indexing and retrieval for repeated content |
US9245064B2 (en) | 2009-11-24 | 2016-01-26 | Ice Edge Business Solutions | Securely sharing design renderings over a network |
US8897563B1 (en) * | 2010-01-15 | 2014-11-25 | Gruntworx, Llc | Systems and methods for automatically processing electronic documents |
US20110255782A1 (en) * | 2010-01-15 | 2011-10-20 | Copanion, Inc. | Systems and methods for automatically processing electronic documents using multiple image transformation algorithms |
US8571317B2 (en) * | 2010-01-15 | 2013-10-29 | Gruntworx, Llc | Systems and methods for automatically processing electronic documents using multiple image transformation algorithms |
US11103787B1 (en) | 2010-06-24 | 2021-08-31 | Gregory S. Rabin | System and method for generating a synthetic video stream |
US8452089B2 (en) * | 2011-03-31 | 2013-05-28 | Sony Corporation | System and method for effectively performing an image identification procedure |
US20120250987A1 (en) * | 2011-03-31 | 2012-10-04 | Sony Corporation | System and method for effectively performing an image identification procedure |
US9036925B2 (en) | 2011-04-14 | 2015-05-19 | Qualcomm Incorporated | Robust feature matching for visual search |
US9189571B2 (en) | 2011-06-11 | 2015-11-17 | Ice Edge Business Solutions, Ltd. | Automated re-use of structural components |
US8706711B2 (en) | 2011-06-22 | 2014-04-22 | Qualcomm Incorporated | Descriptor storage and searches of k-dimensional trees |
US9058331B2 (en) | 2011-07-27 | 2015-06-16 | Ricoh Co., Ltd. | Generating a conversation in a social network based on visual search results |
US8612475B2 (en) | 2011-07-27 | 2013-12-17 | Ricoh Co., Ltd. | Generating a discussion group in a social network based on metadata |
US8892595B2 (en) | 2011-07-27 | 2014-11-18 | Ricoh Co., Ltd. | Generating a discussion group in a social network based on similar source materials |
US9019568B2 (en) | 2012-03-22 | 2015-04-28 | Intuit Inc. | Accurately extracting information based on orthogonal images |
US9355197B2 (en) | 2013-01-25 | 2016-05-31 | Dirtt Environmental Solutions, Ltd | Real-time depth of field effects within design software |
US9619920B2 (en) | 2013-01-31 | 2017-04-11 | Ice Edge Business Solutions, Ltd. | Method and system for efficient modeling of specular reflection |
US9164653B2 (en) | 2013-03-15 | 2015-10-20 | Inspace Technologies Limited | Three-dimensional space for navigating objects connected in hierarchy |
US9747680B2 (en) | 2013-11-27 | 2017-08-29 | Industrial Technology Research Institute | Inspection apparatus, method, and computer program product for machine vision inspection |
US10922450B2 (en) | 2014-06-09 | 2021-02-16 | Dirtt Environmental Solutions, Ltd. | Associating computer-executable objects with timber frames within an architectural design environment |
US20160026628A1 (en) * | 2014-07-22 | 2016-01-28 | Verizon Patent And Licensing Inc. | Providing content based on image item |
US9330113B2 (en) * | 2014-07-22 | 2016-05-03 | Verizon Patent And Licensing Inc. | Providing content based on image item |
Also Published As
Publication number | Publication date |
---|---|
US20090019402A1 (en) | 2009-01-15 |
JP2009043242A (en) | 2009-02-26 |
JP5181888B2 (en) | 2013-04-10 |
EP2015227B8 (en) | 2018-06-13 |
EP2015227B1 (en) | 2018-05-02 |
EP2015227A2 (en) | 2009-01-14 |
EP2015227A3 (en) | 2011-06-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8276088B2 (en) | User interface for three-dimensional navigation | |
US8184155B2 (en) | Recognition and tracking using invisible junctions | |
US9373029B2 (en) | Invisible junction feature recognition for document security or annotation | |
US8144921B2 (en) | Information retrieval using invisible junctions and geometric constraints | |
US8086038B2 (en) | Invisible junction features for patch recognition | |
CN101297318B (en) | Data organization and access for mixed media document system | |
US7669148B2 (en) | System and methods for portable device for mixed media system | |
US7639387B2 (en) | Authoring tools using a mixed media environment | |
US7702673B2 (en) | System and methods for creation and use of a mixed media environment | |
US7551780B2 (en) | System and method for using individualized mixed document | |
US7917554B2 (en) | Visibly-perceptible hot spots in documents | |
US8195659B2 (en) | Integration and use of mixed media documents | |
US20110081892A1 (en) | System and methods for use of voice mail and email in a mixed media environment | |
EP1917636A1 (en) | Method and system for image matching in a mixed media environment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: RICOH CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KE, QIFA;HART, PETER H.;HULL, JONATHAN J.;AND OTHERS;REEL/FRAME:019887/0290;SIGNING DATES FROM 20070815 TO 20070816 Owner name: RICOH CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KE, QIFA;HART, PETER H.;HULL, JONATHAN J.;AND OTHERS;SIGNING DATES FROM 20070815 TO 20070816;REEL/FRAME:019887/0290 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |