WO1999012107A1 - Method and device for parsing natural language sentences and other sequential symbolic expressions - Google Patents

Method and device for parsing natural language sentences and other sequential symbolic expressions Download PDF

Info

Publication number
WO1999012107A1
WO1999012107A1 PCT/US1998/017865 US9817865W WO9912107A1 WO 1999012107 A1 WO1999012107 A1 WO 1999012107A1 US 9817865 W US9817865 W US 9817865W WO 9912107 A1 WO9912107 A1 WO 9912107A1
Authority
WO
WIPO (PCT)
Prior art keywords
symbol
picture
relation
grouping
cognitive
Prior art date
Application number
PCT/US1998/017865
Other languages
French (fr)
Inventor
Douglas E. Brash
Original Assignee
Brash Douglas E
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Brash Douglas E filed Critical Brash Douglas E
Priority to AU92095/98A priority Critical patent/AU9209598A/en
Publication of WO1999012107A1 publication Critical patent/WO1999012107A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Definitions

  • the invention relates to a method and device for parsing natural language sentences and other symbolic expressions, and more specifically, the invention relates to a method and device for parsing natural language sentences and other symbolic expressions into cognitive units.
  • a parsing device converts a linear symbol sequence into an organized structure.
  • the diagrammed sentence encountered in grade-school grammar is one kind of parse.
  • This organized structure can then be displayed for instructional purposes, or other devices can use it to initiate commands, store information, or make logical inferences.
  • a parsing device is a prerequisite for using natural language to communicate with computers and, moreover, a simple parsing device is needed if natural language is to be used for controlling household devices such as kitchen appliances, videocassette recorders, lighting, and heating or air conditioning.
  • the central problem in parsing is to achieve both accuracy and speed.
  • synthetic languages such as computer languages, many effective solutions are available. However, the complexity and ambiguity of natural languages have made them resistant to efficient parsing.
  • Automated sentence parsing devices usually general purpose digital computers — have operated by matching parts of the input sentence with a very large number of stored rules. If the first rule does not apply, the next is chosen for examination.
  • Three broad classes of devices have been used; these employ increasingly deep levels of sentence analysis.
  • the most common approach is Syntactic Parsing, in which the rules driving the parsing device describe the organization of words in a particular language. In practice, achieving a parse also requires resolving the possible senses of ambiguous words. This is done by a semantic interpreter, which uses information about word meaning. Semantic Parsing aims at the level of meaning directly, driving the parser with rules relating ordered words to the underlying meaning of the sentence.
  • Conceptual Analysis also aims directly for the level of meaning, but does not attempt to preserve word order.
  • Conceptual analysis parsers identify key words as tokens for the underlying objects or actions, and the organization sought is that of the objects being discussed rather than that of the words used to discuss them.
  • the processor reconstructs this organization by using "encyclopedia information" about the objects, rather than "dictionary information” about the words.
  • the stored rules can be either template-patterns, which must be matched exactly, or general rules, which are applied to parts of the sentence in succession to build a tree-like parsed structure of increasingly finer detail, from a sentence to its phrases to words.
  • the successive-rule strategy termed "generative,” is combinatorial since rules can be used more than once in various combinations. Rules can be chosen so as to divide a sentence into smaller units ("top-down parsing") or, less frequently, to assemble words into higher-level units (“bottom-up” parsing).
  • parsing method and device which would utilize a small number of rules, minimal semantic information, and minimal computing power.
  • Such parsing method and device would require minimal disambiguation of words having multiple meanings and would not rely on specific knowledge bases, word cooccurrence probability data, selection-restrictions, frames, or expectations about sentence content.
  • Such parsing method and device would also process symbol strings in a time proportional to sentence length.
  • Another object of the invention is to provide a system to parse thought expressions, such as natural language, using minimal semantic information.
  • Still another object of the invention is to provide a system to parse thought expressions, such as natural language, using minimal computing power.
  • Yet another object of the invention is to provide a system to parse thought expressions, such as natural language, with minimal disambiguation of words having multiple meanings.
  • Another object of the invention is to provide a system to parse thought expressions, such as natural language, without limiting the vocabulary to a specific knowledge base.
  • Still another object of the invention is to provide a system of the above character to parse even surprising thought expressions by avoiding word co-occurrence probability data, selectional restrictions, frames, and expectations.
  • Yet another object of the invention is to provide a system of the above character to parse while the expression is being entered into a processor.
  • Another object of the invention is to provide a system of the above character to parse in a time period corresponding linearly to sentence length.
  • the invention is a system for parsing symbol sequences representing thoughts, using a small set of rules herein termed "cognitive rules.”
  • the basic rule distinguishes only between symbols for pictures and symbols for relations.
  • the rules use the symbols for relations as signals for particular parsing procedures.
  • the parser i) divides the sentence into words for pictures and words for the relations between pictures and then ii) assembles picture-relation-picture symbol sequences into symbol groups that constitute new picture symbols, using rules for detecting picture borders.
  • the invention is thus a "bottom-up generative cognitive parser.”
  • cognitive parsing which distinguishes primarily between symbols for pictures (such as “squirrel” or “justice”) and symbols for relations (such as “above” or “pushed”).
  • a picture can be a sound, feeling, or past experience as well as a visual image.
  • symbols are words, such pictures are symbolized by nouns, adjectives, and pronouns, including relative pronouns such as "that” and “which.”
  • Relations are typically symbolized by prepositions and verbs.
  • Cognitive parsing operates at a level more basic than syntactic parsing or conceptual analysis. This simplicity allows it to use only three rules, together with about a dozen criteria for deciding when to apply these rules.
  • the cognitive parser's fundamental operations are a) scanning the input symbol sequence for "picture” and “relation” symbols, b) using the relation symbols to segregate the picture symbols, and c) assembling picture and relation symbol sequences into symbol groups that constitute new picture symbols. Since the distinction between "picture” and “relation” is a primitive form of meaning, and symbol order is preserved, a cognitive parser used for words is a type of semantic parser. The primitive nature of these cognitive rules also allows a cognitive parser to apply them to thought- expressions other than words, such as a string of signs in American Sign Language. The terminology used to describe the operations performed by the parser is understood in terms of the desired final parsed structure, termed "cognitive form.”
  • the sentence “Squirrels bury nuts” can be divided into picture-relation-picture form: (Squirrels) bury (nuts).
  • the verbal relation "bury” acts as a parsing element in languages that use this word order.
  • the picture-relation-picture form hereafter abbreviated "P rel P”
  • P rel P can itself be a "picture", (P rel P).
  • Such a picture is referred to herein as a "picture-symbol group”; e.g., ((Squirrels) bury (nuts)).
  • More complex sentences can be shown to have the same (P rel P) form, with various other types of relations such as prepositions acting as parsing elements: ((Squirrels) bury (nuts)) under (trees). [((Squirrels) bury (nuts)) under (trees)] when [(it) is (Fall)].
  • the square parentheses are used for clarity and do not differ in principle from ordinary parentheses.
  • These structures can be written using an immediate-constituent rule: P rel P — > P. This rule differs from those of immediate-constituent grammars in the prior art because i) it always generates a tri-forked tree and ii) each level of the tree uses the same immediate-constituent rule.
  • the three lines of this form are generated on separate tracks in the cognitive parser.
  • the final structure is a (P rel P) form: ⁇ ⁇ associated-with ⁇ ⁇ .
  • the (P rel P) form motivates a simple procedure to parse symbolic expressions. Instead of, for example, dividing a sentence into subjects, predicates, objects, and clauses, it is divided into pictures. This can be done readily, since between words for any two pictures lies a word for a relation.
  • a local rule can be used to implement what will be termed the picture- symbol grouping procedure:
  • the present embodiment of the parser writes outward-facing parentheses on both sides of any word that denotes a relation. For example: "Squirrels) bury (nuts". The result is that each parenthesis indicates the border of a picture-word; a right-facing parenthesis indicates the beginning, and a left-facing parenthesis indicates the end.
  • the choice of parentheses as markers herein is arbitrary; therefore, other markers may be used to denote the same thing.
  • the picture- symbol grouping procedure alone, gives a flat structure, such as "(Squirrels) bury (nuts) before (Winter)."
  • Higher- level pictures are built by a hierarchy-building procedure in which, each time a relation is encountered, a hierarchy- building decision is made for the pictures that preceded it: Each time the parser encounters a relation, it not only adds the outward-facing parentheses but may also build a new picture by bracketing any P rel P preceding the relation.
  • Relation- words can be complete, strong, or weak picture builders depending on whether successive hierarchy-building continues all the way back to the beginning of the sentence or stops at a prior word that acts as a border.
  • complete builders are: the period between sentences; verbs; higher relations (relations that can join statements, such as "because”); path prepositions (such as "through”); and the comma at the end of a parenthetical expression.
  • Strong builders include most prepositions, and are blocked by "the.”
  • Weak builders are blocked by any relation except 1; these prepositions, such as "of,” closely relate the pictures following and preceding. Additional borders arising in other situations can be determined by one skilled in the art, according to the method of the present invention.
  • the parser uses the (P rel P) well-formedness criterion to regenerate missing relations.
  • the (P rel P) criterion tells us to insert a relation.
  • the relation omitted in English seems to be always either 1 or
  • the 2-P procedure If two P's are encountered in succession, the parser inserts either "composed of or "component of between them. The more specific or sensation-related of the two P's is presumed to be a component of the other. If they are equally specific, the relation is 1. If the second P is a determiner, the relation is T. Note that one of skill in the art could derive various scales of abstractness from references such as Roget's Thesaurus or the machine-readable edition of the Longman Dictionary of Contemporary English.
  • the 2-P procedure generates: (red) 1 (windmill); (the) T (windmill); they made ... (him) [ (director); our ... (son) [ (John); (dog) 1 (house).
  • the illustrative embodiment of the parser uses the following empirically-derived hierarchy-building borders: When the 2-P procedure inserts 1, hierarchy-building builds weakly.
  • Introductory phrases are also correctly parsed by these procedures, and circumvent a hierarchy-building conflict.
  • the sentence "He planned the intermission between shows” can have a different meaning from "Between shows, he planned the intermission.”
  • the phrase after the comma is put on a sidetrack:
  • the bottom picture can be substituted into the () of the top picture to give a parse with an unambiguous meaning.
  • a track can contain only one thought, so the second verb "drank" signals a return to the main track and initiates hierarchy-building as a verb:
  • Cognitive parsers are distinct from syntactic, semantic, and conceptual parsers because the final result is in cognitive form (a tree of symbols for pictures and symbols for relations) rather than: a head-and- modifier structure (such as a diagrammed sentence), a tree of syntactic categories, a tree of semantic categories, a network of semantic primitives, or a network of conceptual primitives or dependencies.
  • Cognitive parsers are further distinguished from syntactic, semantic, and conceptual parsers in using symbols for relations to initiate parsing procedures, to determine the applicable rule for selecting the start- and stop-sites of a parsing procedure, and as the actual start- and stop-sites for a parsing procedure.
  • Cognitive parsers are additionally distinct from syntactic parsers because: they do not use information about the lexical category of a word or the syntactic category of a phrase; they do not use information about agreement in number, gender, case, or tense; they do not distinguish between different nouns, unless a relation has been omitted; and most cognitive parsing operations do not distinguish between nouns and pronouns, adjectives and articles, or verbs and prepositions.
  • Cognitive parsers are further distinguished from syntactic parsers in not identifying functional roles of words and phrases, such as subjects and predicates or agents and objects. Cognitive parsers are additionally distinguished from syntactic parsers in inserting words and symbols not present in the original sentence.
  • Cognitive parsers are additionally distinct from pattern-matching or transition- network syntactic parsers, since they construct a hierarchical parse. They are distinct from recursive transition networks in using nesting rather than loops, so that, for example, multiple adjectives are nested rather than concatenated. Cognitive parsers are distinct from immediate-constituent syntactic parsers in using a single immediate- constituent rule (for English, P rel P — . P) for multiple levels of a hierarchical parse.
  • Cognitive parsers are distinct from transformational-grammar parsers, since they do not perform word-order transformations. They are distinct from phrase structure grammar parsers, since they do not use derived syntactic categories to leave gaps for subsequent long-distance dependencies. Cognitive parsers are distinct from augmented transition networks, since they do not use tables of word features and roles. They are distinct from tree-adjoining-grammar parsers in not beginning with a set of preconstructed syntactic trees. They are distinct from Marcus-type deterministic syntactic parsers in not constructing a tree of syntactic categories. Cognitive parsers are distinct from parsers that use tables of word co-occurrence frequencies in that they do not rely on such tables.
  • Cognitive parsers are additionally distinct from syntactic parsers supplemented by semantic interpretation because they do not use pre-constructed semantic networks, set membership hierarchies, or word-based selection-restriction rules. Indeed, they are insensitive to most semantic information: they do not distinguish between "The squirrel buried a nut” and “My aunts sent a telegram,” and they allow "Colorless green ideas sleep furiously.” For the same reason, cognitive parsers are distinct from devices that construct predicate-argument structures. Cognitive parsers are further distinguished from many prior-art semantic interpretation parsers in being generative rather than matching template-patterns for semantic case, mood, transitivity, theme, information, or voice. Cognitive parsers do not build parsed structures by assembling overlapping parse fragments.
  • Cognitive parsers are distinct from semantic parsers because they are insensitive to most semantic information; as noted above, they do not distinguish between "The squirrel buried a nut” and "My uncles sent a telegram.” Cognitive parsers are additionally distinct from semantic parsers because they are not limited to parsing words; sequences of icons or sign language configurations can also be parsed.
  • Cognitive parsers are distinct from conceptual analyzers because of their insensitivity to most semantic information. They are additionally distinct from conceptual analyzers because they preserve word order. Cognitive parsers furthermore do not use pre-constructed templates for commonly-encountered messages; conceptual selection-restriction rules; or conceptual case structures. They also do not seek key words on which to base slot-filler structures.
  • the procedures that the cognitive parser of the present invention uses generate several novel results unanticipated from the prior art.
  • Certain word combinations automatically leave gaps at sites of long-distance dependencies, which a semantic or conceptual analyzer can fill in.
  • the cognitive parser insert words or symbols for missing relations. Disambiguation is avoidable for a large fraction of cognitive parsing; for many symbolic expressions, disambiguation in fact occurs as a side-product of parsing.
  • Cognitive parsing entails a novel class of rules (hierarchy-building and picture-border procedures) which govern the parser's grouping of newly-built structures.
  • the resulting cognitive forms differ from standard grammar representations of sentences; for example, multiple adjectives are nested rather than concatenated, and adverb phrases are related to entire preceding statement rather than modifying just the verb.
  • Related syntactic structures such as passive voice, perfect tense, and indirect object phrasings — become parsed in a way showing that they are conceptually different rather than just syntactically rearranged.
  • a parser for English sentences, processing time varies linearly with respect to sentence length.
  • FIG. 1 shows the overall structure of an illustrative embodiment of a cognitive parser, a cognitive parser for English sentences.
  • FIG. 2 shows a flowchart of the instruction set of the cognitive parser of FIG. 1, which manipulates symbols in the computing element workspace during cognitive parsing.
  • FIG. 3 shows the structure of the shift-registers for words used by the cognitive parser of FIG. 1, the contents of which are manipulated during cognitive parsing in the illustrative embodiment.
  • FIG. 4 shows a flowchart of the picture- symbol grouping procedure of the FIG. 2 instruction set, the core operation in cognitive parsing.
  • FIG. 5 shows a flowchart of the hierarchy-building procedure of the FIG. 2 instruction set, in which low-level pictures are combined into higher-level ones.
  • FIG. 6 shows a flowchart of the parallel-track procedure of the FIG. 2 instruction set, in which control passes to additional word shift-registers.
  • FIG. 7 shows a flowchart of the track-folding step of the FIG. 6 parallel- track procedure, in which the contents of the additional word shift-registers are combined back into the main shift-register.
  • FIG. 8 shows a flowchart of the comma processing procedure of the FIG. 2 instruction set, in which a comma initiates one of several processing pathways.
  • FIG. 9 shows a flowchart of the 2-P procedure of the FIG. 2 instruction set, in which the presence of two successive words for pictures initiates one of several processing pathways.
  • FIG. 10 shows a flowchart of the relative-pronoun processing procedure of the FIG. 2 instruction set, in which a relative pronoun initiates one of several processing pathways.
  • FIG. 11 shows a specific example of the operation of the sidetrack shift- registers of FIG. 3 during cognitive parsing.
  • FIG. 12 shows a specific example of the operation of the subtrack shift- registers of FIG. 3 during cognitive parsing.
  • FIG. 13 shows illustrative devices that incorporate a cognitive parser such as the one shown in FIG. 1.
  • FIG. 1 shows the overall structure of a cognitive parser for parsing symbol sequences.
  • the cognitive parser includes an input device 1, such as a keyboard, for inputting a sentence or other symbol sequence 2, a computing device 3 which includes a microprocessor executing an instruction set 4 (also referred to herein as "parsing procedure") for sentence-processing steps, memory, preferably an array of shift registers 5 within which words or symbols are placed and moved as the sentence is parsed, a lexicon look-up function 6 for seeking information from a lexicon 7, preferably stored in memory.
  • an input device such as a keyboard
  • a computing device 3 which includes a microprocessor executing an instruction set 4 (also referred to herein as "parsing procedure") for sentence-processing steps, memory, preferably an array of shift registers 5 within which words or symbols are placed and moved as the sentence is parsed, a lexicon look-up function 6 for seeking information from a lexicon 7, preferably stored in memory.
  • an instruction set 4 also
  • the cognitive parser of the present invention also includes an output device 9, such as a display screen, for outputting the parsed sentence or other parsed symbol sequence 8.
  • output device 9 can also be devices such as a voice-recognition system, devices for electronic transfer of text, or a printer.
  • Entries of the lexicon 7 contain: the words or other symbols (which, for simplicity, will be referred to as "words" in the following); notation whether the word represents a picture, a relation, or a picture accompanied by a relation (the entity-type); notation of the word's position on a continuum between sensory-related and abstract; and auxiliary information about the word's hierarchy- building properties and parallel-track utilization.
  • FIG. 2 shows the symbol processing steps of the instruction set 4.
  • a START module 10 sets up initial parameters, such as declaring and initializing variables, inserting spaces to separate punctuation from letters, reading a user's choice of screen display options, and defining the array of shift registers 5.
  • a picture-grouping procedure 12 reads one word of the sentence 1 and looks it up in the lexicon 7. Subsequent processing of each word follows one of the possible paths (arrows and procedure boxes) from the picture-grouping procedure 12 to a STOP module 28, or back to the picture-grouping procedure 12 whereupon the next word is read. The particular path followed depends on decisions made within each procedure module, as described in detail in subsequent Figures.
  • a decision step in the picture-grouping procedure 12 notes whether the newly-read word denotes a picture or a relation. If the word is a picture- word, it is placed into the shift register 5 in an unfilled picture-position of an alternating pattern of "picture-word/relation- word/picture-word/relation-word ." Fitting this pattern may require that a gap be left.
  • a second decision step determines whether the previous word was also a picture. If not, the next word is read. If so, control passes to a 2-P procedure 24, which usually inserts the missing relation and returns control to the picture- grouping procedure 12, where the next word is read.
  • control passes to a relative pronoun processing procedure 26 for insertion of the missing relation. After the missing relation is inserted, control returns to the picture-grouping procedure 12, where the next word is read.
  • the decision step in the picture-grouping procedure 12 notes that the newly- read word denotes a relation, it is placed into the shift register 5 in an unfilled relation- position of the alternating "picture- word/relation- word/picture-word/relation- word." pattern.
  • the picture-grouping procedure 12 uses the newly placed relation- word to mark a boundary between picture-words, thereby parsing the sentence 1 into pictures and relations. For example, if parentheses are used, the result is (P) rel (P) rel
  • This procedure examines the current and preceding parsed structures to find a (P) rel (P) sequence. If found, the sequence is marked to build a ((P) rel (P)) picture-symbol group. Another decision step then determines whether the newly-read relation was one of several special relations, to be described hereinafter. If not, control returns to the picture-grouping procedure 12, where the next word is read. If so, control passes to a parallel-track procedure 20.
  • the parser begins utilizing a new workspace in the register 5 (a sidetrack 36 or a subtrack 38; see FIG.
  • Control then passes to the picture- grouping procedure 12, where the next word is read. If the special relation is a comma, control passes from the hierarchy-building procedure 14 to a comma processing procedure 22, which has its own hierarchy- building, subtrack, and sidetrack procedures. Control then returns to the picture- grouping procedure 12, where the next word is read. If the special relation is a period, the hierarchy-building procedure 14 reiterates as many times as possible, the parallel- track procedure 20 rejoins all sidetracks and subtracks to the original track, and control passes to a STOP module 28. The STOP module puts the parsed sentence 8 into a readily readable format, displays it on the output device 9, and halts processing.
  • FIG. 3 shows the structure of the shift-register for words 5.
  • Each 1- dimensional portion of the shift-register will be referred to as a "track”; most processing occurs on the main track 30.
  • Each track is divided into alternating compartments, termed P-boxes 32 & 35 and rel-boxes 34.
  • the cognitive parser places the first picture-denoting word of the sentence 1 into the left-most P-box 32 of the main track 30; the first relation- word into the left-most rel-box 34; the second P- word into the second P-box 35; and so forth.
  • Tracks adjacent to the main track 30 are termed sidetracks 36 & 37; tracks below the main track are termed subtracks 38 & 39.
  • Subtracks can have sidetracks and vice versa, so the entire 3-D array is filled with tracks.
  • Certain relation-words initiate a shift in track usage so that, for example, after partially filling the main track 30 the cognitive parser begins placing the remaining words of a sentence 1 on subtracks 38 and 39.
  • FIG. 4 shows the detailed structure of the picture- symbol grouping procedure 12 of FIG. 2.
  • sentence-processing is described in terms of operations on the shift-register for words 5, but other methods of grouping symbols and inserting missing relations could be used.
  • the picture- symbol grouping procedure 12 begins at step 40 by placing a "(" into the first P-box 32.
  • the first word or punctuation symbol of the sentence 2 is read.
  • Step 44 accesses the lexicon lookup procedure 6, which then looks up this word or punctuation symbol in the lexicon 7.
  • step 46 determines that the lexicon has marked the word as being better processed after being replaced by other words, as for an adjective denoting a picture plus a relation
  • step 48 replaces the original word by its replacement(s) from the lexicon and step 42 re-reads it.
  • the parser reads the word's entity-type entry in the lexicon 7. If the entity-type is "picture,” step 52 copies the word in the first P-box 32 of the main track 30. It also notes the current P-box's location in a variable such as "lastploc.” If step 50 finds that the word's entity-type is "relation,” step 62 then copies the word in the first rel-box 34. The current rel-box's location is noted in a variable such as "lastrloc.” If step 50 finds that the word can be either a picture or a relation, the word is copied into the left-most box of either type that does not yet contain a word. That is, the parser acts as if the newly-read word is of the type required to fit an unbroken (P) rel (P) rel (P) ... pattern.
  • P unbroken
  • Step 54 determines whether the word preceding a newly-read P-word was another P. If not, then step 56 determines whether the word is a relative pronoun preceded by a comma. If not, then step 58 advances the word reader to the next word or symbol in the sentence 2. If so, step 60 retains the relative pronoun's referent by replacing the relative pronoun with the previous picture constructed by the cognitive parser. If step 54 finds that the word preceding a P-word was also a P-word, control passes to the 2-P stage (24 of FIG. 2; FIG. 9).
  • step 62 has copied the newly-read relation- word into a rel-box.
  • step 64 then puts a ")" after the P- word in the P-box 32 that lies to the left of the current rel-box 34.
  • This step also places a "(" in the empty P-box 35 that lies to the right of the current rel-box 34.
  • Control then passes to the hierarchy-building stage (14 of FIG. 2; FIG. 5).
  • step 52 copying P-words into the left-most P-box not already containing a word and step 62 copying rel-words into the left-rel P-box not already containing a word.
  • FIG. 5 shows the detailed structure of the hierarchy-building procedure 14.
  • Step 70 looks up the current relation-word's lexicon entry, to determine whether the word hierarchy-builds completely (periods, verbs, higher relations, path prepositions), strongly (most prepositions), or weakly (binding prepositions such as "of"), or is a comma. Each hierarchy-building property results in control being passed to a different one of three hierarchy-building modules, 72 & 74 or 72-76-74 or 72-78-74. If step 70 determines that a complete builder is a verb, it also adds 1 to a variable such as "vbct" for later use. If step 70 determines that the relation is a comma, control passes to the comma processing procedure (22 of FIG. 2; FIG. 8).
  • step 72 determines whether the shift register 5 contains to the left of the current relation- word: a P-word, a previous relation- word, and another P-word. If not, control returns to step 58 of the picture-grouping procedure 12, where the next word is read.
  • step 58 determines whether the shift register 5 contains a P-word, relation- word, and another P-word to the left of the current relation- word.
  • the parser tests for the presence of a word that acts as a picture-boundary 76 & 78. This picture-boundary condition determines where leftward hierarchy-building stops, and varies from one module to the next as seen by comparing steps 76 and 78.
  • step 76 of the parser tests for a "(the)" not already incorporated into a larger (P rel P).
  • step 78 tests whether the previous relation is a relation other than "1". If a picture-boundary is found, control returns to step 58 of the picture-grouping procedure 12 where the next word is read. Third, if step 76 or step 78 determines that no picture-boundary is present, then step 74 brackets the P rel P with parentheses to give (P rel P).
  • bracketing is achieved by i) shifting the previous relation- word (e.g. the one in rel-box 34) and second P (e.g. the one in P-box 35) into the P-box containing the first P (e.g. P-box 32) and ii) placing a "(" and ")" at the left and right ends, respectively, of that P- box (e.g. P-box 32).
  • the current relation is then also shifted leftward to the rel-box formerly occupied by the previous relation (e.g. rel-box 34).
  • step 74 After thus hierarchy-building once, the parser loops from step 74 back to step 72 to find additional sets of P, rel, and P to the left of the current relation-word. For complete builders, this loop continues until the beginning of the sentence 2 is reached; in addition, step 72 passes control to the parallel-track procedure (20 of FIG. 2; FIG. 6).
  • FIG. 6 shows the detailed structure of the parallel-track procedure 20.
  • the cognitive parser reaches this point after hierarchy-building for a complete-builder relation.
  • Step 80 then tests whether this relation is the second verb on the current track, because such verbs are signals for ending tracks. If not, step 82 tests whether the relation is a sentience verb, higher relation, or period by looking up the word in the lexicon 7. If none of these is the case, control returns to step 58 of the picture- grouping procedure 12 where the next word is read.
  • step 84 initiates two new subtracks for processing the remainder of the sentence. If processing is currently occurring on the main track 30, these subtracks are 38 & 39 (FIG. 3).
  • the sentience verb is moved to the first rel-box of the next subtrack down (current subtrack+1); next the sentience verb's "(" is moved to the first P-box of the second subtrack down (current subtrack+2). Details are given in the description of FIG. 12. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence 1 are processed on the subtrack+2.
  • the sentience verb on subtrack+1 connects the main track 30 to subtrack+2.
  • step 86 initiates two new sidetracks such as 36 & 37.
  • the higher relation is moved to the first rel-box of the next sidetrack over (current sidetrack+1); then the higher relation's "(" is moved to the first P-box of the second sidetrack over (current sidetrack+2). Details are given in the description of FIG. 11. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence are processed into sidetrack+2.
  • the higher relation on sidetrack+1 connects the main track 30 to sidetrack+2.
  • step 92 determines whether processing is occurring on the main track 30. If not, processing on the subtrack or sidetrack will end and that track will be rejoined to the previous track by a track-folding procedure (94 and FIG. 7). Since multiple layers of tracks can have been created during processing, track rejoining repeats until step 92 finds that the main track 30 has been reached. Control then passes to step 96, which determines whether this point was reached because the current relation is a period, as just discussed, or because it was the second verb on a track (discussed next). If a period, control passes to a STOP module 28 which prepares the now completely parsed sentence 9 for an output device 8.
  • Step 80 can find that the current relation- word is the second verb on the current track, a signal to end that track.
  • Step 92 determines whether processing is occurring on the main track 30. If not, processing on the subtrack or sidetrack will end and that track will be rejoined to the previous track by the track-folding procedure (94 and FIG. 7). Track rejoining repeats until step 92 finds that the main track 30 has been reached. Control then passes to step 96, which determines whether this point was reached because the c rent relation was the second verb on a track, or because it was a period. If the relation was the second verb on a track, control passes to step 82.
  • FIG. 7 shows the detailed structure of the track-folding procedure, step 94 of FIG. 6.
  • step 100 rejoins the current subtrack to the next-shallowest subtrack by moving the following items to the lowest available rel and P boxes of the next- shallowest subtrack:
  • the connecting relation (the relation that created the subtrack, either a sentience verb, as discussed with FIG. 6, or a 1 relation, to be discussed hereinafter); the subtrack's P; the current relation (the relation initiating track-folding); and the "(" inserted by that relation. Details are given in the description of FIG. 12.
  • a hierarchy-building module 72 & 74 then hierarchy-builds completely.
  • Step 102 tests whether processing is now on the shallowest subtrack of the current sidetrack. If not, processing returns to step 100 and the operations repeat until step 102 finds that all subtracks have been rejoined to the sidetrack.
  • Step 104 then rejoins the sidetrack to its next-shallowest sidetrack by similar operations.
  • the following items are moved to the first available rel and P boxes of the next-shallowest sidetrack: The sidetrack's connecting relation (the higher relation that created the sidetrack); the sidetrack's P; the current relation (the relation initiating track- folding); and the "(" inserted by that relation. Details are given in the description of FIG. 11.
  • Another hierarchy-building module 72 & 74 then hierarchy-builds completely.
  • Control then returns to step 92 (FIG. 6) in order to determine whether the main track 30 has now been reached. If not, the subtrack and sidetrack rejoining procedures repeat, beginning with step 100, until step 92 (FIG. 6) determines that all tracks have been rejoined to the main track.
  • FIG. 8 shows the detailed structure of the comma processing procedure 22 of FIG. 2.
  • the cognitive parser reaches this point after the hierarchy-building stage 14 has determined that the current relation is a comma.
  • the cognitive parser uses the lexicon 7 to determine the type of word following the comma.
  • step 114 stores the assembled picture for later use by pronouns in step 162 of FIG. 10. Step 116 then initiates two new sidetracks for processing the remainder of the sentence.
  • the comma is moved to the first rel-box of the next sidetrack over (current sidetrack+1; e.g. 36 if processing is currently on the main track 30).
  • the comma is then changed to a "cents" sign so that the parser can later distinguish it from other commas; this choice of replacement symbol is arbitrary.
  • the comma's "(" is moved into the first P-box of the second sidetrack over (current sidetrack+2; e.g. 37). Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence 2 are processed on the sidetrack+2.
  • the comma on sidetrack+1 connects the previous track, e.g. 30, to sidetrack+2.
  • step 110 determines that the word after the comma is a verb
  • the parser assumes that the comma is the final comma of a parenthetical expression or an appositive. As a result, it assumes that processing has been occurring on a sidetrack that should now be terminated.
  • the parser therefore executes hierarchy-building steps 72 & 74 to completely build the pictures preceding the comma.
  • step 118 tests whether processing is occurring on the main track 30; if not, a track-folding procedure 120 is initiated.
  • Procedure 120 is the same as track-folding procedure 94 (FIG. 7) for folding of subtracks, such as 38 & 39. However, there is no folding of sidetracks such as 36 & 37. Because parenthetical expressions and appositives are separate from a sentence's main thought, the parser leaves them in their sidetrack, e.g. 37, rather than being rejoined to a shallower sidetrack. Specifically, the connecting comma (the previous comma that initiated the sidetrack) and the P are not moved. The current comma and its "(" are deleted. The following verb and its "(" are placed in the first available rel- and P-boxes of the next-shallowest sidetrack.
  • step 118 Track-folding repeats until step 118 determines that the main track 30 has been reached. Because the verb following the comma was entered into the register at step 110, step 118 also advances the word counter by one so that the next- word reader will skip a word. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read. The remaining words of the sentence 2 are processed into the main track 30.
  • FIG. 9 shows the detailed structure of the 2-P procedure 24. If the picture- grouping procedure 12 (FIG. 2) has determined that there have been two picture- words in a row, step 130 then uses the lexicon 7 to determine whether one of the picture-words is a relative pronoun. If so, control passes to the relative-pronoun processing procedure 26 (FIG. 2). If not, step 132 determines whether the second picture- word - the current word - is a determiner such as "a" or "the".
  • step 134 examines the lexicon 7 to determine each picture-word's cognitive category. These categories are pre-assigned a rank ranging from sensation-related to abstract concepts. Other similar lists could be derived either from a general principle or from empirical comparison to the resulting accuracy in parsing a particular language. The more specific or sensation-related of the two P's is parsed as a component of the other. Thus, if the first picture-word is more sensation-related (as in "red windmill"), step 136 inserts the symbols ), 1, and ( between the two picture-words, in the appropriate P- and rel-boxes of the register. The same relation is inserted if the two words have the same category ("dog house").
  • a hierarchy-building module 72-78-74 then assembles the pictures preceding the newly- inserted relation into larger pictures.
  • Step 78 tests for a picture-border, ending hierarchy-building at the first relation other than ].
  • step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture-grouping procedure 12 where the next word is read.
  • step 138 inserts the symbols ), [, and (.
  • Hierarchy-building module 72-140-74 follows, and step 140 allows hierarchy-building of pictures preceding the inserted relation only if the left-hand P ⁇ the P preceding the first of the 2-P's — is not a determiner and the relation to its right is [ or I.
  • step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture- grouping procedure 12 where the next word is read.
  • Step 132 could find that the second of the 2-P's is a determiner (e.g., the third and fourth words of "bought his wife a card"). If so, step 142 inserts the symbols ), T, and (. A hierarchy-building module 72-144-74 assembles the pictures preceding this inserted relation until step 144 determines that a verb has been reached; the verb acts as a picture border. When step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture-grouping procedure 12 where the next word is read.
  • a determiner e.g., the third and fourth words of "bought his wife a card"
  • FIG. 10 shows the detailed structure of the relative-pronoun processing procedure 26.
  • step 150 uses the lexicon 7 to determine the pattern fit by the two P's. If the pattern is "that"-P, with the P unquantitated (not plural), then the parser treats the relative pronoun as a demonstrative adjective.
  • Step 152 inserts the symbols ), [, and ( into the appropriate P- and rel-boxes of the register between the two picture-words. There is no hierarchy-building and control returns directly to step 58 of the picture-grouping procedure 12 where the next word is read.
  • step 150 determines that the pattern is relative pronoun/P, with P quantitated, or is P/relative pronoun then step 154 inserts the symbols ), 1, and (.
  • Step 158 then initiates two new subtracks for processing the remainder of the sentence.
  • the 1 is moved down to the next subtrack (cu ⁇ ent subtrack+1), in the first rel-box; then the second P and its "(" are moved to the second subtrack down (current subtrack+2), in the first P-box.
  • the 1 on subtrack+1 serves to connect the previous track to subtrack+2.
  • step 160 uses the lexicon 7 to determine whether the second P is a relative pronoun. If not, control simply returns to step 58 of the picture-grouping procedure 12 where the next word is read. If so, and if the replacement option has been chosen, step 162 replaces the relative pronoun by the picture built in the register 5 just prior to beginning the subtrack. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read. For either decision made by step 160, the remaining words of the sentence are processed into subtrack+2.
  • the cognitive parser reads a sentence, or other symbolic expression of thought, one symbol at a time.
  • the parser's structure results in a series of procedures that enact the cognitive rules (Terminology).
  • the result is a sentence or other expression that has been parsed into cognitive form, a hierarchy of (P rel P) structures.
  • the hierarchical structure of this form is represented by the placement of markers, here parentheses.
  • the procedures manipulate words within the shift-registers for words 5 (FIG. 3). The operation of the cognitive parser, and the procedures it uses, will be illustrated with two example sentences.
  • FIG. 11 shows the basic picture-grouping operations, and additionally shows the sidetrack procedures.
  • the example sentence 2 is "The squrrrel buried a nut because Fall came.”
  • the first P-box 32 of the register's main track 30 is initialized with "(", which results from the relation ".” ending the preceding sentence.
  • the parser's picture-grouping stage 12 finds in the lexicon 7 that the first word, "The”, represents a picture. The parser therefore places this word into the first P-box 32, after the parenthesis.
  • the second word, "squirrel" is also found to represent a picture and so is placed into the second P-box 35.
  • FIG. 11F shows that the parser shifts once to build "[(a) [ (nut)]”
  • FIG. 11G shows that it shifts again to build " ⁇ [(The) [ (squirrel)] buried [(a) [ (nut)] ⁇ ".
  • FIG. 11H shows that the parser initiates the new sidetracks 36 & 37 by moving "because” from the first rel-box 34 of the main track 30 to the first rel-box of the first sidetrack 36.
  • the "(" contributed by "because” is moved into the first P- box of the second sidetrack 37.
  • FIG. 11 1 shows that the parser adds the next two words, "Fall” and "came”, to the new sidetrack 37 and processes them in the usual way.
  • the final period also contributes two parentheses. Because "came” is intransitive, it has no object.
  • the result is the form "()" in the second P-box of sidetrack 37. This form is a space for the "itself” that is the unspoken object of an intransitive verb.
  • the period then instructs the hierarchy-building procedure 14 to go to completion.
  • the period first hierarchy-builds in its own sidetrack 37.
  • FIG. UK shows that the track-folding procedure 94 then folds the contents of the sidetracks 36 & 37 back into the vacant rel- and P- boxes 34 & 35, respectively, of the main track 30.
  • the higher relation "because" has returned to the position 34 it occupied in FIG. 11G, but it is now followed by words already built into a picture.
  • FIG. 11L shows that the hierarchy-building procedure 14 continues on the main track 30, with the result residing in the main track's first picture- box 32. This result is the original sentence now parsed into cognitive form, achieved by a regular and deterministic series of procedures.
  • FIG. 12 shows the operation of the subtrack procedures.
  • the example sentence 2 taken from a predicate logic text, is "The man who wrote Waverley drank port.” It is an example of what linguists consider an embedded sentence: "who wrote Waverley” is embedded in the larger sentence.
  • FIG. 12A shows that the parser processes "The man” in the usual way. At “who”, the parser finds in the lexicon 7 that the word is a pronoun representing a picture; the parser places "who” into the third P- box.
  • the parser's 2-P procedure 24 (FIG.
  • FIG. 12B shows that the relative pronoun processing procedure 26 then initiates usage of subtracks.
  • the parser moves 1 from its rel-box on the main track 30 into the first rel-box of the first subtrack 38; in addition, "(who” is shifted into the first P-box of the second subtrack 39.
  • FIG. 12C shows that the parser then enters and processes the words "wrote Waverley drank" on the new subtrack 39 in the usual way.
  • the second verb on the subtrack 39 signals the end of that subtrack. That is, the second verb instructs the parallel-track procedure 20 to terminate the subtrack 39 and rejoin subtracks 38 & 39 to the main track 30.
  • FIG. 12D shows that the parallel-track procedure 20 therefore first hierarchy-builds to completion on the verb's own track 39.
  • FIG. 12E shows that the parser next folds the contents of the subtracks 38 & 39 back into the lowest vacant rel- and P-boxes of the main track 30.
  • FIGS. 12F & 12G show that the verb "drank” also instructs the hierarchy-building procedure to go to completion on the main track 30, using two shift-and-hierarchy-build operations.
  • the sequence of operations driven by the second verb, "drank" is now complete.
  • FIG. 12H shows that the parser then reads "port” and the period in the usual way. In the next step, the period would cause hierarchy- building 14 to go to completion (as in FIG. 11L).
  • FIG. 13A shows the structure of command-operated devices, which use cognitively-parsed natural language to generate the specific commands native to the device. These commands drive simple devices, such as kitchen appliances, lighting, heating or air conditioning, and videocassette recorders, or more complex devices such as computers.
  • a command in natural language 2 is the input to a cognitive parser 170 of the type described in FIGS. 1-12.
  • the parsed symbol sequence 8 outputted from the cognitive parser is the input to a command- translation unit 172 containing a small lexicon listing equivalent phrasings for each of the limited number of native commands used by the device.
  • the command-translation unit 172 outputs a command native to the controlled device 174.
  • FIG. 13B shows the structure of a natural language translation device, for which the input symbol sequence 2 is a natural language text.
  • the cognitive parser 170 produces a parsed symbol sequence 8, using a syntactic and semantic analyzer unit 176 to resolve words that can be either a picture or a relation, rather than using syntactic and semantic analysis to decide among grammar rules.
  • the parsed symbol sequence 8 is then the input into a natural language translation unit 178, which translates the expression in the first language to an equivalent expression in a second language and displays the result.
  • FIG. 13C shows the structure of a knowledge- storage device that stores facts in cognitive form.
  • the cognitive parser 170 produces a parsed symbol sequence in cognitive form 8, again using a syntactic and semantic analyzer unit 176 to resolve entity-type ambiguities.
  • the parsed symbol sequence is the input to a storage unit 180, which thus stores in cognitive form the facts in the input text 2.
  • FIG. 13D shows the structure of a logical inference device that uses a cognitive parser.
  • the input is a natural language text representing facts 2, which the cognitive parser 170 and syntactic-semantic analyzer unit 176 convert to a parsed symbol sequence in cognitive form 8.
  • This symbol sequence is the input to a semantic representation unit 182, including a lexicon that lists the semantic primitives and logical operations equivalent to potential input symbol sequences.
  • Various equivalent natural language inputs 2 thus lead to the same semantic representation output.
  • This semantic representation is the input to a logical inference engine 184, which performs logical deductions from stored and newly-presented facts.
  • the cognitive parser provides a simple means of parsing natural language and other symbolic expressions of thought using a small number of stored rules, minimal stored semantic information, and minimal requirement for disambiguation, and is capable of parsing while the symbolic expression is being entered into the device. Its operation is not limited to a specific knowledge base or to anticipatable thoughts.
  • the cognitive parser's simplicity allows it to be used in personal computers and in non- computer devices equipped with only minimal computing power.
  • the parsed natural language structure produced would make it easier for untrained users to use a computer for storing information, querying a database, or finding logical inferences. By replacing multiple layers of menus, it would make computers easier to use even for trained users and allow control of kitchen appliances, videocassette recorders, home lighting, and heating and air conditioning.
  • Another use for the parsing operations would be as a programming language that resembled natural language.
  • cognitive form reflects the structure of the thoughts underlying language more accurately than does traditional grammar, the cognitive parser would be useful in teaching grammar. Because those sentences which can be parsed using cognitive rules without semantic analysis are easiest to understand, the cognitive parser could be used as a reading-level checker.
  • the cognitive parser could parse non-grammatical or idiomatic sentences. Combination with a device performing semantic interpretation or conceptual dependency analysis could constitute an input to a device performing logical inference involving the properties of objects.
  • input and output could use voice-recognition or speech-production devices, electronic scanning or transfer of text, graphical display of the hierarchy of symbol groups, or a printer. Processing could begin during sentence entry or after a period.
  • the lexicon entries could be arranged by individual words or word roots.
  • An ordinary dictionary-like lexicon could be used by modifying the cognitive parsing rules to treat nouns, pronouns, and adjectives as pictures and to treat verbs and prepositions as relations.
  • the parsing operations and registers could be implemented in hardware rather than software, by direct manipulation of memory addresses and pointers rather than by manipulating words or parentheses, or in any instruction format acceptable to the computing element instead of in a high-level programming language.
  • the parsing operations could be implemented as immediate-constituent production rules rather than as procedures.
  • Markers for the borders of symbol groups could be overlines, nested boxes, or other symbols different from parentheses.
  • the parser could place grouping-markers by calculating marker positions rather than by manipulating words, or symbol group borders could be embodied by discontinuities in computing element addresses rather than by markers.
  • Word manipulation could be performed by manipulating computing element addresses for words rather than manipulating the words themselves.
  • Symbols parsed could be icons or American sign language configurations rather than words. Languages with sentence patterns different from those of English could be parsed by grouping symbols into (P P rel) or (rel P P) symbol groups.
  • the device could also be operated in reverse for sentence production, using as input a symbol sequence divided into symbol-groups and having one word designated as the sentence's subject.

Abstract

A parsing method and apparatus for symbolic expressions of thought such as English-language sentences is provided. The basic procedure distinguishes only between symbols for pictures (such as 'squirrel' or 'justice') and symbols for relations (such as 'above' or 'pushed'). For example, to the cognitive parser, the sentences 'The squirrel buried a nut' and 'My aunts sent a telegram' are equivalent. The parser thus operates at a level more basic than syntactic parsing, making it simpler. The illustrative embodiment, a cognitive parser for English sentences, comprises: a microprocessor (3), a stored lexicon (7) including symbols and associated entity-types (picture or relation), an input device (1) for a symbol sequence (2), and a procedure (4) executing on the microprocessor for grouping the inputted symbols according to the rules based on entity type. A method comprising operation of the cognitive parser is also provided.

Description

METHOD AND DEVICE FOR PARSING NATURAL LANGUAGE SENTENCES AND OTHER SEQUENTIAL SYMBOLIC EXPRESSIONS
Field of the Invention
The invention relates to a method and device for parsing natural language sentences and other symbolic expressions, and more specifically, the invention relates to a method and device for parsing natural language sentences and other symbolic expressions into cognitive units.
Background of the Invention
A parsing device converts a linear symbol sequence into an organized structure. The diagrammed sentence encountered in grade-school grammar is one kind of parse. This organized structure can then be displayed for instructional purposes, or other devices can use it to initiate commands, store information, or make logical inferences. A parsing device is a prerequisite for using natural language to communicate with computers and, moreover, a simple parsing device is needed if natural language is to be used for controlling household devices such as kitchen appliances, videocassette recorders, lighting, and heating or air conditioning. The central problem in parsing is to achieve both accuracy and speed. For synthetic languages, such as computer languages, many effective solutions are available. However, the complexity and ambiguity of natural languages have made them resistant to efficient parsing.
Automated sentence parsing devices — usually general purpose digital computers — have operated by matching parts of the input sentence with a very large number of stored rules. If the first rule does not apply, the next is chosen for examination. Three broad classes of devices have been used; these employ increasingly deep levels of sentence analysis. The most common approach is Syntactic Parsing, in which the rules driving the parsing device describe the organization of words in a particular language. In practice, achieving a parse also requires resolving the possible senses of ambiguous words. This is done by a semantic interpreter, which uses information about word meaning. Semantic Parsing aims at the level of meaning directly, driving the parser with rules relating ordered words to the underlying meaning of the sentence. Conceptual Analysis also aims directly for the level of meaning, but does not attempt to preserve word order. Conceptual analysis parsers identify key words as tokens for the underlying objects or actions, and the organization sought is that of the objects being discussed rather than that of the words used to discuss them. The processor reconstructs this organization by using "encyclopedia information" about the objects, rather than "dictionary information" about the words.
In each approach, the stored rules can be either template-patterns, which must be matched exactly, or general rules, which are applied to parts of the sentence in succession to build a tree-like parsed structure of increasingly finer detail, from a sentence to its phrases to words. The successive-rule strategy, termed "generative," is combinatorial since rules can be used more than once in various combinations. Rules can be chosen so as to divide a sentence into smaller units ("top-down parsing") or, less frequently, to assemble words into higher-level units ("bottom-up" parsing). By storing the result of each successful pattern-matching or rule application, the parse accumulates aspects of the sentence's organization.
The disadvantages of present parsing methods are complexity, large required processor size, limited vocabulary, and slow speed. These disadvantages stem from the large number of syntactic rules used and the incorporation of semantic information to resolve ambiguities.
What is desired, therefore, is a parsing method and device which would utilize a small number of rules, minimal semantic information, and minimal computing power. Such parsing method and device would require minimal disambiguation of words having multiple meanings and would not rely on specific knowledge bases, word cooccurrence probability data, selection-restrictions, frames, or expectations about sentence content. Such parsing method and device would also process symbol strings in a time proportional to sentence length.
Summary of the Invention
Accordingly, it is an object of the invention to provide a system to parse thought expressions, such as natural language, using a small number of rules, thereby minimizing rule-choice decisions.
Another object of the invention is to provide a system to parse thought expressions, such as natural language, using minimal semantic information.
Still another object of the invention is to provide a system to parse thought expressions, such as natural language, using minimal computing power.
Yet another object of the invention is to provide a system to parse thought expressions, such as natural language, with minimal disambiguation of words having multiple meanings.
Another object of the invention is to provide a system to parse thought expressions, such as natural language, without limiting the vocabulary to a specific knowledge base.
Still another object of the invention is to provide a system of the above character to parse even surprising thought expressions by avoiding word co-occurrence probability data, selectional restrictions, frames, and expectations.
Yet another object of the invention is to provide a system of the above character to parse while the expression is being entered into a processor.
Another object of the invention is to provide a system of the above character to parse in a time period corresponding linearly to sentence length.
The invention is a system for parsing symbol sequences representing thoughts, using a small set of rules herein termed "cognitive rules." The basic rule distinguishes only between symbols for pictures and symbols for relations. In addition, the rules use the symbols for relations as signals for particular parsing procedures. In the illustrative embodiment of the invention discussed hereinbelow, which processes English sentences, the parser: i) divides the sentence into words for pictures and words for the relations between pictures and then ii) assembles picture-relation-picture symbol sequences into symbol groups that constitute new picture symbols, using rules for detecting picture borders. The invention is thus a "bottom-up generative cognitive parser."
TERMINOLOGY
Against a background of complex natural language parsers is introduced cognitive parsing, which distinguishes primarily between symbols for pictures (such as "squirrel" or "justice") and symbols for relations (such as "above" or "pushed"). In the general sense used here, a picture can be a sound, feeling, or past experience as well as a visual image. When the symbols are words, such pictures are symbolized by nouns, adjectives, and pronouns, including relative pronouns such as "that" and "which." Relations are typically symbolized by prepositions and verbs. The underlying idea of cognitive parsing can be quickly conveyed by noting that the cognitive parser does not distinguish between "The squirrel buried a nut" and "My aunt sent a telegram." In both cases a relation (the verb) separates two pictures, thereby parsing them. Cognitive parsing operates at a level more basic than syntactic parsing or conceptual analysis. This simplicity allows it to use only three rules, together with about a dozen criteria for deciding when to apply these rules.
The cognitive parser's fundamental operations are a) scanning the input symbol sequence for "picture" and "relation" symbols, b) using the relation symbols to segregate the picture symbols, and c) assembling picture and relation symbol sequences into symbol groups that constitute new picture symbols. Since the distinction between "picture" and "relation" is a primitive form of meaning, and symbol order is preserved, a cognitive parser used for words is a type of semantic parser. The primitive nature of these cognitive rules also allows a cognitive parser to apply them to thought- expressions other than words, such as a string of signs in American Sign Language. The terminology used to describe the operations performed by the parser is understood in terms of the desired final parsed structure, termed "cognitive form."
Cognitive Form
The sentence "Squirrels bury nuts" can be divided into picture-relation-picture form: (Squirrels) bury (nuts). The verbal relation "bury" acts as a parsing element in languages that use this word order. The picture-relation-picture form, hereafter abbreviated "P rel P," can itself be a "picture", (P rel P). Such a picture is referred to herein as a "picture-symbol group"; e.g., ((Squirrels) bury (nuts)). More complex sentences can be shown to have the same (P rel P) form, with various other types of relations such as prepositions acting as parsing elements: ((Squirrels) bury (nuts)) under (trees). [((Squirrels) bury (nuts)) under (trees)] when [(it) is (Fall)]. The square parentheses are used for clarity and do not differ in principle from ordinary parentheses. These structures can be written using an immediate-constituent rule: P rel P — > P. This rule differs from those of immediate-constituent grammars in the prior art because i) it always generates a tri-forked tree and ii) each level of the tree uses the same immediate-constituent rule.
Rules similar to those for the illustrative embodiment, which parses English sentences, could be used to parse a language or symbol system whose cognitive form is instead constructed of, for example, (P P rel) symbol groups. Such rule modifications would be apparent to one skilled in linguistics in view of the cognitive parsing procedures described herein; therefore, the procedures for their implementation will not be discussed herein.
One reason that the usefulness of (P rel P) structures for parsing has not been noticed is that the relations are often omitted from sentences. A phrase such as "red squirrel" appears to lack on its face a P rel P structure. However, upon closer examination, the P rel P structure is evident since the phrase means "(redness) component-of (squirrel)". That is, an adjective carries with it both a picture and the relation needed to complete the P rel P structure. If, following Frege, the number two is treated as representing the set of all pairs, then "two red squirrels" is: (two) composed-of [(red) component-of (squirrels)]. For simplicity, "composed-of ' will be abbreviated [, which can be thought of as enfolding the component. "Component-of is abbreviated 1. The result is: (two) [ [(red) 1 (squirrels)]. Gerunds, participles, and infinitives also have a (P rel P) structure. The gerund "burying" means "something buries something," which has the form (P verb P). The two "something"-s are of course omitted in English. The full form accounts for the ability of "burying" to have an object or adverb.
For some sentences, it is appropriate to have the parser place portions of the sentence on separate processing tracks, separate but identical workspaces. The parallel processing tracks are useful for commas and relative clauses. For example, the two statements in "The man, who wrote Waverley, drank port" are independent and are thus merely associated with each other. The sentence is represented in cognitive form as follows: {[(The) T (man)] drank [port]} associated-with { [who] wrote [Waverley] }
The three lines of this form are generated on separate tracks in the cognitive parser. The final structure is a (P rel P) form: { } associated-with { } .
Since easy-to-understand sentences have a nested (P rel P) structure, this structure can be used as a well-formedness criterion to fill in the missing portions of poorly-formed sentences. For example, "They made him director" must be missing a relation between "him" and "director." Since the job "director" has been instantiated with "him," the missing relation is "component-of." Similarly, the well-formedness criterion facilitates disambiguation. In "He saw the saw," "(He)" is a picture and so should be followed by a relation such as the verb sense of "saw." Because "the" represents "(the) I ", it should be followed by a picture such as the noun sense of "saw." Disambiguation between different noun or verb senses (e.g., river bank and financial institution) is not needed, because cognitive parsing distinguishes only between pictures and relations.
Picture-Symbol Grouping Procedure
The (P rel P) form motivates a simple procedure to parse symbolic expressions. Instead of, for example, dividing a sentence into subjects, predicates, objects, and clauses, it is divided into pictures. This can be done readily, since between words for any two pictures lies a word for a relation. A local rule can be used to implement what will be termed the picture- symbol grouping procedure: The present embodiment of the parser writes outward-facing parentheses on both sides of any word that denotes a relation. For example: "Squirrels) bury (nuts". The result is that each parenthesis indicates the border of a picture-word; a right-facing parenthesis indicates the beginning, and a left-facing parenthesis indicates the end. Note that the choice of parentheses as markers herein is arbitrary; therefore, other markers may be used to denote the same thing.
An unanticipated consequence of this parsing procedure is that gaps are left for the pictures missing from sentences having intransitive verbs and imperatives: (He) ran (). () Answer (the telephone). Such gaps, created by more elaborate methods in some syntactic parsing methods of the prior art, are useful for subsequent semantic interpretation. Hierarchy-Building Procedure
The picture- symbol grouping procedure, alone, gives a flat structure, such as "(Squirrels) bury (nuts) before (Winter)." Higher- level pictures are built by a hierarchy-building procedure in which, each time a relation is encountered, a hierarchy- building decision is made for the pictures that preceded it: Each time the parser encounters a relation, it not only adds the outward-facing parentheses but may also build a new picture by bracketing any P rel P preceding the relation.
This augmentation gives: (Squirrels) bury (nuts
[(Squirrels) bury (nuts)] before ( {[(Squirrels) bury (nuts)] before (Winter)}.
The additional parentheses thus indicate borders for larger picture- symbol groups.
Picture-Borders
Relation- words can be complete, strong, or weak picture builders depending on whether successive hierarchy-building continues all the way back to the beginning of the sentence or stops at a prior word that acts as a border. In the currently preferred embodiment, complete builders are: the period between sentences; verbs; higher relations (relations that can join statements, such as "because"); path prepositions (such as "through"); and the comma at the end of a parenthetical expression. Strong builders include most prepositions, and are blocked by "the." Weak builders are blocked by any relation except 1; these prepositions, such as "of," closely relate the pictures following and preceding. Additional borders arising in other situations can be determined by one skilled in the art, according to the method of the present invention.
2-P Procedure
The parser uses the (P rel P) well-formedness criterion to regenerate missing relations. When two successive picture-words appear in a sentence, the (P rel P) criterion tells us to insert a relation. The relation omitted in English seems to be always either 1 or | .
We will term this the 2-P procedure: If two P's are encountered in succession, the parser inserts either "composed of or "component of between them. The more specific or sensation-related of the two P's is presumed to be a component of the other. If they are equally specific, the relation is 1. If the second P is a determiner, the relation is T. Note that one of skill in the art could derive various scales of abstractness from references such as Roget's Thesaurus or the machine-readable edition of the Longman Dictionary of Contemporary English. The 2-P procedure generates: (red) 1 (windmill); (the) T (windmill); they made ... (him) [ (director); our ... (son) [ (John); (dog) 1 (house).
The illustrative embodiment of the parser uses the following empirically-derived hierarchy-building borders: When the 2-P procedure inserts 1, hierarchy-building builds weakly.
When the 2-P procedure inserts ϊ at "P-determiner," hierarchy-building is blocked by verbs.
When the 2-P procedure inserts ϊ at non-determiner P's, hierarchy-building is blocked by determiners and by all relations except 1 and [.
When the 2-P procedure inserts [ at "that P" with P unquantified (i.e., "that" used as a demonstrative adjective), there is no hierarchy-building.
Parallel-Track Procedures
In statements, the verb asserts that the clause is true. Picture hierarchies can be built from pairs of such statements, as in "(The squirrel buried a nut) because (it was Fall)." In these situations, the currently preferred embodiment of the parser uses a higher relation, a comma, or a relative pronoun as a signal to switch from building a (P V P) picture on one processing track to building a second (P V P) picture on another processing track.
[(The squirrel) buried (a nut)] because [(it) was (Fall)].
Each parallel track, which we will term a "sidetrack," is picture-grouped separately. At the end of the sentence, the two thoughts are joined by rejoining the sidetracks: {[(The squirrel) buried (a nut)] because [(it) was (Fall)] }.
The same procedures result in correct parsing of non-essential relative clauses, as in "The man, who wrote Waverley, drank port." The commas initiate a sidetrack, preventing "man"and "who" from being adjacent and thus preventing the 2-P procedure from being triggered inappropriately: {[(The) T (man)] drank [port]} associated-with { [who] wrote [Waverley] }
"Who" can be replaced by the picture preceding it if desired; here, the previous picture is "the man."
Introductory phrases are also correctly parsed by these procedures, and circumvent a hierarchy-building conflict. For example, the sentence "He planned the intermission between shows" can have a different meaning from "Between shows, he planned the intermission." To parse the second picture, the phrase after the comma is put on a sidetrack:
[() Between (shows)] associated-with [(he) planned (the intermission)].
Once parsing is complete, the bottom picture can be substituted into the () of the top picture to give a parse with an unambiguous meaning.
Clauses that are not asserted to be true - "embedded sentences" — are parsed using a second category of parallel track which will be termed a "subtrack". These are initiated by: a) a relative pronoun (creating an essential clause) or b) a class of verbs that can be followed by a second action. The latter - such as "make," "want," "know," and "is" — express causation, volition, emotion, or being rather than action, and will be termed "sentience verbs."
In "The man who wrote Waverley drank port", for example, the combination "man who" triggers the 2-P procedure. The relation inserted is 1, since "man" is more specific than the pronoun: (The) [ (man) 1 (who. Hierarchy-building at such relative pronoun 2-P insertions is blocked by a verb or a "the" that is not already part of a picture.
The relative pronoun "who" then initiates a subtrack connected to the main track by the inserted (. There is no hierarchy-building: (The) T (man)
1 (who Parsing continues on the subtrack: (The) T (man)
1
[(who) wrote (Waverley)] drank (
A track can contain only one thought, so the second verb "drank" signals a return to the main track and initiates hierarchy-building as a verb:
(The) T (man) 1 [(who) wrote (Waverley)] drank ( {(The) T <(man) 1 [(who) wrote (Waverley)]>] drank (
This rejoining behavior differs from that of a sidetrack. If desired, "who" could have been replaced by the picture previous to it, "man":
{ [(The) T [(man) ] [(man) wrote (Waverley)]]] drank (port)}.
This parse is the correct cognitive form, because the man is being defined by his participation in the activity of writing Waverley. For this reason, the scope of "the" is the entire clause "man who wrote Waverley."
Sentience verbs are simpler, since they hierarchy-build completely when starting the subtrack:
(I) know { (you) read [(the) ( (book)] } .
DISTINCTIONS FROM PRIOR APPROACHES TO SENTENCE PARSING
Many differences are apparent between the procedures used by a cognitive parser of the present invention and those used in the prior art. Cognitive parsers are distinct from syntactic, semantic, and conceptual parsers because the final result is in cognitive form (a tree of symbols for pictures and symbols for relations) rather than: a head-and- modifier structure (such as a diagrammed sentence), a tree of syntactic categories, a tree of semantic categories, a network of semantic primitives, or a network of conceptual primitives or dependencies. Cognitive parsers are further distinguished from syntactic, semantic, and conceptual parsers in using symbols for relations to initiate parsing procedures, to determine the applicable rule for selecting the start- and stop-sites of a parsing procedure, and as the actual start- and stop-sites for a parsing procedure. Cognitive parsers are additionally distinct from syntactic parsers because: they do not use information about the lexical category of a word or the syntactic category of a phrase; they do not use information about agreement in number, gender, case, or tense; they do not distinguish between different nouns, unless a relation has been omitted; and most cognitive parsing operations do not distinguish between nouns and pronouns, adjectives and articles, or verbs and prepositions. Cognitive parsers are further distinguished from syntactic parsers in not identifying functional roles of words and phrases, such as subjects and predicates or agents and objects. Cognitive parsers are additionally distinguished from syntactic parsers in inserting words and symbols not present in the original sentence.
Cognitive parsers are additionally distinct from pattern-matching or transition- network syntactic parsers, since they construct a hierarchical parse. They are distinct from recursive transition networks in using nesting rather than loops, so that, for example, multiple adjectives are nested rather than concatenated. Cognitive parsers are distinct from immediate-constituent syntactic parsers in using a single immediate- constituent rule (for English, P rel P — . P) for multiple levels of a hierarchical parse.
Cognitive parsers are distinct from transformational-grammar parsers, since they do not perform word-order transformations. They are distinct from phrase structure grammar parsers, since they do not use derived syntactic categories to leave gaps for subsequent long-distance dependencies. Cognitive parsers are distinct from augmented transition networks, since they do not use tables of word features and roles. They are distinct from tree-adjoining-grammar parsers in not beginning with a set of preconstructed syntactic trees. They are distinct from Marcus-type deterministic syntactic parsers in not constructing a tree of syntactic categories. Cognitive parsers are distinct from parsers that use tables of word co-occurrence frequencies in that they do not rely on such tables.
Cognitive parsers are additionally distinct from syntactic parsers supplemented by semantic interpretation because they do not use pre-constructed semantic networks, set membership hierarchies, or word-based selection-restriction rules. Indeed, they are insensitive to most semantic information: they do not distinguish between "The squirrel buried a nut" and "My aunts sent a telegram," and they allow "Colorless green ideas sleep furiously." For the same reason, cognitive parsers are distinct from devices that construct predicate-argument structures. Cognitive parsers are further distinguished from many prior-art semantic interpretation parsers in being generative rather than matching template-patterns for semantic case, mood, transitivity, theme, information, or voice. Cognitive parsers do not build parsed structures by assembling overlapping parse fragments.
Cognitive parsers are distinct from semantic parsers because they are insensitive to most semantic information; as noted above, they do not distinguish between "The squirrel buried a nut" and "My aunts sent a telegram." Cognitive parsers are additionally distinct from semantic parsers because they are not limited to parsing words; sequences of icons or sign language configurations can also be parsed.
Cognitive parsers are distinct from conceptual analyzers because of their insensitivity to most semantic information. They are additionally distinct from conceptual analyzers because they preserve word order. Cognitive parsers furthermore do not use pre-constructed templates for commonly-encountered messages; conceptual selection-restriction rules; or conceptual case structures. They also do not seek key words on which to base slot-filler structures.
The procedures that the cognitive parser of the present invention uses generate several novel results unanticipated from the prior art. Certain word combinations automatically leave gaps at sites of long-distance dependencies, which a semantic or conceptual analyzer can fill in. At some word combinations, the cognitive parser insert words or symbols for missing relations. Disambiguation is avoidable for a large fraction of cognitive parsing; for many symbolic expressions, disambiguation in fact occurs as a side-product of parsing. Cognitive parsing entails a novel class of rules (hierarchy-building and picture-border procedures) which govern the parser's grouping of newly-built structures. The resulting cognitive forms differ from standard grammar representations of sentences; for example, multiple adjectives are nested rather than concatenated, and adverb phrases are related to entire preceding statement rather than modifying just the verb. Related syntactic structures — such as passive voice, perfect tense, and indirect object phrasings — become parsed in a way showing that they are conceptually different rather than just syntactically rearranged. In the illustrative embodiment, a parser for English sentences, processing time varies linearly with respect to sentence length. Brief Description of the Drawings
FIG. 1 shows the overall structure of an illustrative embodiment of a cognitive parser, a cognitive parser for English sentences.
FIG. 2 shows a flowchart of the instruction set of the cognitive parser of FIG. 1, which manipulates symbols in the computing element workspace during cognitive parsing.
FIG. 3 shows the structure of the shift-registers for words used by the cognitive parser of FIG. 1, the contents of which are manipulated during cognitive parsing in the illustrative embodiment.
FIG. 4 shows a flowchart of the picture- symbol grouping procedure of the FIG. 2 instruction set, the core operation in cognitive parsing.
FIG. 5 shows a flowchart of the hierarchy-building procedure of the FIG. 2 instruction set, in which low-level pictures are combined into higher-level ones.
FIG. 6 shows a flowchart of the parallel-track procedure of the FIG. 2 instruction set, in which control passes to additional word shift-registers.
FIG. 7 shows a flowchart of the track-folding step of the FIG. 6 parallel- track procedure, in which the contents of the additional word shift-registers are combined back into the main shift-register.
FIG. 8 shows a flowchart of the comma processing procedure of the FIG. 2 instruction set, in which a comma initiates one of several processing pathways.
FIG. 9 shows a flowchart of the 2-P procedure of the FIG. 2 instruction set, in which the presence of two successive words for pictures initiates one of several processing pathways.
FIG. 10 shows a flowchart of the relative-pronoun processing procedure of the FIG. 2 instruction set, in which a relative pronoun initiates one of several processing pathways.
FIG. 11 shows a specific example of the operation of the sidetrack shift- registers of FIG. 3 during cognitive parsing. FIG. 12 shows a specific example of the operation of the subtrack shift- registers of FIG. 3 during cognitive parsing.
FIG. 13 shows illustrative devices that incorporate a cognitive parser such as the one shown in FIG. 1.
Detailed Description of the Invention
FIG. 1 shows the overall structure of a cognitive parser for parsing symbol sequences. The cognitive parser includes an input device 1, such as a keyboard, for inputting a sentence or other symbol sequence 2, a computing device 3 which includes a microprocessor executing an instruction set 4 (also referred to herein as "parsing procedure") for sentence-processing steps, memory, preferably an array of shift registers 5 within which words or symbols are placed and moved as the sentence is parsed, a lexicon look-up function 6 for seeking information from a lexicon 7, preferably stored in memory.
The cognitive parser of the present invention also includes an output device 9, such as a display screen, for outputting the parsed sentence or other parsed symbol sequence 8. It should be apparent to those skilled in the art that input and output devices 1 & 9 can also be devices such as a voice-recognition system, devices for electronic transfer of text, or a printer. Entries of the lexicon 7 contain: the words or other symbols (which, for simplicity, will be referred to as "words" in the following); notation whether the word represents a picture, a relation, or a picture accompanied by a relation (the entity-type); notation of the word's position on a continuum between sensory-related and abstract; and auxiliary information about the word's hierarchy- building properties and parallel-track utilization.
The operation of the instruction set 4 and the array of shift registers 5 will be described in greater detail hereinfollowing.
FIG. 2 shows the symbol processing steps of the instruction set 4. A START module 10 sets up initial parameters, such as declaring and initializing variables, inserting spaces to separate punctuation from letters, reading a user's choice of screen display options, and defining the array of shift registers 5. A picture-grouping procedure 12 reads one word of the sentence 1 and looks it up in the lexicon 7. Subsequent processing of each word follows one of the possible paths (arrows and procedure boxes) from the picture-grouping procedure 12 to a STOP module 28, or back to the picture-grouping procedure 12 whereupon the next word is read. The particular path followed depends on decisions made within each procedure module, as described in detail in subsequent Figures.
After a word has been read and looked up in the lexicon 7, a decision step in the picture-grouping procedure 12 notes whether the newly-read word denotes a picture or a relation. If the word is a picture- word, it is placed into the shift register 5 in an unfilled picture-position of an alternating pattern of "picture-word/relation- word/picture-word/relation-word ...." Fitting this pattern may require that a gap be left. A second decision step determines whether the previous word was also a picture. If not, the next word is read. If so, control passes to a 2-P procedure 24, which usually inserts the missing relation and returns control to the picture- grouping procedure 12, where the next word is read. If, however, one of the two successive P- words can be a relative pronoun, control passes to a relative pronoun processing procedure 26 for insertion of the missing relation. After the missing relation is inserted, control returns to the picture-grouping procedure 12, where the next word is read.
If the decision step in the picture-grouping procedure 12 notes that the newly- read word denotes a relation, it is placed into the shift register 5 in an unfilled relation- position of the alternating "picture- word/relation- word/picture-word/relation- word...." pattern. The picture-grouping procedure 12 then uses the newly placed relation- word to mark a boundary between picture-words, thereby parsing the sentence 1 into pictures and relations. For example, if parentheses are used, the result is (P) rel (P) rel
(p) ....
After a relation-word has been read and the boundary marked, control passes to a hierarchy-building procedure 14. This procedure examines the current and preceding parsed structures to find a (P) rel (P) sequence. If found, the sequence is marked to build a ((P) rel (P)) picture-symbol group. Another decision step then determines whether the newly-read relation was one of several special relations, to be described hereinafter. If not, control returns to the picture-grouping procedure 12, where the next word is read. If so, control passes to a parallel-track procedure 20. Here, the parser begins utilizing a new workspace in the register 5 (a sidetrack 36 or a subtrack 38; see FIG. 3), or terminates use of a subtrack or a sidetrack and returns to utilizing the original one (the main track 30). Control then passes to the picture- grouping procedure 12, where the next word is read. If the special relation is a comma, control passes from the hierarchy-building procedure 14 to a comma processing procedure 22, which has its own hierarchy- building, subtrack, and sidetrack procedures. Control then returns to the picture- grouping procedure 12, where the next word is read. If the special relation is a period, the hierarchy-building procedure 14 reiterates as many times as possible, the parallel- track procedure 20 rejoins all sidetracks and subtracks to the original track, and control passes to a STOP module 28. The STOP module puts the parsed sentence 8 into a readily readable format, displays it on the output device 9, and halts processing.
FIG. 3 shows the structure of the shift-register for words 5. Each 1- dimensional portion of the shift-register will be referred to as a "track"; most processing occurs on the main track 30. Each track is divided into alternating compartments, termed P-boxes 32 & 35 and rel-boxes 34. The cognitive parser places the first picture-denoting word of the sentence 1 into the left-most P-box 32 of the main track 30; the first relation- word into the left-most rel-box 34; the second P- word into the second P-box 35; and so forth. Tracks adjacent to the main track 30 are termed sidetracks 36 & 37; tracks below the main track are termed subtracks 38 & 39. Subtracks can have sidetracks and vice versa, so the entire 3-D array is filled with tracks. Certain relation-words, as will be described hereinafter, initiate a shift in track usage so that, for example, after partially filling the main track 30 the cognitive parser begins placing the remaining words of a sentence 1 on subtracks 38 and 39.
FIG. 4 shows the detailed structure of the picture- symbol grouping procedure 12 of FIG. 2. For clarity, sentence-processing is described in terms of operations on the shift-register for words 5, but other methods of grouping symbols and inserting missing relations could be used. The picture- symbol grouping procedure 12 begins at step 40 by placing a "(" into the first P-box 32. At step 42, the first word or punctuation symbol of the sentence 2 is read. Step 44 accesses the lexicon lookup procedure 6, which then looks up this word or punctuation symbol in the lexicon 7. If step 46 determines that the lexicon has marked the word as being better processed after being replaced by other words, as for an adjective denoting a picture plus a relation, then step 48 replaces the original word by its replacement(s) from the lexicon and step 42 re-reads it.
At step 50, the parser reads the word's entity-type entry in the lexicon 7. If the entity-type is "picture," step 52 copies the word in the first P-box 32 of the main track 30. It also notes the current P-box's location in a variable such as "lastploc." If step 50 finds that the word's entity-type is "relation," step 62 then copies the word in the first rel-box 34. The current rel-box's location is noted in a variable such as "lastrloc." If step 50 finds that the word can be either a picture or a relation, the word is copied into the left-most box of either type that does not yet contain a word. That is, the parser acts as if the newly-read word is of the type required to fit an unbroken (P) rel (P) rel (P) ... pattern.
Step 54 determines whether the word preceding a newly-read P-word was another P. If not, then step 56 determines whether the word is a relative pronoun preceded by a comma. If not, then step 58 advances the word reader to the next word or symbol in the sentence 2. If so, step 60 retains the relative pronoun's referent by replacing the relative pronoun with the previous picture constructed by the cognitive parser. If step 54 finds that the word preceding a P-word was also a P-word, control passes to the 2-P stage (24 of FIG. 2; FIG. 9).
The core function of the picture- grouping stage 12 occurs after step 62 has copied the newly-read relation- word into a rel-box. Step 64 then puts a ")" after the P- word in the P-box 32 that lies to the left of the current rel-box 34. This step also places a "(" in the empty P-box 35 that lies to the right of the current rel-box 34. The words up to and including the current word have now been parsed into pictures and relations. Control then passes to the hierarchy-building stage (14 of FIG. 2; FIG. 5).
Each remaining word of the sentence 2 is processed similarly, but with step 52 copying P-words into the left-most P-box not already containing a word and step 62 copying rel-words into the left-rel P-box not already containing a word.
FIG. 5 shows the detailed structure of the hierarchy-building procedure 14. Step 70 looks up the current relation-word's lexicon entry, to determine whether the word hierarchy-builds completely (periods, verbs, higher relations, path prepositions), strongly (most prepositions), or weakly (binding prepositions such as "of"), or is a comma. Each hierarchy-building property results in control being passed to a different one of three hierarchy-building modules, 72 & 74 or 72-76-74 or 72-78-74. If step 70 determines that a complete builder is a verb, it also adds 1 to a variable such as "vbct" for later use. If step 70 determines that the relation is a comma, control passes to the comma processing procedure (22 of FIG. 2; FIG. 8).
Hierarchy-building modules appear at many locations in the cognitive parser, and have three parts. First, step 72 determines whether the shift register 5 contains to the left of the current relation- word: a P-word, a previous relation- word, and another P-word. If not, control returns to step 58 of the picture-grouping procedure 12, where the next word is read. Second, if the shift register 5 does contain a P-word, relation- word, and another P-word to the left of the current relation- word, the parser then tests for the presence of a word that acts as a picture-boundary 76 & 78. This picture-boundary condition determines where leftward hierarchy-building stops, and varies from one module to the next as seen by comparing steps 76 and 78. For complete-builder relations 72 & 74, there is no test for a picture boundary because hierarchy-building goes to completion. For strong builders, step 76 of the parser tests for a "(the)" not already incorporated into a larger (P rel P). For weak builders, step 78 tests whether the previous relation is a relation other than "1". If a picture-boundary is found, control returns to step 58 of the picture-grouping procedure 12 where the next word is read. Third, if step 76 or step 78 determines that no picture-boundary is present, then step 74 brackets the P rel P with parentheses to give (P rel P).
It should be apparent to those skilled in the art that many methods of bracketing can be envisioned other than those described herein. In the preferred embodiment, bracketing is achieved by i) shifting the previous relation- word (e.g. the one in rel-box 34) and second P (e.g. the one in P-box 35) into the P-box containing the first P (e.g. P-box 32) and ii) placing a "(" and ")" at the left and right ends, respectively, of that P- box (e.g. P-box 32). The current relation is then also shifted leftward to the rel-box formerly occupied by the previous relation (e.g. rel-box 34). These operations are described in detail in the section on the operation of the cognitive parser, Figs. 11 and 12.
After thus hierarchy-building once, the parser loops from step 74 back to step 72 to find additional sets of P, rel, and P to the left of the current relation-word. For complete builders, this loop continues until the beginning of the sentence 2 is reached; in addition, step 72 passes control to the parallel-track procedure (20 of FIG. 2; FIG. 6).
FIG. 6 shows the detailed structure of the parallel-track procedure 20. The cognitive parser reaches this point after hierarchy-building for a complete-builder relation. Step 80 then tests whether this relation is the second verb on the current track, because such verbs are signals for ending tracks. If not, step 82 tests whether the relation is a sentience verb, higher relation, or period by looking up the word in the lexicon 7. If none of these is the case, control returns to step 58 of the picture- grouping procedure 12 where the next word is read.
If step 82 finds the relation to be a sentience verb, step 84 initiates two new subtracks for processing the remainder of the sentence. If processing is currently occurring on the main track 30, these subtracks are 38 & 39 (FIG. 3). In general, in the illustrative embodiment, the sentience verb is moved to the first rel-box of the next subtrack down (current subtrack+1); next the sentience verb's "(" is moved to the first P-box of the second subtrack down (current subtrack+2). Details are given in the description of FIG. 12. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence 1 are processed on the subtrack+2. The sentience verb on subtrack+1 connects the main track 30 to subtrack+2.
Similarly, if step 82 finds the relation to be a higher relation, step 86 initiates two new sidetracks such as 36 & 37. The higher relation is moved to the first rel-box of the next sidetrack over (current sidetrack+1); then the higher relation's "(" is moved to the first P-box of the second sidetrack over (current sidetrack+2). Details are given in the description of FIG. 11. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence are processed into sidetrack+2. The higher relation on sidetrack+1 connects the main track 30 to sidetrack+2.
If step 82 finds the relation to be a period, step 92 determines whether processing is occurring on the main track 30. If not, processing on the subtrack or sidetrack will end and that track will be rejoined to the previous track by a track-folding procedure (94 and FIG. 7). Since multiple layers of tracks can have been created during processing, track rejoining repeats until step 92 finds that the main track 30 has been reached. Control then passes to step 96, which determines whether this point was reached because the current relation is a period, as just discussed, or because it was the second verb on a track (discussed next). If a period, control passes to a STOP module 28 which prepares the now completely parsed sentence 9 for an output device 8.
Step 80 can find that the current relation- word is the second verb on the current track, a signal to end that track. Step 92 then determines whether processing is occurring on the main track 30. If not, processing on the subtrack or sidetrack will end and that track will be rejoined to the previous track by the track-folding procedure (94 and FIG. 7). Track rejoining repeats until step 92 finds that the main track 30 has been reached. Control then passes to step 96, which determines whether this point was reached because the c rent relation was the second verb on a track, or because it was a period. If the relation was the second verb on a track, control passes to step 82. FIG. 7 shows the detailed structure of the track-folding procedure, step 94 of FIG. 6. First, step 100 rejoins the current subtrack to the next-shallowest subtrack by moving the following items to the lowest available rel and P boxes of the next- shallowest subtrack: The connecting relation (the relation that created the subtrack, either a sentience verb, as discussed with FIG. 6, or a 1 relation, to be discussed hereinafter); the subtrack's P; the current relation (the relation initiating track-folding); and the "(" inserted by that relation. Details are given in the description of FIG. 12. A hierarchy-building module 72 & 74 then hierarchy-builds completely. Step 102 tests whether processing is now on the shallowest subtrack of the current sidetrack. If not, processing returns to step 100 and the operations repeat until step 102 finds that all subtracks have been rejoined to the sidetrack.
Step 104 then rejoins the sidetrack to its next-shallowest sidetrack by similar operations. The following items are moved to the first available rel and P boxes of the next-shallowest sidetrack: The sidetrack's connecting relation (the higher relation that created the sidetrack); the sidetrack's P; the current relation (the relation initiating track- folding); and the "(" inserted by that relation. Details are given in the description of FIG. 11. Another hierarchy-building module 72 & 74 then hierarchy-builds completely.
Control then returns to step 92 (FIG. 6) in order to determine whether the main track 30 has now been reached. If not, the subtrack and sidetrack rejoining procedures repeat, beginning with step 100, until step 92 (FIG. 6) determines that all tracks have been rejoined to the main track.
FIG. 8 shows the detailed structure of the comma processing procedure 22 of FIG. 2. The cognitive parser reaches this point after the hierarchy-building stage 14 has determined that the current relation is a comma. At step 110 the cognitive parser uses the lexicon 7 to determine the type of word following the comma.
If that word is a P or a relation that is not a verb, the parser assumes that the comma is an initial or middle comma for a list, parenthetical expression, or appositive. The parser therefore executes hierarchy-building steps 72-112-74 to build the pictures preceding the comma. Hierarchy-building continues as long as step 112, a picture- border test, determines that the relation preceding the comma is [ or 1. When hierarchy-building stops, step 114 stores the assembled picture for later use by pronouns in step 162 of FIG. 10. Step 116 then initiates two new sidetracks for processing the remainder of the sentence. The comma is moved to the first rel-box of the next sidetrack over (current sidetrack+1; e.g. 36 if processing is currently on the main track 30). In the illustrative embodiment, the comma is then changed to a "cents" sign so that the parser can later distinguish it from other commas; this choice of replacement symbol is arbitrary. The comma's "(" is moved into the first P-box of the second sidetrack over (current sidetrack+2; e.g. 37). Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read, so that the remaining words of the sentence 2 are processed on the sidetrack+2. The comma on sidetrack+1 connects the previous track, e.g. 30, to sidetrack+2.
If step 110 instead determines that the word after the comma is a verb, the parser assumes that the comma is the final comma of a parenthetical expression or an appositive. As a result, it assumes that processing has been occurring on a sidetrack that should now be terminated. The parser therefore executes hierarchy-building steps 72 & 74 to completely build the pictures preceding the comma. When hierarchy- building is complete, step 118 tests whether processing is occurring on the main track 30; if not, a track-folding procedure 120 is initiated.
Procedure 120 is the same as track-folding procedure 94 (FIG. 7) for folding of subtracks, such as 38 & 39. However, there is no folding of sidetracks such as 36 & 37. Because parenthetical expressions and appositives are separate from a sentence's main thought, the parser leaves them in their sidetrack, e.g. 37, rather than being rejoined to a shallower sidetrack. Specifically, the connecting comma (the previous comma that initiated the sidetrack) and the P are not moved. The current comma and its "(" are deleted. The following verb and its "(" are placed in the first available rel- and P-boxes of the next-shallowest sidetrack. Track-folding repeats until step 118 determines that the main track 30 has been reached. Because the verb following the comma was entered into the register at step 110, step 118 also advances the word counter by one so that the next- word reader will skip a word. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read. The remaining words of the sentence 2 are processed into the main track 30.
FIG. 9 shows the detailed structure of the 2-P procedure 24. If the picture- grouping procedure 12 (FIG. 2) has determined that there have been two picture- words in a row, step 130 then uses the lexicon 7 to determine whether one of the picture-words is a relative pronoun. If so, control passes to the relative-pronoun processing procedure 26 (FIG. 2). If not, step 132 determines whether the second picture- word - the current word - is a determiner such as "a" or "the".
If the second picture- word is not a determiner, step 134 examines the lexicon 7 to determine each picture-word's cognitive category. These categories are pre-assigned a rank ranging from sensation-related to abstract concepts. Other similar lists could be derived either from a general principle or from empirical comparison to the resulting accuracy in parsing a particular language. The more specific or sensation-related of the two P's is parsed as a component of the other. Thus, if the first picture-word is more sensation-related (as in "red windmill"), step 136 inserts the symbols ), 1, and ( between the two picture-words, in the appropriate P- and rel-boxes of the register. The same relation is inserted if the two words have the same category ("dog house"). A hierarchy-building module 72-78-74 then assembles the pictures preceding the newly- inserted relation into larger pictures. Step 78 tests for a picture-border, ending hierarchy-building at the first relation other than ]. When step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture-grouping procedure 12 where the next word is read.
If, instead, the second picture-word is more sensation-related than the first (as in "the windmill"), step 138 inserts the symbols ), [, and (. Hierarchy-building module 72-140-74 follows, and step 140 allows hierarchy-building of pictures preceding the inserted relation only if the left-hand P ~ the P preceding the first of the 2-P's — is not a determiner and the relation to its right is [ or I. When step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture- grouping procedure 12 where the next word is read.
Step 132 could find that the second of the 2-P's is a determiner (e.g., the third and fourth words of "bought his wife a card"). If so, step 142 inserts the symbols ), T, and (. A hierarchy-building module 72-144-74 assembles the pictures preceding this inserted relation until step 144 determines that a verb has been reached; the verb acts as a picture border. When step 72 determines that hierarchy-building is complete, control returns to step 58 of the picture-grouping procedure 12 where the next word is read.
FIG. 10 shows the detailed structure of the relative-pronoun processing procedure 26. If the 2-P procedure 24 (FIG. 2) has determined that one of the two picture-words is a relative pronoun, step 150 then uses the lexicon 7 to determine the pattern fit by the two P's. If the pattern is "that"-P, with the P unquantitated (not plural), then the parser treats the relative pronoun as a demonstrative adjective. Step 152 inserts the symbols ), [, and ( into the appropriate P- and rel-boxes of the register between the two picture-words. There is no hierarchy-building and control returns directly to step 58 of the picture-grouping procedure 12 where the next word is read. If, instead, step 150 determines that the pattern is relative pronoun/P, with P quantitated, or is P/relative pronoun then step 154 inserts the symbols ), 1, and (. A hierarchy-building module 72-156-74 hierarchy-builds to the left of the inserted relation until step 156 determines that processing has reached either a verb or a "(the)" that is not already contained within a larger picture. Step 158 then initiates two new subtracks for processing the remainder of the sentence. In analogy to step 84 (FIG. 6), the 1 is moved down to the next subtrack (cuπent subtrack+1), in the first rel-box; then the second P and its "(" are moved to the second subtrack down (current subtrack+2), in the first P-box. The 1 on subtrack+1 serves to connect the previous track to subtrack+2.
It is often desirable, for later uses, to replace the relative pronoun with its referent. Therefore, before proceeding, step 160 uses the lexicon 7 to determine whether the second P is a relative pronoun. If not, control simply returns to step 58 of the picture-grouping procedure 12 where the next word is read. If so, and if the replacement option has been chosen, step 162 replaces the relative pronoun by the picture built in the register 5 just prior to beginning the subtrack. Control then returns to step 58 of the picture-grouping procedure 12 where the next word is read. For either decision made by step 160, the remaining words of the sentence are processed into subtrack+2.
Operation of Invention : FIGS . 11-12
The cognitive parser reads a sentence, or other symbolic expression of thought, one symbol at a time. The parser's structure (FIGS. 1-10) results in a series of procedures that enact the cognitive rules (Terminology). The result is a sentence or other expression that has been parsed into cognitive form, a hierarchy of (P rel P) structures. The hierarchical structure of this form is represented by the placement of markers, here parentheses. In the preferred embodiment, the procedures manipulate words within the shift-registers for words 5 (FIG. 3). The operation of the cognitive parser, and the procedures it uses, will be illustrated with two example sentences.
FIG. 11 shows the basic picture-grouping operations, and additionally shows the sidetrack procedures. The example sentence 2 is "The squrrrel buried a nut because Fall came." At the outset, FIG. 11 A, the first P-box 32 of the register's main track 30 is initialized with "(", which results from the relation "." ending the preceding sentence. The parser's picture-grouping stage 12 finds in the lexicon 7 that the first word, "The", represents a picture. The parser therefore places this word into the first P-box 32, after the parenthesis. The second word, "squirrel", is also found to represent a picture and so is placed into the second P-box 35. FIG. IIB shows that the parser next notes that two successive P-words have been entered, and so activates the 2-P procedure 24. Because the lexicon 7 states that "The" is more abstract than "squirrel", the 2-P procedure inserts the missing relation [ along with its outward- facing parentheses. The \ is inserted into the rel-box 34 between the two P-boxes. The ")" is inserted into the first P-box 32 after "(The", resulting in a completely- bracketed picture, "(The)". The "(" is inserted into the second P-box, resulting in "(squirrel". Note that the local contribution of one parenthesis by each relation results in long-range assembly of fully-bracketed pictures. In FIG. 11C, the parser determines that the next word, "buried", is a relation and so inserts it into the next available rel-box and also inserts the outward-facing parentheses.
Because the lexicon 7 states that "buried" is a verb, the parser next initiates the hierarchy-building procedure 14, with the result shown in FIG. 11D. " and "(squirrel)" are both shifted leftward into the first P-box 32, after (The)", "buried" is then shifted to the now-vacant first rel-box 34 and and its "(" is shifted to the now- vacant second P-box 35. Finally, an additional set of parentheses is added to the left and right ends of the first P-box 32, thereby assembling "[(The) [ (squirrel)]" into a picture-symbol group, a higher-level picture. In a similar way, "a" and "nut" are read and another missing [ is inserted as shown in FIG. HE. The inserted relation [ does not hierarchy-build "The squirrel buried a" into a larger picture, because the cognitive rules state that "a" acts as a picture-border for [ in this situation. The relation "because" is therefore read into the third rel-box, contributing its outward-facing parentheses.
The lexicon 7 states that "because" is a higher relation ~ a relation that joins two statements. As a result, hierarchy-building 14 goes to completion. FIG. 11F shows that the parser shifts once to build "[(a) [ (nut)]" and FIG. 11G shows that it shifts again to build "{[(The) [ (squirrel)] buried [(a) [ (nut)]}".
After initiating hierarchy-building, higher relations such as "because" transfer control of the parser to the parallel-track procedure 20 (FIG. 2), in order to use sidetracks as a way of parsing separately the two separate statements in the sentence. FIG. 11H shows that the parser initiates the new sidetracks 36 & 37 by moving "because" from the first rel-box 34 of the main track 30 to the first rel-box of the first sidetrack 36. In addition, the "(" contributed by "because" is moved into the first P- box of the second sidetrack 37. FIG. 11 1 shows that the parser adds the next two words, "Fall" and "came", to the new sidetrack 37 and processes them in the usual way. The final period also contributes two parentheses. Because "came" is intransitive, it has no object. The result is the form "()" in the second P-box of sidetrack 37. This form is a space for the "itself" that is the unspoken object of an intransitive verb.
The period then instructs the hierarchy-building procedure 14 to go to completion. As shown in FIG. 11J, the period first hierarchy-builds in its own sidetrack 37. FIG. UK shows that the track-folding procedure 94 then folds the contents of the sidetracks 36 & 37 back into the vacant rel- and P- boxes 34 & 35, respectively, of the main track 30. The higher relation "because" has returned to the position 34 it occupied in FIG. 11G, but it is now followed by words already built into a picture. Finally, FIG. 11L shows that the hierarchy-building procedure 14 continues on the main track 30, with the result residing in the main track's first picture- box 32. This result is the original sentence now parsed into cognitive form, achieved by a regular and deterministic series of procedures.
FIG. 12 shows the operation of the subtrack procedures. The example sentence 2, taken from a predicate logic text, is "The man who wrote Waverley drank port." It is an example of what linguists consider an embedded sentence: "who wrote Waverley" is embedded in the larger sentence. FIG. 12A shows that the parser processes "The man" in the usual way. At "who", the parser finds in the lexicon 7 that the word is a pronoun representing a picture; the parser places "who" into the third P- box. The parser's 2-P procedure 24 (FIG. 2) begins to insert the relation missing between "man" and "who"; since the lexicon 7 states that "who" is a relative pronoun, the relative pronoun processing procedure 26 becomes active and determines that this relation is 1. That is, the man is being defined as the one who participated in the action following "who". For this reason, the inserted 1 must not hierarchy-build "The man" into its own picture, separate from the one to follow. The cognitive rules achieve this because, in the relative pronoun processing procedure 26, "The" is a picture-border for ] in this situation.
FIG. 12B shows that the relative pronoun processing procedure 26 then initiates usage of subtracks. The parser moves 1 from its rel-box on the main track 30 into the first rel-box of the first subtrack 38; in addition, "(who" is shifted into the first P-box of the second subtrack 39. FIG. 12C shows that the parser then enters and processes the words "wrote Waverley drank" on the new subtrack 39 in the usual way.
Because a track can contain only one statement, the second verb on the subtrack 39, "drank", signals the end of that subtrack. That is, the second verb instructs the parallel-track procedure 20 to terminate the subtrack 39 and rejoin subtracks 38 & 39 to the main track 30. FIG. 12D shows that the parallel-track procedure 20 therefore first hierarchy-builds to completion on the verb's own track 39. FIG. 12E shows that the parser next folds the contents of the subtracks 38 & 39 back into the lowest vacant rel- and P-boxes of the main track 30.
Finally, FIGS. 12F & 12G show that the verb "drank" also instructs the hierarchy-building procedure to go to completion on the main track 30, using two shift-and-hierarchy-build operations. The sequence of operations driven by the second verb, "drank", is now complete. FIG. 12H shows that the parser then reads "port" and the period in the usual way. In the next step, the period would cause hierarchy- building 14 to go to completion (as in FIG. 11L).
These illustrations of the cognitive parser's operation should not be construed as indicating that the symbols used by the shift-register are limited to words, or that the symbols parsed by the cognitive parser are limited to words and sentences. Nor should these illustrations be construed as restricting the cognitive parser to devices implemented with shift-registers.
Devices Incorporating Invention: FIG. 13
As described in the Background, a cognitive parser can be incorporated as a key component of another device, simplifying and speeding its operation compared to devices using syntactic or semantic parsers alone. Four illustrative classes of device are shown in FIG. 13. FIG. 13A shows the structure of command-operated devices, which use cognitively-parsed natural language to generate the specific commands native to the device. These commands drive simple devices, such as kitchen appliances, lighting, heating or air conditioning, and videocassette recorders, or more complex devices such as computers. A command in natural language 2 is the input to a cognitive parser 170 of the type described in FIGS. 1-12. The parsed symbol sequence 8 outputted from the cognitive parser is the input to a command- translation unit 172 containing a small lexicon listing equivalent phrasings for each of the limited number of native commands used by the device. For a given natural language command, the command-translation unit 172 outputs a command native to the controlled device 174.
FIG. 13B shows the structure of a natural language translation device, for which the input symbol sequence 2 is a natural language text. The cognitive parser 170 produces a parsed symbol sequence 8, using a syntactic and semantic analyzer unit 176 to resolve words that can be either a picture or a relation, rather than using syntactic and semantic analysis to decide among grammar rules. The parsed symbol sequence 8 is then the input into a natural language translation unit 178, which translates the expression in the first language to an equivalent expression in a second language and displays the result.
FIG. 13C shows the structure of a knowledge- storage device that stores facts in cognitive form. Beginning with a natural language text representing facts 2, the cognitive parser 170 produces a parsed symbol sequence in cognitive form 8, again using a syntactic and semantic analyzer unit 176 to resolve entity-type ambiguities. The parsed symbol sequence is the input to a storage unit 180, which thus stores in cognitive form the facts in the input text 2.
FIG. 13D shows the structure of a logical inference device that uses a cognitive parser. As in FIG. 13C, the input is a natural language text representing facts 2, which the cognitive parser 170 and syntactic-semantic analyzer unit 176 convert to a parsed symbol sequence in cognitive form 8. This symbol sequence is the input to a semantic representation unit 182, including a lexicon that lists the semantic primitives and logical operations equivalent to potential input symbol sequences. Various equivalent natural language inputs 2 thus lead to the same semantic representation output. This semantic representation is the input to a logical inference engine 184, which performs logical deductions from stored and newly-presented facts.
Conclusion, Ramifications, and Scope of Invention
The cognitive parser provides a simple means of parsing natural language and other symbolic expressions of thought using a small number of stored rules, minimal stored semantic information, and minimal requirement for disambiguation, and is capable of parsing while the symbolic expression is being entered into the device. Its operation is not limited to a specific knowledge base or to anticipatable thoughts. The cognitive parser's simplicity allows it to be used in personal computers and in non- computer devices equipped with only minimal computing power.
The parsed natural language structure produced would make it easier for untrained users to use a computer for storing information, querying a database, or finding logical inferences. By replacing multiple layers of menus, it would make computers easier to use even for trained users and allow control of kitchen appliances, videocassette recorders, home lighting, and heating and air conditioning. Another use for the parsing operations would be as a programming language that resembled natural language. Because cognitive form reflects the structure of the thoughts underlying language more accurately than does traditional grammar, the cognitive parser would be useful in teaching grammar. Because those sentences which can be parsed using cognitive rules without semantic analysis are easiest to understand, the cognitive parser could be used as a reading-level checker. In combination with a device performing semantic interpretation, lexical-functional grammar, or preference semantics, the cognitive parser could parse non-grammatical or idiomatic sentences. Combination with a device performing semantic interpretation or conceptual dependency analysis could constitute an input to a device performing logical inference involving the properties of objects.
Although the description above contains many specificities, these should not be construed as limiting the scope of the invention but as merely illustrating capabilities of the presently preferred embodiment of the cognitive parser. It will be understood by those having skill in the art that changes can be made to the specific embodiment without departing from the spirit or scope of the invention. For example, input and output could use voice-recognition or speech-production devices, electronic scanning or transfer of text, graphical display of the hierarchy of symbol groups, or a printer. Processing could begin during sentence entry or after a period. The lexicon entries could be arranged by individual words or word roots. An ordinary dictionary-like lexicon could be used by modifying the cognitive parsing rules to treat nouns, pronouns, and adjectives as pictures and to treat verbs and prepositions as relations. The parsing operations and registers could be implemented in hardware rather than software, by direct manipulation of memory addresses and pointers rather than by manipulating words or parentheses, or in any instruction format acceptable to the computing element instead of in a high-level programming language. The parsing operations could be implemented as immediate-constituent production rules rather than as procedures.
Markers for the borders of symbol groups could be overlines, nested boxes, or other symbols different from parentheses. The parser could place grouping-markers by calculating marker positions rather than by manipulating words, or symbol group borders could be embodied by discontinuities in computing element addresses rather than by markers. Word manipulation could be performed by manipulating computing element addresses for words rather than manipulating the words themselves. Symbols parsed could be icons or American sign language configurations rather than words. Languages with sentence patterns different from those of English could be parsed by grouping symbols into (P P rel) or (rel P P) symbol groups. The device could also be operated in reverse for sentence production, using as input a symbol sequence divided into symbol-groups and having one word designated as the sentence's subject.
Accordingly, the scope of the invention should be determined by the appended claims and their legal equivalents, rather than by the embodiments illustrated.

Claims

What is claimed is:
1. An apparatus for parsing symbol sequences such as natural language sentences, comprising: a microprocessor; a memory addressable by said microprocessor; a lexicon, stored in said memory, including symbols and an entity-type coπesponding to each symbol, the entity-type selected from the group comprising a picture, a relation, and a picture accompanied by a relation; an input device for inputting the symbol sequence to said microprocessor; a procedure, executing on said microprocessor, for determining the entity-type coπesponding to each inputted symbol by searching said lexicon, and for grouping the inputted symbols according to rules based on entity-type; and an output device for outputting the symbol grouping.
2. The apparatus of Claim 1 , wherein said input device comprises a keyboard.
3. The apparatus of Claim 1, wherein said input device comprises a voice- recognition system.
4. The apparatus of Claim 1, wherein said output device comprises a display screen.
5. The apparatus of Claim 1 , wherein said output device comprises a software program for using the symbol grouping.
6. The apparatus of Claim 5, wherein said software program operates a control unit of another device.
7. The apparatus of Claim 1, wherein said output device comprises an information storage device.
8. The apparatus of Claim 1 , wherein said lexicon includes a cognitive category valuation for each symbol which is a picture, the cognitive category substantially ranging from sensation-related to abstract concepts; and wherein said program includes a consecutive picture-symbol routine that uses the cognitive category valuation for inserting a relation-symbol between consecutive symbols which are pictures.
9. The apparatus of Claim 1 , wherein said procedure includes a hierarchy-building routine for building a symbol group having pictures and a relation.
10. The apparatus of Claim 1 , wherein said lexicon includes a grouping property coπesponding to at least some of the symbols; and wherein said procedure also determines the grouping property coπesponding to an inputted symbol and also groups the inputted symbols according to the grouping property.
11. A lexicon for use in a parser for symbol sequences such as natural language sentences comprising: a memory device; and a table, stored in said memory device, including a plurality of symbols and an entity- type coπesponding to each symbol of the plurality of symbols, the entity- type selected from the group comprising a picture, a relation, and a picture accompanied by a relation.
12. The lexicon of Claim 11 , wherein said table further comprises a cognitive category valuation for each symbol of the plurality of symbols which is a picture, the cognitive category valuation substantially ranging from sensation-related to abstract concepts.
13. The lexicon of Claim 11 , wherein said table further comprises a grouping property associated with at least one of the plurality of symbols.
14. An apparatus operated by symbol sequences such as natural language sentences, comprising: a microprocessor; a memory addressable by said microprocessor; a lexicon, stored in said memory, including symbols and an entity-type coπesponding to each symbol, the entity-type selected from the group comprising a picture, a relation, and a picture accompanied by a relation; an input device for inputting the symbol sequence to said microprocessor; and a program, executing on said microprocessor, for determining the entity-type coπesponding to each inputted symbol by searching said lexicon, for grouping the inputted symbols by entity-type, and for generating an instruction for operating the apparatus in response to the symbol grouping.
15. The apparatus of Claim 14, wherein said input device comprises a keyboard.
16. The apparatus of Claim 14, wherein said input device comprises a voice- recognition system.
17. A method for parsing symbol sequences such as natural language sentences comprising the steps of: receiving a symbol sequence; extracting a symbol from the received symbol sequence; searching a lexicon with the extracted symbol and retrieving an entity-type coπesponding to the symbol; grouping the symbols according to the retrieved entity-types; repeating the extracting, searching, retrieving, and grouping steps for each symbol in the received symbol sequence; and outputting said symbol grouping.
18. The method of Claim 17, wherein the retrieving step comprises retrieving a cognitive category valuation for the symbol; and wherein the method includes the step of inserting a relation- symbol between consecutive symbols in the received symbol sequence.
19. The method of Claim 17, wherein said grouping step includes building a symbol group having pictures and a relation.
20. The method of Claim 17, wherein said searching step further comprises retrieving a grouping property coπesponding to the symbol; and wherein said grouping step comprises grouping the symbols according to the retrieved entity-types and grouping properties.
PCT/US1998/017865 1997-09-03 1998-08-28 Method and device for parsing natural language sentences and other sequential symbolic expressions WO1999012107A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU92095/98A AU9209598A (en) 1997-09-03 1998-08-28 Method and device for parsing natural language sentences and other sequential symbolic expressions

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/922,494 US5960384A (en) 1997-09-03 1997-09-03 Method and device for parsing natural language sentences and other sequential symbolic expressions
US08/922,494 1997-09-03

Publications (1)

Publication Number Publication Date
WO1999012107A1 true WO1999012107A1 (en) 1999-03-11

Family

ID=25447117

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1998/017865 WO1999012107A1 (en) 1997-09-03 1998-08-28 Method and device for parsing natural language sentences and other sequential symbolic expressions

Country Status (3)

Country Link
US (1) US5960384A (en)
AU (1) AU9209598A (en)
WO (1) WO1999012107A1 (en)

Families Citing this family (119)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6173261B1 (en) * 1998-09-30 2001-01-09 At&T Corp Grammar fragment acquisition using syntactic and semantic clustering
US7089218B1 (en) 2004-01-06 2006-08-08 Neuric Technologies, Llc Method for inclusion of psychological temperament in an electronic emulation of the human brain
US8725493B2 (en) * 2004-01-06 2014-05-13 Neuric Llc Natural language parsing method to provide conceptual flow
US6556963B1 (en) * 1997-09-24 2003-04-29 International Business Machines Corporation User state sensitive system and method for nutrient analysis using natural language interface
US6185592B1 (en) * 1997-11-18 2001-02-06 Apple Computer, Inc. Summarizing text documents by resolving co-referentiality among actors or objects around which a story unfolds
US6128596A (en) * 1998-04-03 2000-10-03 Motorola, Inc. Method, device and system for generalized bidirectional island-driven chart parsing
US6134709A (en) * 1998-06-29 2000-10-17 Cisco Technology, Inc. Method and apparatus for parsing commands
JP3781561B2 (en) * 1998-08-13 2006-05-31 日本電気株式会社 Natural language analysis device, system and recording medium
US6499013B1 (en) * 1998-09-09 2002-12-24 One Voice Technologies, Inc. Interactive user interface using speech recognition and natural language processing
GB9821969D0 (en) * 1998-10-08 1998-12-02 Canon Kk Apparatus and method for processing natural language
US6442524B1 (en) 1999-01-29 2002-08-27 Sony Corporation Analyzing inflectional morphology in a spoken language translation system
US6243669B1 (en) * 1999-01-29 2001-06-05 Sony Corporation Method and apparatus for providing syntactic analysis and data structure for translation knowledge in example-based language translation
US6278968B1 (en) 1999-01-29 2001-08-21 Sony Corporation Method and apparatus for adaptive speech recognition hypothesis construction and selection in a spoken language translation system
US6282507B1 (en) 1999-01-29 2001-08-28 Sony Corporation Method and apparatus for interactive source language expression recognition and alternative hypothesis presentation and selection
US6266642B1 (en) 1999-01-29 2001-07-24 Sony Corporation Method and portable apparatus for performing spoken language translation
US6356865B1 (en) * 1999-01-29 2002-03-12 Sony Corporation Method and apparatus for performing spoken language translation
US6275791B1 (en) * 1999-02-26 2001-08-14 David N. Weise Natural language parser
US6654731B1 (en) * 1999-03-01 2003-11-25 Oracle Corporation Automated integration of terminological information into a knowledge base
US6374224B1 (en) 1999-03-10 2002-04-16 Sony Corporation Method and apparatus for style control in natural language generation
US7058573B1 (en) * 1999-04-20 2006-06-06 Nuance Communications Inc. Speech recognition system to selectively utilize different speech recognition techniques over multiple speech recognition passes
US20020032564A1 (en) * 2000-04-19 2002-03-14 Farzad Ehsani Phrase-based dialogue modeling with particular application to creating a recognition grammar for a voice-controlled user interface
AU5451800A (en) 1999-05-28 2000-12-18 Sehda, Inc. Phrase-based dialogue modeling with particular application to creating recognition grammars for voice-controlled user interfaces
US6327566B1 (en) * 1999-06-16 2001-12-04 International Business Machines Corporation Method and apparatus for correcting misinterpreted voice commands in a speech recognition system
US6327561B1 (en) * 1999-07-07 2001-12-04 International Business Machines Corp. Customized tokenization of domain specific text via rules corresponding to a speech recognition vocabulary
JP2001075957A (en) * 1999-08-24 2001-03-23 Internatl Business Mach Corp <Ibm> Display method and device for structure of natural language
US6684183B1 (en) 1999-12-06 2004-01-27 Comverse Ltd. Generic natural language service creation environment
US6834280B2 (en) 2000-02-07 2004-12-21 Josiah Lee Auspitz Systems and methods for determining semiotic similarity between queries and database entries
US6556973B1 (en) * 2000-04-19 2003-04-29 Voxi Ab Conversion between data representation formats
US6675159B1 (en) 2000-07-27 2004-01-06 Science Applic Int Corp Concept-based search and retrieval system
US20020046019A1 (en) * 2000-08-18 2002-04-18 Lingomotors, Inc. Method and system for acquiring and maintaining natural language information
US6733389B2 (en) * 2000-10-12 2004-05-11 Igt Gaming device having a first game scheme involving a symbol generator, a second game and a first game terminator
US7027974B1 (en) 2000-10-27 2006-04-11 Science Applications International Corporation Ontology-based parser for natural language processing
US7318019B1 (en) * 2000-11-17 2008-01-08 Semantic Compaction Systems Word output device and matrix keyboard for use therein
US20020091509A1 (en) * 2001-01-02 2002-07-11 Yacov Zoarez Method and system for translating text
US7085723B2 (en) * 2001-01-12 2006-08-01 International Business Machines Corporation System and method for determining utterance context in a multi-context speech application
US7249018B2 (en) * 2001-01-12 2007-07-24 International Business Machines Corporation System and method for relating syntax and semantics for a conversational speech application
US7257537B2 (en) * 2001-01-12 2007-08-14 International Business Machines Corporation Method and apparatus for performing dialog management in a computer conversational interface
US7127402B2 (en) * 2001-01-12 2006-10-24 International Business Machines Corporation Method and apparatus for converting utterance representations into actions in a conversational system
US6950793B2 (en) * 2001-01-12 2005-09-27 International Business Machines Corporation System and method for deriving natural language representation of formal belief structures
US6766316B2 (en) 2001-01-18 2004-07-20 Science Applications International Corporation Method and system of ranking and clustering for document indexing and retrieval
US6961694B2 (en) * 2001-01-22 2005-11-01 Microsoft Corporation Method and apparatus for reducing latency in speech-based applications
IL142421A0 (en) * 2001-04-03 2002-03-10 Linguistic Agents Ltd Linguistic agent system
US7609829B2 (en) * 2001-07-03 2009-10-27 Apptera, Inc. Multi-platform capable inference engine and universal grammar language adapter for intelligent voice application execution
US20030007609A1 (en) * 2001-07-03 2003-01-09 Yuen Michael S. Method and apparatus for development, deployment, and maintenance of a voice software application for distribution to one or more consumers
US20050234727A1 (en) * 2001-07-03 2005-10-20 Leo Chiu Method and apparatus for adapting a voice extensible markup language-enabled voice system for natural speech recognition and system response
US20050216456A1 (en) * 2001-08-10 2005-09-29 T-Systemev, Ltd. Method for entering, recording, distributing and reporting data
US7024351B2 (en) * 2001-08-21 2006-04-04 Microsoft Corporation Method and apparatus for robust efficient parsing
US6785643B2 (en) * 2001-10-15 2004-08-31 Motorola, Inc. Chart parsing using compacted grammar representations
US7949513B2 (en) * 2002-01-22 2011-05-24 Zi Corporation Of Canada, Inc. Language module and method for use with text processing devices
US7805302B2 (en) * 2002-05-20 2010-09-28 Microsoft Corporation Applying a structured language model to information extraction
US20030237055A1 (en) * 2002-06-20 2003-12-25 Thomas Lange Methods and systems for processing text elements
US6881039B2 (en) * 2002-09-23 2005-04-19 Cooligy, Inc. Micro-fabricated electrokinetic pump
US8464781B2 (en) 2002-11-01 2013-06-18 Cooligy Inc. Cooling systems incorporating heat exchangers and thermoelectric layers
US20050211418A1 (en) * 2002-11-01 2005-09-29 Cooligy, Inc. Method and apparatus for efficient vertical fluid delivery for cooling a heat producing device
US7836597B2 (en) 2002-11-01 2010-11-23 Cooligy Inc. Method of fabricating high surface to volume ratio structures and their integration in microheat exchangers for liquid cooling system
US20050096910A1 (en) * 2002-12-06 2005-05-05 Watson Kirk L. Formed document templates and related methods and systems for automated sequential insertion of speech recognition results
US7444285B2 (en) * 2002-12-06 2008-10-28 3M Innovative Properties Company Method and system for sequential insertion of speech recognition results to facilitate deferred transcription services
US7774694B2 (en) 2002-12-06 2010-08-10 3M Innovation Properties Company Method and system for server-based sequential insertion processing of speech recognition results
US20040117173A1 (en) * 2002-12-18 2004-06-17 Ford Daniel Alexander Graphical feedback for semantic interpretation of text and images
US6917936B2 (en) * 2002-12-18 2005-07-12 Xerox Corporation Method and apparatus for measuring similarity between documents
US8849648B1 (en) 2002-12-24 2014-09-30 At&T Intellectual Property Ii, L.P. System and method of extracting clauses for spoken language understanding
US8818793B1 (en) 2002-12-24 2014-08-26 At&T Intellectual Property Ii, L.P. System and method of extracting clauses for spoken language understanding
US7044196B2 (en) * 2003-01-31 2006-05-16 Cooligy,Inc Decoupled spring-loaded mounting apparatus and method of manufacturing thereof
US20040162724A1 (en) * 2003-02-11 2004-08-19 Jeffrey Hill Management of conversations
US7606714B2 (en) * 2003-02-11 2009-10-20 Microsoft Corporation Natural language classification within an automated response system
US7337157B2 (en) * 2003-02-19 2008-02-26 Kurzweil Technologies, Inc. System, method, and product of manufacture for implementing an EAIL (enhanced artificial intelligence language) engine
US7707135B2 (en) * 2003-03-04 2010-04-27 Kurzweil Technologies, Inc. Enhanced artificial intelligence language
US7421393B1 (en) 2004-03-01 2008-09-02 At&T Corp. System for developing a dialog manager using modular spoken-dialog components
US7591302B1 (en) * 2003-07-23 2009-09-22 Cooligy Inc. Pump and fan control concepts in a cooling system
WO2005033909A2 (en) * 2003-10-08 2005-04-14 Any Language Communications Inc. Relationship analysis system and method for semantic disambiguation of natural language
US20050163136A1 (en) * 2003-11-17 2005-07-28 Leo Chiu Multi-tenant self-service VXML portal
US7697673B2 (en) 2003-11-17 2010-04-13 Apptera Inc. System for advertisement selection, placement and delivery within a multiple-tenant voice interaction service system
US7412393B1 (en) * 2004-03-01 2008-08-12 At&T Corp. Method for developing a dialog manager using modular spoken-dialog components
US7613719B2 (en) * 2004-03-18 2009-11-03 Microsoft Corporation Rendering tables with natural language commands
JP4525936B2 (en) * 2004-03-18 2010-08-18 日本電気株式会社 Text mining apparatus, method and program thereof
US8706680B1 (en) * 2004-04-05 2014-04-22 Donald E. Macfarlane Automated report generation system using a structured lexicon of active lexemes and method
US7552116B2 (en) * 2004-08-06 2009-06-23 The Board Of Trustees Of The University Of Illinois Method and system for extracting web query interfaces
US7580837B2 (en) 2004-08-12 2009-08-25 At&T Intellectual Property I, L.P. System and method for targeted tuning module of a speech recognition system
US7242751B2 (en) 2004-12-06 2007-07-10 Sbc Knowledge Ventures, L.P. System and method for speech recognition-enabled automatic call routing
US8473449B2 (en) * 2005-01-06 2013-06-25 Neuric Technologies, Llc Process of dialogue and discussion
US7751551B2 (en) 2005-01-10 2010-07-06 At&T Intellectual Property I, L.P. System and method for speech-enabled call routing
US7869989B1 (en) * 2005-01-28 2011-01-11 Artificial Cognition Inc. Methods and apparatus for understanding machine vocabulary
US7657020B2 (en) 2005-06-03 2010-02-02 At&T Intellectual Property I, Lp Call routing system and method of using the same
US10319252B2 (en) 2005-11-09 2019-06-11 Sdl Inc. Language capability assessment and training apparatus and techniques
US8346555B2 (en) * 2006-08-22 2013-01-01 Nuance Communications, Inc. Automatic grammar tuning using statistical language model generation
US8073681B2 (en) 2006-10-16 2011-12-06 Voicebox Technologies, Inc. System and method for a cooperative conversational voice user interface
US20080154853A1 (en) * 2006-12-22 2008-06-26 International Business Machines Corporation English-language translation of exact interpretations of keyword queries
US7818176B2 (en) 2007-02-06 2010-10-19 Voicebox Technologies, Inc. System and method for selecting and presenting advertisements based on natural language processing of voice-based input
JP4398992B2 (en) * 2007-03-29 2010-01-13 株式会社東芝 Information search apparatus, information search method, and information search program
US8140335B2 (en) 2007-12-11 2012-03-20 Voicebox Technologies, Inc. System and method for providing a natural language voice user interface in an integrated voice navigation services environment
US9305548B2 (en) 2008-05-27 2016-04-05 Voicebox Technologies Corporation System and method for an integrated, multi-modal, multi-device natural language voice services environment
US8370128B2 (en) * 2008-09-30 2013-02-05 Xerox Corporation Semantically-driven extraction of relations between named entities
US8336021B2 (en) * 2008-12-15 2012-12-18 Microsoft Corporation Managing set membership
US8326637B2 (en) 2009-02-20 2012-12-04 Voicebox Technologies, Inc. System and method for processing multi-modal device interactions in a natural language voice services environment
US20110112823A1 (en) * 2009-11-06 2011-05-12 Tatu Ylonen Oy Ltd Ellipsis and movable constituent handling via synthetic token insertion
US10417646B2 (en) 2010-03-09 2019-09-17 Sdl Inc. Predicting the cost associated with translating textual content
US9928296B2 (en) * 2010-12-16 2018-03-27 Microsoft Technology Licensing, Llc Search lexicon expansion
US11003838B2 (en) 2011-04-18 2021-05-11 Sdl Inc. Systems and methods for monitoring post translation editing
WO2013096892A1 (en) * 2011-12-21 2013-06-27 Federated Media Publishing, Llc Method and apparatus for rating documents and authors
CN103365834B (en) * 2012-03-29 2017-08-18 富泰华工业(深圳)有限公司 Language Ambiguity eliminates system and method
US10261994B2 (en) 2012-05-25 2019-04-16 Sdl Inc. Method and system for automatic management of reputation of translators
US9280520B2 (en) * 2012-08-02 2016-03-08 American Express Travel Related Services Company, Inc. Systems and methods for semantic information retrieval
US9152622B2 (en) 2012-11-26 2015-10-06 Language Weaver, Inc. Personalized machine translation via online adaptation
US9213694B2 (en) 2013-10-10 2015-12-15 Language Weaver, Inc. Efficient online domain adaptation
CA2949782C (en) 2014-04-25 2023-09-05 Mayo Foundation For Medical Education And Research Enhancing reading accuracy, efficiency and retention
WO2016044321A1 (en) 2014-09-16 2016-03-24 Min Tang Integration of domain information into state transitions of a finite state transducer for natural language processing
EP3195145A4 (en) 2014-09-16 2018-01-24 VoiceBox Technologies Corporation Voice commerce
US9747896B2 (en) 2014-10-15 2017-08-29 Voicebox Technologies Corporation System and method for providing follow-up responses to prior natural language inputs of a user
US10431214B2 (en) 2014-11-26 2019-10-01 Voicebox Technologies Corporation System and method of determining a domain and/or an action related to a natural language input
US10614799B2 (en) 2014-11-26 2020-04-07 Voicebox Technologies Corporation System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance
US10360501B2 (en) * 2015-12-31 2019-07-23 International Business Machines Corporation Real-time capture and translation of human thoughts and ideas into structured patterns
US10606952B2 (en) 2016-06-24 2020-03-31 Elemental Cognition Llc Architecture and processes for computer learning and understanding
US10331784B2 (en) 2016-07-29 2019-06-25 Voicebox Technologies Corporation System and method of disambiguating natural language processing requests
US10997227B2 (en) 2017-01-18 2021-05-04 Google Llc Systems and methods for processing a natural language query in data tables
US10832658B2 (en) * 2017-11-15 2020-11-10 International Business Machines Corporation Quantized dialog language model for dialog systems
WO2019156131A1 (en) * 2018-02-07 2019-08-15 日本電気株式会社 Information processing device, information processing method, and computer-readable storage medium
CN109062904B (en) * 2018-08-23 2022-05-20 上海互教教育科技有限公司 Logic predicate extraction method and device
US11386902B2 (en) * 2020-04-28 2022-07-12 Bank Of America Corporation System for generation and maintenance of verified data records
US11531821B2 (en) * 2020-08-13 2022-12-20 Salesforce, Inc. Intent resolution for chatbot conversations with negation and coreferences

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4852170A (en) * 1986-12-18 1989-07-25 R & D Associates Real time computer speech recognition system
US4914704A (en) * 1984-10-30 1990-04-03 International Business Machines Corporation Text editor for speech input
US5047953A (en) * 1989-07-31 1991-09-10 Detroit Institute Of Children Augmentive communications system and method
US5083268A (en) * 1986-10-15 1992-01-21 Texas Instruments Incorporated System and method for parsing natural language by unifying lexical features of words
US5210689A (en) * 1990-12-28 1993-05-11 Semantic Compaction Systems System and method for automatically selecting among a plurality of input modes
US5299125A (en) * 1990-08-09 1994-03-29 Semantic Compaction Systems Natural language processing system and method for parsing a plurality of input symbol sequences into syntactically or pragmatically correct word messages
US5309546A (en) * 1984-10-15 1994-05-03 Baker Bruce R System for method for producing synthetic plural word messages
US5345538A (en) * 1992-01-27 1994-09-06 Krishna Narayannan Voice activated control apparatus
US5457768A (en) * 1991-08-13 1995-10-10 Kabushiki Kaisha Toshiba Speech recognition apparatus using syntactic and semantic analysis

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4829423A (en) * 1983-01-28 1989-05-09 Texas Instruments Incorporated Menu-based natural language understanding system
US4887212A (en) * 1986-10-29 1989-12-12 International Business Machines Corporation Parser for natural language text
US4864502A (en) * 1987-10-07 1989-09-05 Houghton Mifflin Company Sentence analyzer
US4868750A (en) * 1987-10-07 1989-09-19 Houghton Mifflin Company Collocational grammar system
US4994966A (en) * 1988-03-31 1991-02-19 Emerson & Stern Associates, Inc. System and method for natural language parsing by initiating processing prior to entry of complete sentences
US4914590A (en) * 1988-05-18 1990-04-03 Emhart Industries, Inc. Natural language understanding system
EP0361570B1 (en) * 1988-09-15 1997-08-06 Océ-Nederland B.V. A system for grammatically processing a sentence composed in natural language
JPH02240769A (en) * 1989-03-14 1990-09-25 Canon Inc Device for preparing natural language sentence
US5146406A (en) * 1989-08-16 1992-09-08 International Business Machines Corporation Computer method for identifying predicate-argument structures in natural language text
US5297040A (en) * 1991-10-23 1994-03-22 Franklin T. Hu Molecular natural language processing system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5309546A (en) * 1984-10-15 1994-05-03 Baker Bruce R System for method for producing synthetic plural word messages
US4914704A (en) * 1984-10-30 1990-04-03 International Business Machines Corporation Text editor for speech input
US5083268A (en) * 1986-10-15 1992-01-21 Texas Instruments Incorporated System and method for parsing natural language by unifying lexical features of words
US4852170A (en) * 1986-12-18 1989-07-25 R & D Associates Real time computer speech recognition system
US5047953A (en) * 1989-07-31 1991-09-10 Detroit Institute Of Children Augmentive communications system and method
US5299125A (en) * 1990-08-09 1994-03-29 Semantic Compaction Systems Natural language processing system and method for parsing a plurality of input symbol sequences into syntactically or pragmatically correct word messages
US5210689A (en) * 1990-12-28 1993-05-11 Semantic Compaction Systems System and method for automatically selecting among a plurality of input modes
US5457768A (en) * 1991-08-13 1995-10-10 Kabushiki Kaisha Toshiba Speech recognition apparatus using syntactic and semantic analysis
US5345538A (en) * 1992-01-27 1994-09-06 Krishna Narayannan Voice activated control apparatus

Also Published As

Publication number Publication date
AU9209598A (en) 1999-03-22
US5960384A (en) 1999-09-28

Similar Documents

Publication Publication Date Title
US5960384A (en) Method and device for parsing natural language sentences and other sequential symbolic expressions
Berwick Principle-based parsing
Grishman Computational linguistics: an introduction
Winograd Understanding natural language
Briscoe Lexical issues in natural language processing
JP2012520527A (en) Question answering system and method based on semantic labeling of user questions and text documents
Habash Generation-heavy hybrid machine translation
Torr Wide-coverage statistical parsing with Minimalist Grammars
Van Halteren et al. Linguistic Exploitation of Syntactic Databases: The Use of the Nijmegen LDB Program
Hayes et al. A tutorial on techniques and applications for natural language processing
Roberts et al. Visualising Collocation for Close Writing.
Kuboň Problems of robust parsing of Czech
Delmonte Deep & shallow linguistically based parsing
Van Delden et al. Combining finite state automata and a greedy learning algorithm to determine the syntactic roles of commas
Kovář Automatic syntactic analysis for real-world applications
Hoard Language understanding and the emerging alignment of linguistics and natural language processing
Du et al. Syntax-directed machine translation of natural language: Effect of garden path phenomenon on sentence structure
Dods Automatically inferring grammar specifications for adnominal possession from interlinear glossed text
Fraj et al. Parsing Arabic Texts Using Real Patterns of Syntactic Trees
Kuhn Two ways of formalizing OT syntax in the LFG framework
Sengupta On lexical and syntactic processing of Bangla language by computer
Morsch et al. Crisis Action Message Analyzer Exploratory Development Model
Bonfante et al. CHAPTER TWELVE SUPERTAGGING WITH CONSTRAINTS
Idicula Design and Development of an Adaptable Frame-Based System for Dravidian Language-Processing
Chesi Complexity and Determinism in Linguistic Computation

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AU CA CN IL JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: CA