US5721938A - Method and device for parsing and analyzing natural language sentences and text - Google Patents
Method and device for parsing and analyzing natural language sentences and text Download PDFInfo
- Publication number
- US5721938A US5721938A US08/487,263 US48726395A US5721938A US 5721938 A US5721938 A US 5721938A US 48726395 A US48726395 A US 48726395A US 5721938 A US5721938 A US 5721938A
- Authority
- US
- United States
- Prior art keywords
- words
- word
- text
- program
- meaning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/268—Morphological analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/211—Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
Definitions
- This invention provides a computer-implemented method and device for parsing and analyzing natural language sentences.
- the program is text-based and divides language into binary elements.
- the method and device are applied to the parsing and analyzing of English language sentences through a computer program.
- any language may be similarly parsed, and analyzed.
- the results of the parsing and analysis give information about the sentence or text in connection with word order and meaning.
- the method and device further provide for determining alternative constructions to a particular sentence, so that the user is made aware of possible alternative choices in controlling the meaning of a sentence or text, or potential ambiguities in the sentence meaning.
- the user becomes aware of subtle nuances in meaning or multiple perspectives and then may choose among the alternatives in order to convey the meaning intended or meaning otherwise not anticipated.
- prescriptive grammar Another aspect of school grammar, which we refer to here as prescriptive grammar is its insistence t-hat old, even now archaic, forms must continue in use and that many new formations must be excluded . . . . It prescriptive grammar! fails to recognize that language is an innate attribute of human nature Prescriptive grammar is thus not very much concerned with the nature of language as such, nor with the nature of English in particular. It is interested instead in "correct English,” that is, in enforcing the use of one particular dialect. The interest here in surpassing prescriptive grammar! is in the vast body of structural and syntactic principles which are common to all varieties of English rather than in the minor details which differentiate them. English transformational Grammar, Jacobs and Rosenbaum, Blaisdell, 1968!
- Prescriptive a term of art, assumes a right or a wrong grammatical use in sentences. The usefulness of a prescriptive system is predicated upon the mistaken assumption that the goal in writing is correct structure rather than an understanding of the structure of meaning. Further, a problem grows from the number of possible combinations of meaning grows exponentially with the numbers of words in a sentence. Thus, in systems such as Hu's, a means to limit the number of words in a sentence subscribes to the presumption that a simple sentence is a better sentence. b a prescriptive system of "right” and "wrong", the corpus of rules assumes that English words have one meaning per word. Because English operates very differently, having multiple meanings for words, the prescriptive grammars create a problem.
- the prescriptive grammars need to control by limiting what happens when a word with one of its particular meanings in a given situation binds another word to a particular meaning. Thereby, such a limit actually tries to prevent the occurrence of multiple meanings. The mistake occurs because limiting a word to how it binds or obligates adjacent words limits the number of possible combinations. Further, controlling language meaning, by limiting its adjacent combinations, obligates English to function in a manner opposed to its actual nature. While this may certainly be argued to be true in the Hu system it unnecessarily limits the variety of words that a parser may handle. Also, it restricts the invention to limited language domains where only one meaning of a word would serve the user. For example, the system may be limited to airline reservation systems.
- the present inventive program overcomes the deficiencies of the above prior art.
- the present program provides a natural language text parser, which may be used for all natural languages, and for the syntactic analysis of text which is simpler, faster, and more compact in memory transfer than those of prior art.
- the present program syntactically categorizes and depicts all possible constituents, i.e., all possible word and word pattern combinations, which comprise natural language. It also depicts, in algorithms and on the screen, the order in which all possible constituents, i.e. all possible word and word pattern combinations, can combine.
- the present program provides a method and device to provide linguistic and morphological analysis on a text corpus to derive part-of-speech, and ultimately semantic information, even if a match is not found in a look-up dictionary for a particular word in the text.
- the present program also provides a language analyzer, parser, and editor which use a finite set of very simple grammatical and syntactical rules, rather than a large, set of complicated grammatical rules, which merely apply a better approximation. Instead, simple rules in the present program encompass and include the individual variation of language use. Thus the program allows the analysis of any text in natural language.
- the present program provides a language analyzer which is context-based, rather than object-based. In such a system, it is word order that determines meaning in context, rather than having the lexical definition alone determine the word order. In the same way, the present program treats the sentence as a whole initially and then classifies the individual words and word patterns, as opposed to focusing solely on individual words and how they bind to other words.
- the program breaks the paradigm of chaining words together from left to right to create meaning, and thereby allows the reader to overlap and connect versus chaining meaning. Therefore, the user chooses the word-pattern framework.
- the present program gives the user the tools necessary to recognize word groups identified in traditional grammatical terminology within a sentence, as well as named as the invention's constituent word groups.
- the present program enhances the user's ability to perceive the four basic constituents of language as binary elements of what are termed herein and defined below as Primary Image, Conditional Image, Process Detail, and Background Detail placed in binary positions, where binary refers to the two-part Nounness positions of Doer or Receiver, and the second position as Qualifier.
- Qualifier refers to the vernacular term "modifier", including single-word adjectives and adverbs, but herein also encompasses constituent-element word groups placed as modifiers.
- Verbness as used herein by the program, encompasses two completely different uses of the vernacular term for verb as defined as a word expressing action or state of being. Whether such a word operates by formula with spelling changes, or in combination with auxiliaries, to form verbs in tense, or whether such a word follows the constituent element patterns used herein as Process Details and called in conventional grammatical terms the verbals (the participle, past or present, gerund, or infinitive), the term Verbness includes interchangeably the power of the verb to function as a constituent element of Image or separately, by formula, as a Detail constituent element.
- Equalizers include the terms and lists in conventional grammar, known as conjunctive adverbs and coordinating conjunctions, and encompassing all words or phrases used to introduce or join Primary Image with an Equalized relationship. These words Trigger relationship which signal equally weighted idea relationships. Please see Appendix 1 for a list of these.
- Conditional Triggers as used herein by the program include the terms and lists in conventional grammar, known as subordinating conjunctions and relative pronouns, and encompassing all words or phrases used to introduce, create, and join Conditional Image to sentences. These words Trigger relationships of ideas which either support or limit by circumstance the meanings to which they are attached as Nounness or Qualifier. Again, please see Appendix 2 for a list of these.
- Process Detail Triggers as used herein by the program function by formula as added to the structure of verbs preventing them from use as verb in tense.
- the Process Details add meaning of past, on-going, or future processes to other parts of the sentence to which they are attached as Nounness or Qualifier. Again, please see Appendix 3 for a list of these.
- the Background Detail Triggers used herein by the program include the terms and lists in conventional grammar, known as prepositions, and encompassing all words and phrases used to introduce, create, and join Background Detail to a sentence. These words Trigger and add relationships of time, place, and grouping as meanings to other parts of the sentence to which they are attached as Nounness or Qualifier. Again, please see Appendix 4 for a list of these.
- the present program categorizes parts of an inputted text into one of four patterns symbolized by colors on the computer screen.
- the present program classifies syntax in natural languages such as English.
- the invention may be advantageously used, for example, in the context of a syntax checker or natural-language processor.
- the invention comprises software, which dissects text, classifying its components into two units denoted herein as: Nounness and Verbness. These units are optionally combined in the formation of different types of word patterns.
- a provision for positioning the word patterns of the text in different locations allows the user to vary the text's meaning.
- Such analysis provides a way to communicate the textual meaning of a document in terms understandable by a digital computer.
- the two basic binary elements of natural language combine in two Image word patterns, denoted as Primary Image and Conditional Image, and two Detail word patterns, denoted as Process Detail and Background Detail.
- the two basic elements, Nounness and Verbness function binarily, either in combinations for the two Image word patterns or separately in the two Detail word patterns. All word patterns, outside the verb-in-tense, function binarily in one of two positions: as Nounness, called Nesting, or as modifiers, called Qualifying. Since meaning in a natural language sentence is determined by word- or word-pattern location, the meaning can be changed by moving word patterns from one location to another, called "Flipping".
- Natural language thus organized into binary units constitutes the parsing method and device of the present invention as expressed a program for purposes such as, but not limited to, parsing sentences, analysis of meaning, natural language processing, context analysis of documents, data-base searching, syntax analysis of documents, and the teaching of writing in the English language.
- Embodied in the present invention lies a variety of applications.
- the present invention establishes a base line for emulating textual thought which moves the use of text past the current field of simply text processing.
- Text becomes an extension of thought represented on the screen as bigger patterns of meaning in context combinations rather than solely a string of words from left to right. Therefore, an extended variety of applications includes any representation of text as inputted by voice, for example, or as text intended to be communicated in written form which the user rearranges and re-orders for varied and multiple options of connected and developed thought.
- voice processing technology Once combined with voice processing technology, the user will be able to speak text while seeing its context framework arranged and highlighted in color on a computer screen.
- the present invention addresses the occurrence of homonyms because fixed word order clearly determines by context that the word "to” could never be used interchangeably with the word “too”, for example. Even a spoken homonym must have fixed word order placement which the present invention correctly identifies.
- Further applications include natural language processing and programming, machine translation, data-base searching, "text engine” document search, artificial intelligence and other language-related applications. Currently, the industry is searching for natural language solutions for these applications, but the object-based solutions cannot solve problems created when vocabulary words either sound alike or have more than one meaning.
- the present invention allows an architecture for symbolizing constituent elements and components of thought reducing the flow of data on optic fibers, for example, to very small pieces of information.
- Data flow with language can become compressed in a binary fashion with the present invention adapted to telecommunication.
- machine translation for foreign languages frustrates users because the literal word-for-word translations cannot account for the differences between inflected languages and a language which depends on word order for meaning.
- the present invention provides a means of identifying syntax in English. Therefore, the present program can serve as the foundation for comparing English syntax to that of other languages for the purposes of teaching syntax from one language to another or for machine translation. Further, the binary nature of the present invention offers a solution for creating a natural-language programming code which can dependably operate without debilitating exceptions. An example provides herein a sample of how symbols could represent larger pieces of meaning for symbolic encryption.
- the goal of the present invention is to create a computer program to identify the algorithmic architecture of language which identifies bigger blocks of meaning than single words.
- the computer program by means of this invention, operates from blocks, or patterns, of meaning which, in turn, operate by symbolic formulae.
- the present invention analyzes what the user originates through an inputted text. In such a fashion, the present invention changes the technology from an object-based premise to a context-based foundation for examining the chain of events which the user has connected as a particular thought.
- a context-based architecture creates a technology where the user can predict new directions of thought and meaning beyond the object-based definitions of words in isolation which does not provide any means for the user to control changes in options of thought.
- the present invention therefore, accounts automatically for the differences in meaning inherent in single words as they connect to produce varied meaning by changes in context.
- the present invention identifies the differences in multiple meanings for identical forms. Without the context placement, the user cannot know which meaning of "wind” or “clock” applies. In the case of “wind,” the pronunciation changes as the context changes; in the case of "clock,” however, the meaning change does not coincide with a change in pronunciation to reflect the change in meaning.
- the current program creates the context meaning without the necessity of pronounced verbal clues.
- the current program uses the written text and context structure to indicate the differences in word definitions from fixed-word-order placement.
- the present invention organizes bigger blocks of meaning by defining the syntactical architecture which has given these famous, nonsensical lines from Lewis Carroll's "Jabberwocky" context-based meaning. Therefore, the individual words do not have lexical meaning but, on the contrary, are not nonsensical lines because they have context meaning.
- context-based architecture of the present invention creates a syntactical paradigm which supersedes literal, object-based programming.
- FIG. 1 shows a diagram of the structures into which the present program categorizes natural language.
- FIG. 2 shows a diagram of the various binary constituents of FIG. 1 and their flow to binary placements.
- FIG. 3 is a table of binary constituents which describe the constituent elements and comparatively align their elements.
- FIG. 4 is a flip chart illustrating how the computer determines options to move or Flip data from one pattern to another and then how to place these patterns once structured.
- FIG. 5 is a diagram showing a type of data compression.
- the present invention is a computer-implemented parser for natural languages which supersedes traditional, generative, and object-based analysis by creating a binary architecture which defines context and parts of speech by word order in any use of language text.
- it may be implemented on a computer, for example, within a software program or coordinated with a voice-synthesizing program or product.
- a requisite amount of memory is available, it may also be implemented directly in a hard-wired system, such as in ROM chips.
- ROM chips read-only memory
- An embodiment of the present invention was first written in the language C and was enhanced in C++. A system of 4 MB would be sufficient to run the program. Please refer to the attached example 2 for a detailed description of the hardware specifications.
- a first step in the program is the inputting of a stream of alphanumeric data delimited by a punctuation mark. More than likely, in the typical text to be analyzed, there will be a number of punctuation marks indicating a number of different sentences.
- This text may come from a number of sources. For example, it may be imported from an ASCII text file, or it may come from a text file having a predetermined format, such as that from a standard word processor program. Alternatively, the user may create an original document within the program itself. Additionally, it may be imported from a magnetic tape, or from a data file stored in a semiconductor memory. Of course, numerous other methods may be used to store the data file as is well-known in the art.
- the menu and tool bar options are described in detail in example 2.
- the screen colors serve as an integral aspect of the actual presentation and working architecture of the logic which presents the results of the analysis to the user. In no way can the colors on the screen be considered as merely secondary or decorative graphics, but instead they serve as an integral element of the program's operating system as the source of context meaning for the user. The colors and their significance are described more fully later.
- the program of the present invention can be run by a typical CPU with a standard keyboard used to effect choices of menu or tool bar commands.
- the program operates on the premise that the verb is the "axis." As such, the location of the possible verbs must be located first. This is done in conjunction with a dictionary look-up system in which the dictionary comprises words categorized as one or more parts of speech. This dictionary contains no lexical definitions. Accompanying the dictionary look-up system is a word-order analyzer which separates the possible verbs from other words by word order and morphology, to be explained later with reference to a specific embodiment. In this way, the definition and isolation of the verb is performed by the system.
- the Help system is not a tutorial in the embodiment described of the program; it serves as an actual interface between the logic of the system and the user's needs.
- the user makes changes in meaning from among the options offered by the Help system while, within the working of the internal programs, the order of the searching logic describes and re-describes the constituent elements of the user's text.
- Example 2 contains a detailed description of the mechanism and order of the searching logic.
- a similar search and word-order analysis is performed in this respect in separate searching steps regarding the remaining parts of speech, including, and preferably in this order: noun, pronoun, adjective, adverb, conjunction, preposition, interjection.
- the program moves on to analyze the constituent elements, i.e., the word patterns, of the text.
- These constituent elements are of four basic types: Primary Image, Conditional Image, Process Detail, and Background Detail.
- a word search is also conducted to determine the presence and location of certain words which function as Triggers for Conditional Images, Background Details and Process Details, and between Primary Images.
- Conditional Triggers are words which signal Conditional word group patterns for conditional relationships among Images.
- Conditional Triggers include subordinating conjunctions and relative pronouns and any combinations of words which serve as Conditional Triggers and are listed as such.
- Background Triggers are words and combinations of words which signal Background word group patterns for background data added to other constituent-element word groups. Background Triggers include prepositions and are listed as such.
- Equalizing Trigger Another type of Trigger for which the program searches is the Equalizing Trigger which signals Equalized and paired relationships among words and groups of words. These include conjunctive adverbs and coordinating conjunctions and are listed as such in Appendix 1.
- word patterns i.e., constituent elements, determined by the program.
- the four basic types of word patterns, i.e., constituent elements, determined by the program, are the Primary Image, the Conditional Image, the Process Detail, and the Background Detail. These word patterns, or constituent elements, are shown in FIG. 1.
- the failed system had backups which saved the data.
- the present program identifies a Primary Image whenever it finds a function of Nounness followed by a verb, where Nounness is defined as any Conditional Image, Process Detail, Background Detail, Primary Image in quotation marks, single-word nouns or multiple single-word nouns, personal pronouns, indefinite pronouns, or demonstrative pronouns. Any part of speech or word construct which functions in one of the binary Nounness locations, by formula, serves as Nounness as defined within the computer program.
- Nounness is denoted an agent/doer, and classifies as the subject of a clause, i.e., an Image.
- the present program will find a Primary Image rather than a Conditional one when the agent/doer, followed by a verb, is not preceded by a Conditional Trigger. If it is preceded by a Conditional Trigger, a Conditional Image will be identified by the program.
- Nounness has two placements in the program: doer or receiver. All of the examples above demonstrate the doer position but can be "Flipped" by the program to the other, receiver position.
- the receiver position follows a verb in tense in either a Conditional or Primary Image. Moving the above underlined examples to the other side of the verb would change the Nounness to a receiver position and therefore change the meaning of the sentence significantly.
- the doer position of Nounness occurs only in the two Image constituents. Both Image constituent elements contain the receiver position and both Detail constituent elements can contain receivers within the method of the invention.
- the Background Detail contains the receiver as an integral part of its formula but the Process Detail does not have a receiver when the Process Detail has an intransitive property.
- nominative case pronouns function within the method of this invention solely as the doer or following a linking verb in tense, just as the objective case pronoun, as it is known, serves solely in a receiver position.
- nominative case pronouns exist for that purpose, variations may occur because individual use in a dialect structure does not change the context and because fixed-word order has more meaning than do the words chosen to be put in those positions.
- a typical document may say a great deal more in context than vocabulary alone.
- the present invention allows a simple, effective, and powerful contextual insight not otherwise accessible. Examples in this description will illustrate the ease of insight through the color-coded text analysis. However, in the present application, such colors will be denoted by boldface, italics, underlining, small text, and normal text.
- Nounness which also showcases the program's nesting function and the inter-relationship of writing processes, follows. Wording from the first sentence will illustrate. The passage, “for understanding and forecasting the structure, variability and dynamic interaction,” combines the Process Detail with a Background Detail. The word “for” Triggers the Background Detail context which the present invention would color orange. The phrase, “understanding and forecasting the ocean's structure, variability and dynamic interactions,” uses the Process Detail formula to create the context of a process (done by the agency in the document, NOAA) and places it as the receiver of the Background Detail. The present invention colors the Process Detail green, superimposing it on an orange background to illustrate the Nesting, or Nounness, function. The full text of the example follows and a full analysis is given in Appendix 5.
- Conditional Images must be joined to other Images and to each other in a similar fashion, these Conditional Images being initiated and joined by Conditional Triggers, including subordinating conjunctions and relative pronouns.
- the term "join” may refer to adjacent constituent elements of Image in the case of Qualifier; or the term “join” may refer to nonadjacent, that is, stacked, or Nested, Images used as Nounness in which the Conditional Trigger introduces a Conditional Image which itself operates as part of another constituent element.
- Conditional Triggers "where" and “while” specify particular conditions applying to the identified ocean regions, preventing generalization.
- the Image creates the basic unit of an idea in the present program.
- the movement, or verb, within the Image creates the core of that idea.
- the word ⁇ movement ⁇ refers to any verb in tense which expresses mental, physical, or emotional "doing" initiated by a "doer", that is, a form of Nounness fulfilled by a word or constituent element. Without the movement, the program recognizes no Image. In fact, this movement unit serves as the "axis" of the Image. Even an Image with only the movement stated has a doer understood by the computer, such as in commands.
- the verb is the axis of the Image, and every sentence in the program must have at least one Image.
- verbs as action words. The modern user, therefore, can feel lost or confused if verbs do not suggest, visually, physical action. From a lexicon of 155,000 words, the text corpus of a collegiate dictionary, for example, contains approximately 9400 verbs. A significant proportion of these verbs do not in any way express physical, visible movement. Verbs may express movement but without a physical embodiment. Even thoughts and emotions constitute a type of movement. Similarly, existence constitutes a "movement" in time and space. Therefore, the present program identifies "Verbness,” whether as a form of tense in Image or as a Process Detail, to be movement. Examples of such verb movement can be physical, mental, and emotional.
- English has a lexicon of 17 so-called linking verbs. They serve to show equality between the doer and the receiver or a receiver and some prior referent. This lexicon contains a few verbs which distinctly mean "existence" or some shade of existence.
- axis refers to the position of verb as it occurs in tense in Images within the program because the sentence, or multiple clauses within sentences, cannot exist in English without the verb placed in near proximity to its Doer.
- a verb which serves as a command can serve in one word within the program as a full sentence because its Doer is a silent and universally understood "you".
- the verb in tense in this example is the word "is". Contrary to conventional descriptions of English and the current usage of those in the field of linguistics, the verb "is” does not allow for completed meaning in a simple sentence form by any stretch of the imagination. Within the inventive program "is” serves as the axis between the Doer and the Receiver and, thereby, the present invention would color "is” blue to indicate that the Primary Image axis of this sentence example serves to support the relationship between the doer and receiver and cannot, thereby, serve as meaning by itself. In linguistic terms, the verb in tense initiates the verb phrase. However, in the present invention, the verb in tense does not serve as a Trigger for forthcoming meaning. Rather, the verb in tense is an axis around which the meaning revolves for each Image.
- the Image is a combination of Nounness and Verbness.
- Process Details such as infinitives, participles, and gerunds, may function as Nounness.
- the other set of word patterns is the category of Details.
- the program categorizes Details in two different sub-categories: Process Details and Background Details. In contrast to Images, Details are not combinations of Nounness and Verbness.
- the program recognizes two binary elements in English, Nounness and Verbness. Together they form an Image and separately, they form the two Detail patterns.
- a Background Trigger which is conventionally a preposition, typically comprises a Background Trigger preceding an Agent/Receiver. That is, a Background Trigger precedes Nounness. Please see Appendix 4 for a list of those Triggers.
- a Process Detail within the meaning of the program has more of the qualities of Verbness.
- the Process Detail looks like a verb, the Verbness here does not create an event or idea, as does an Image.
- the present program notes that the Process Detail has the qualities of a verb including one of the transitive, intransitive, or linking properties as well as the expression of movement; but the Process Detail never has tense.
- the combination of the word "to" and a verb, with an optional agent/receiver may create a Process Detail.
- verbs ending in "-ing”, with an optional agent/receiver, verbs ending in -"ed", as well as verbs ending in certain specific, irregular endings typically create Process Details and would be colored appropriately in the present program. Process Details have no agent/doer. They may have agent/receivers, however.
- Auxiliary verbs are also considered. These verbs operate in conjunction with verbs in tense without changing the tense meanings. They add possibility, or conjecture, but do not, in this program, in any way alter the patterning of tense use. Please see Appendix 8 for a list of these.
- FIG. 4 shows the allowable Flipping options.
- the reader's understanding drives the meaning as determined by the computer.
- the user of the present program can either write what seems to fulfill the intended meaning, then vary the placement to choose optional meaning, or, the user can read documents from other people to determine by context either the intention of the text or how to respond in like manner, contextually.
- the following sentences illustrate from among the various allowable placements, in fixed word order context, available for the arrangements of constituent elements in thought patterns.
- the Flipping function allows the user to re-format data and thereby to access different contextual relationships and allowing the user multiple varied options for connecting thought.
- the data in the constituent element may, as in this case, stay in exactly the same form from one sentence placement to another. However, the meaning that the information flow creates from the way the constituent element connects to another constituent element by changing placement completely alters the knowledge imparted.
- the present program uses the fact that word pattern location determines meaning to present multiple perspectives to the user while the user analyzes a particular sentence or text. These alternatives are combinations of Images and Details or the data in them in different orders. Obviously, many different meanings can be conveyed by many different constituent element orders. In fact, once a user changes the pattern, the meaning is usually changed also. The user can choose among sentencing options and thereby weigh emphasis and focus, and thereby locate potential ambiguities and create the meaning intended. In this way, the user can choose the least ambiguous combination of words or uncover a completely different interpretation for the same data.
- Images and Details in and of themselves may be used to show the user the underlying, contextual meanings being communicated in the text.
- the present system advantageously colors the Images and Details on the computer screen in such a way as to communicate what kind of Image or Detail is present in any particular text.
- the present method of the program uses a consistent coloring scheme to communicate the borders of the constituent element word patterns. These colors consistently communicate the type of context which the constituent elements symbolize. This specific color scheme is explained later in this paragraph. Just as the program categorizes English wording into one of these four patterns symbolized by colors, nothing that appears on the screen remains uncolored.
- the word “needs” can operate either as a verb in tense which would combine with the prior wording "and the expected increase . . . “ to create a blue Primary image; or, the same wording "and the expected increase . . . “ ties to a prior "both . . . “ and therefore "needs” either requires “of” to be written before it or the writer neglected to complete the Image.
- these coloring schemes may overlap.
- Primary Images are colored blue; Conditional Images are colored pink; Background Details are colored orange; and Process Details are colored green. Thus, when looking at the coloring on the screen, the user is able to see which word patterns are occurring and in what orders.
- a blue Primary Image provides the skeleton of a sentence.
- a number of blue-colored patterns show equality of strength among the different Primary Images.
- Primary Images may serve as central or supporting points, however, but not necessarily as the only point.
- Primary Images occasionally but only rarely Nest or Qualify, whereas they usually serve as a support structure for Nesting and compressing meaning.
- a dominance of Primary Images may signal being too happy or literal. In this case, for example, the user may not be showing enough shades of meaning.
- the present invention would color this group of words blue.
- Conditional Images communicate causes or time constraints. They show contingency. They expand or limit events. They create idea relationships, and dimensions of time and place; they provide supporting information of quantity, quality, rank, degree, cause, contrast, exclusion.
- a phrase colored green indicates a Process Detail which compacts or compresses ideas, or layers meaning.
- the user may thus decide to more explicitly identify agents or causes or sources of action.
- the use of too many Process Details, shown by an abundance of green on the screen, may indicate to the user that they are overemphasizing how something is done rather than who or what caused the events.
- the user may choose to compact multiple sentences into more economical wording with Process Details.
- the present invention would color "the system's failing" green.
- a phrase colored orange indicates Background Detail which helps the user make the text more specific.
- a drawback of the extended use of Background Detail is that the user may obscure the overall picture.
- the audience or substance of the text may demand details of time, place, and grouping.
- the present program would color "in the budget department” orange.
- the present program parses natural languages.
- the identification and location of various word patterns allow the user to arrange them at will according to the meaning actually intended. This contrasts with previous systems which are organized around and rely upon individual word meanings so as to make the problem of natural language analysis lexical and not context-based.
- the program is context-based. That is, the program analyzes language at a higher structural level than prior object-based systems, and is consequently more effective.
- the present invention maximizes as opposed to limits meaning choices.
- the present invention also correctly, and perhaps more accurately than other systems, parses traditional parts of speech.
- An example of how accurately context-, rather than object-based analysis identifies multiple word uses is illustrated in the following sentence:
- the present invention utilizes the screen to demonstrate, by color, the context patterns and interrelationships among them in natural text.
- a handbook accompanies the program to help the user choose options of thought in the wording of that text.
- both the handbook and the help system include reminder descriptions of how each pattern operates. Users may use, e.g., a set of highlighter markers with the same abbreviated descriptions for reminders and reinforcement. These descriptions also have a use symbolically in conjunction with developing technology for voice activation, machine translation, natural language programming and artificial intelligence. Because the tool bar and the message lines in both the DOS and Windows versions utilized these abbreviations to direct the user, this example elucidates and repeats the same abbreviated definitions.
- Example may use a convenient data compression technique as indicated in FIG. 5 which allows the easy look-up of words.
- FIG. 5 An example follows below for identifying any Conditional pattern.
- the Example incidentally, may use a convenient data compression technique as indicated in FIG. 5 which allows the easy look-up of words.
- the transmission for example, of voice synthesized communication via optic fiber requires compressing data into units or blocks of meaning which are as small or compact as possible.
- the constituent word patterns of the present program serve as such units.
- all Conditional Triggers for example, could by symbolized with one type of computer code having symbolic variations to present all of the various Conditional Triggers.
- the program runs on an IBM compatible system but may also run on many other platforms, e.g., Apple Macintosh and UNIX. 4 MB of RAM, or more, are recommended.
- the language is C; however, some of the code has been written in C++.
- the data files consist of a dictionary, approximately 750K, a translation file, 1K, and a configuration file.
- the dictionary contains no definitions, only syntactic information listing all the parts of speech a particular word can be.
- the translation file decodes the encrypted dictionary.
- the software in this example can highlight a 1000 sentence text corpus; however, as fully implemented, the software can handle any size document.
- the software opens up a configuration file which tells certain hardware whether, for example, to spell check or not, to use black and white or color, and to write to the screen using fast video memory. Depending upon what the configuration file tells it, it sets color attributes that will be used. Then the user is prompted for a file name. The file is located, which fills up the display memory with that file; or the user may create a new file; or the user may input a file by any other means.
- Alt+S will begin sentence analysis.
- the software parses one character at a time from the beginning until it locates the ending punctuation for sentences, including ⁇ ?,
- the user may press W to identify each word's syntactic function. It parses one sentence at a time from the first character to the end and locates the end of word punctuation including ⁇ space, comma, semi-colon, quotation marks, or/and period). Then it fills another internal structure called word structure which contains the beginning and end of each word relative to display memory and contains information about the word type from the dictionary and the word's use, e.g., whether a noun is a subject noun or a receiver. As it locates each word, it checks the dictionary to find the first occurrence of the word in the dictionary, then stops when * finds a hit and marks the word with a syntactic part of speech label. Then it goes to the next word.
- the user may opt to highlight Images in their appropriate colors.
- the user chooses "I”. It parses the word structure from the first word to the end of the sentence, looking for Conditional Triggers and Equalizing and Joining Triggers.
- the user may also opt to highlight Details.
- the Detail logic is nearly the same as the clause logic in that the program searches for Detail Triggers as it scans the words from beginning to end, filling up the phrase structure relative to the display memory. Likewise, it identifies Detail type and highlights each Detail on the screen in the appropriate colors.
- a user may opt to examine many sentences simultaneously to look for structure patterns.
- the user may select G for group mode and drag the cursor to the desired amount of text.
- the highlighting occurs identically to the sentence by sentence process except that it runs the logic repeatedly and displays colors for a body of text rather than just one sentence.
- Sentence variation for re-organizing thought serves as a central point for the user of the present program.
- Context patterns change their "weight" of substance by the simple additional change of Triggers and patterns. Without the color on the screen to illustrate these differences, the user may so struggle with the concepts that the present invention would be less than easily used. The color, therefore, provides real and substantive meaning beyond the labeling and parsing in the message lines.
- conventional grammatical labels focus primarily upon single-word use and part-of-speech designation with an incidental nod to group patterns which, although named, seem to have minor significance in the process of communicating. Contrarily, the group patterns, in fact, have more significance that the labels suggest because their interrelationships actually create an ordered meaning which the labels do not well explain.
- Boldface type represents the "blue-colored" Primary Image.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
Abstract
A program is provided for organizing a natural language, such as English, into binary units of two basic elements, Nounness and Verbness, which combine in two idea word patterns, called Primary Image and Conditional Image, and two Detail word patterns, called Process Detail and Background Detail. These two basic elements, Nounness and Verbness, function binarily within the program, either in combination for the two Image word patterns or separately for the two Detail word patterns. All word terns, except the verb-in-tense in the two Image word patterns, function binarily within the program in one of two positions: as Nounness, called Nesting, or as modifiers, called Qualifying. Since meaning in an English sentence is determined solely by word and word pattern location, binary units can be created which allow meaning to be changed by moving words or word patterns from one location to another, called Flipping. Natural language, thus organized into binary units, can be thus analyzed in computer programs for purposes such as, but not limited to, natural language processing which is not restricted to limited language domains, voice activation, machine translation from one natural language to another, context analysis of documents, data base searching, syntax analysis of documents, and the teaching of writing in natural language.
Description
This invention provides a computer-implemented method and device for parsing and analyzing natural language sentences. The program is text-based and divides language into binary elements. In the embodiment described herein, the method and device are applied to the parsing and analyzing of English language sentences through a computer program. However, any language may be similarly parsed, and analyzed. The results of the parsing and analysis give information about the sentence or text in connection with word order and meaning. The method and device further provide for determining alternative constructions to a particular sentence, so that the user is made aware of possible alternative choices in controlling the meaning of a sentence or text, or potential ambiguities in the sentence meaning. By the method and device of the invention, the user becomes aware of subtle nuances in meaning or multiple perspectives and then may choose among the alternatives in order to convey the meaning intended or meaning otherwise not anticipated.
Numerous attempts have been made in the art to classify language into a particular scheme for using a computer system. These attempts have particularly focused on determining meaning, on a word-by-word basis, according to a look-up scheme in a dictionary. While varying degrees of success have been obtained, these systems have been driven by rules of traditional grammar which ignore the flexibility of language, thus making it rigid and difficult to use, not reflecting the natural thought process.
For example, U.S. Pat. No. 4,864,502, issued to Kucera, et al., describes a sentence analyzer which identifies noun phrases and verb groups. A separate analyzer then determines clause types. Similarly, U.S. Pat. No. 4,887,212, issued to Zamora, et al., describes a parser which isolates word groups and provides verb group analysis and clause analysis. Neither of these patents describes a software application that classifies text into binary units and which further classifies combinations of these units into binary word patterns, and binary placements for them. Nor do they present alternate choices of meaning to be presented to the user.
U.S. Pat. No. 5,146,405, issued to Church, describes a method for determining parts of speech based on statistical data of actual usage. The deficiency in such a system is that it is tied to the statistical data relating to the actual usage of individual words. This is because it exclusively relies on such arbitrary data to determine part of speech.
U.S. Pat. No. 5,297,040, issued to Hu, describes a method of processing language, where correct sentence structures, the prescriptive structures, also referred to by Roderick A. Jacobs and Peter S. Rosenbaum, as "school grammar," are those whose words bind to their neighboring words in a particular manner rather than in a descriptive structure.
Another aspect of school grammar, which we refer to here as prescriptive grammar is its insistence t-hat old, even now archaic, forms must continue in use and that many new formations must be excluded . . . . It prescriptive grammar! fails to recognize that language is an innate attribute of human nature Prescriptive grammar is thus not very much concerned with the nature of language as such, nor with the nature of English in particular. It is interested instead in "correct English," that is, in enforcing the use of one particular dialect. The interest here in surpassing prescriptive grammar! is in the vast body of structural and syntactic principles which are common to all varieties of English rather than in the minor details which differentiate them. English transformational Grammar, Jacobs and Rosenbaum, Blaisdell, 1968!
There is a need for an invention which is responsive to Jacobs and Rosenbaum which embraces individual combinations of contextual blocks of meaning whether or not they violate the rules of prescriptive grammar.
"Prescriptive", a term of art, assumes a right or a wrong grammatical use in sentences. The usefulness of a prescriptive system is predicated upon the mistaken assumption that the goal in writing is correct structure rather than an understanding of the structure of meaning. Further, a problem grows from the number of possible combinations of meaning grows exponentially with the numbers of words in a sentence. Thus, in systems such as Hu's, a means to limit the number of words in a sentence subscribes to the presumption that a simple sentence is a better sentence. b a prescriptive system of "right" and "wrong", the corpus of rules assumes that English words have one meaning per word. Because English operates very differently, having multiple meanings for words, the prescriptive grammars create a problem. The prescriptive grammars need to control by limiting what happens when a word with one of its particular meanings in a given situation binds another word to a particular meaning. Thereby, such a limit actually tries to prevent the occurrence of multiple meanings. The mistake occurs because limiting a word to how it binds or obligates adjacent words limits the number of possible combinations. Further, controlling language meaning, by limiting its adjacent combinations, obligates English to function in a manner opposed to its actual nature. While this may certainly be argued to be true in the Hu system it unnecessarily limits the variety of words that a parser may handle. Also, it restricts the invention to limited language domains where only one meaning of a word would serve the user. For example, the system may be limited to airline reservation systems.
U.S. Pat. No. 5,146,406, issued to Jensen, describes a method for determining predicate-argument structures by a double parsing system. The first parse analyzes syntax; the second analyzes argument structure. Like other systems, Jensen also utilizes a "best guess" technique that is not well-suited for handling diverse language systems because often errors result.
U.S. Pat. No. 5,101,349, issued to Tokuume, et. al., describes a method for analyzing language, including grammatical functions and condition functions. Also according to Tokuume, et. al., phrase structures may be generated by the grammatical definition using stored grammatical rules. However, such a system is tied to complicated tree and node analysis that necessarily limits its flexibility because of the vast number of exceptions that such a system cannot account for.
For example, The spaceship photographed Seattle flying to Mars.
THE NEW YORK TIMES, Sep. 1, 1993, in reviewing Microsoft's current research, uses this example to illustrate the object-based research pursued by Microsoft's team of linguists and computer scientists. Microsoft's syntactic analysis uses semantic clues, from dictionary definitions, to conclude whether or not Seattle can fly to Mars and thereby forcing a movement of the sentencing to the following:
Flying to Mars, the spaceship photographed Seattle.
However, presumptions made from linguistic decisions such as literal vocabulary definitions and the existence of verb phrases limits user meaning and the actual descriptive function of meaning in English. In some conceivable instances, Seattle can fly to Mars. If the user intends to say that Seattle does fly to Mars, then the original meaning should stay intact. Object-based programming, however, does not accept any sentencing but that which assumes that Seattle cannot fly. Furthermore, and very importantly, the user may intend a conditional meaning which an automatic re-writing by the computer would obscure. For example, the user may have implied "when the spaceship flies to Mars" or "if the spaceship flies to Mars," and so on with the other Conditional Triggers. Therefore, the object-based assumption of directing meaning by the limits of vocabulary actually prevents the real options for re-framing meaning which exist in the context of English.
The present inventive program overcomes the deficiencies of the above prior art. The present program provides a natural language text parser, which may be used for all natural languages, and for the syntactic analysis of text which is simpler, faster, and more compact in memory transfer than those of prior art. To accomplish this, the present program syntactically categorizes and depicts all possible constituents, i.e., all possible word and word pattern combinations, which comprise natural language. It also depicts, in algorithms and on the screen, the order in which all possible constituents, i.e. all possible word and word pattern combinations, can combine. The present program provides a method and device to provide linguistic and morphological analysis on a text corpus to derive part-of-speech, and ultimately semantic information, even if a match is not found in a look-up dictionary for a particular word in the text. The present program also provides a language analyzer, parser, and editor which use a finite set of very simple grammatical and syntactical rules, rather than a large, set of complicated grammatical rules, which merely apply a better approximation. Instead, simple rules in the present program encompass and include the individual variation of language use. Thus the program allows the analysis of any text in natural language.
The present program provides a language analyzer which is context-based, rather than object-based. In such a system, it is word order that determines meaning in context, rather than having the lexical definition alone determine the word order. In the same way, the present program treats the sentence as a whole initially and then classifies the individual words and word patterns, as opposed to focusing solely on individual words and how they bind to other words.
The program breaks the paradigm of chaining words together from left to right to create meaning, and thereby allows the reader to overlap and connect versus chaining meaning. Therefore, the user chooses the word-pattern framework.
The present program gives the user the tools necessary to recognize word groups identified in traditional grammatical terminology within a sentence, as well as named as the invention's constituent word groups. The present program enhances the user's ability to perceive the four basic constituents of language as binary elements of what are termed herein and defined below as Primary Image, Conditional Image, Process Detail, and Background Detail placed in binary positions, where binary refers to the two-part Nounness positions of Doer or Receiver, and the second position as Qualifier.
In particular, all words and constituent-element word groups in English, except verb in tense, function in the program as Nounness or Qualifiers where Nounness operates as any placement of a word or constituent element as a noun including a word defined as a single word noun naming person, place, or thing.
As used herein by the program, Qualifier refers to the vernacular term "modifier", including single-word adjectives and adverbs, but herein also encompasses constituent-element word groups placed as modifiers.
Further, Verbness, as used herein by the program, encompasses two completely different uses of the vernacular term for verb as defined as a word expressing action or state of being. Whether such a word operates by formula with spelling changes, or in combination with auxiliaries, to form verbs in tense, or whether such a word follows the constituent element patterns used herein as Process Details and called in conventional grammatical terms the verbals (the participle, past or present, gerund, or infinitive), the term Verbness includes interchangeably the power of the verb to function as a constituent element of Image or separately, by formula, as a Detail constituent element.
Further, Equalizers, as used herein by the program, include the terms and lists in conventional grammar, known as conjunctive adverbs and coordinating conjunctions, and encompassing all words or phrases used to introduce or join Primary Image with an Equalized relationship. These words Trigger relationship which signal equally weighted idea relationships. Please see Appendix 1 for a list of these.
Further, the Conditional Triggers as used herein by the program include the terms and lists in conventional grammar, known as subordinating conjunctions and relative pronouns, and encompassing all words or phrases used to introduce, create, and join Conditional Image to sentences. These words Trigger relationships of ideas which either support or limit by circumstance the meanings to which they are attached as Nounness or Qualifier. Again, please see Appendix 2 for a list of these.
Further, the Process Detail Triggers as used herein by the program function by formula as added to the structure of verbs preventing them from use as verb in tense. The Process Details add meaning of past, on-going, or future processes to other parts of the sentence to which they are attached as Nounness or Qualifier. Again, please see Appendix 3 for a list of these.
Further, the Background Detail Triggers used herein by the program include the terms and lists in conventional grammar, known as prepositions, and encompassing all words and phrases used to introduce, create, and join Background Detail to a sentence. These words Trigger and add relationships of time, place, and grouping as meanings to other parts of the sentence to which they are attached as Nounness or Qualifier. Again, please see Appendix 4 for a list of these.
These objects are met by the method and device of the present program for parsing natural language text. This program categorizes parts of an inputted text into one of four patterns symbolized by colors on the computer screen. The present program classifies syntax in natural languages such as English. The invention may be advantageously used, for example, in the context of a syntax checker or natural-language processor. In one embodiment, the invention comprises software, which dissects text, classifying its components into two units denoted herein as: Nounness and Verbness. These units are optionally combined in the formation of different types of word patterns. A provision for positioning the word patterns of the text in different locations allows the user to vary the text's meaning. Such analysis provides a way to communicate the textual meaning of a document in terms understandable by a digital computer.
In the present program, as implemented by a computer program, the two basic binary elements of natural language, Nounness and Verbness, combine in two Image word patterns, denoted as Primary Image and Conditional Image, and two Detail word patterns, denoted as Process Detail and Background Detail. The two basic elements, Nounness and Verbness, function binarily, either in combinations for the two Image word patterns or separately in the two Detail word patterns. All word patterns, outside the verb-in-tense, function binarily in one of two positions: as Nounness, called Nesting, or as modifiers, called Qualifying. Since meaning in a natural language sentence is determined by word- or word-pattern location, the meaning can be changed by moving word patterns from one location to another, called "Flipping". Natural language thus organized into binary units constitutes the parsing method and device of the present invention as expressed a program for purposes such as, but not limited to, parsing sentences, analysis of meaning, natural language processing, context analysis of documents, data-base searching, syntax analysis of documents, and the teaching of writing in the English language.
Embodied in the present invention lies a variety of applications. The present invention establishes a base line for emulating textual thought which moves the use of text past the current field of simply text processing. Text becomes an extension of thought represented on the screen as bigger patterns of meaning in context combinations rather than solely a string of words from left to right. Therefore, an extended variety of applications includes any representation of text as inputted by voice, for example, or as text intended to be communicated in written form which the user rearranges and re-orders for varied and multiple options of connected and developed thought. Once combined with voice processing technology, the user will be able to speak text while seeing its context framework arranged and highlighted in color on a computer screen. The present invention, e.g., addresses the occurrence of homonyms because fixed word order clearly determines by context that the word "to" could never be used interchangeably with the word "too", for example. Even a spoken homonym must have fixed word order placement which the present invention correctly identifies. Further applications include natural language processing and programming, machine translation, data-base searching, "text engine" document search, artificial intelligence and other language-related applications. Currently, the industry is searching for natural language solutions for these applications, but the object-based solutions cannot solve problems created when vocabulary words either sound alike or have more than one meaning.
The present invention allows an architecture for symbolizing constituent elements and components of thought reducing the flow of data on optic fibers, for example, to very small pieces of information. Data flow with language can become compressed in a binary fashion with the present invention adapted to telecommunication. Also in the field, machine translation for foreign languages frustrates users because the literal word-for-word translations cannot account for the differences between inflected languages and a language which depends on word order for meaning.
The present invention provides a means of identifying syntax in English. Therefore, the present program can serve as the foundation for comparing English syntax to that of other languages for the purposes of teaching syntax from one language to another or for machine translation. Further, the binary nature of the present invention offers a solution for creating a natural-language programming code which can dependably operate without debilitating exceptions. An example provides herein a sample of how symbols could represent larger pieces of meaning for symbolic encryption.
The goal of the present invention is to create a computer program to identify the algorithmic architecture of language which identifies bigger blocks of meaning than single words. The computer program, by means of this invention, operates from blocks, or patterns, of meaning which, in turn, operate by symbolic formulae. The present invention analyzes what the user originates through an inputted text. In such a fashion, the present invention changes the technology from an object-based premise to a context-based foundation for examining the chain of events which the user has connected as a particular thought. A context-based architecture creates a technology where the user can predict new directions of thought and meaning beyond the object-based definitions of words in isolation which does not provide any means for the user to control changes in options of thought. The present invention, therefore, accounts automatically for the differences in meaning inherent in single words as they connect to produce varied meaning by changes in context.
The following examples illustrate how a context-based word order influences the meaning of connected words in context.
a) The dog bit the man.
The man bit the dog.
In these two sentences the identical vocabulary has completely different meaning in context.
b) We will wind the clock.
The wind blows.
Clock my time as I run.
The present invention identifies the differences in multiple meanings for identical forms. Without the context placement, the user cannot know which meaning of "wind" or "clock" applies. In the case of "wind," the pronunciation changes as the context changes; in the case of "clock," however, the meaning change does not coincide with a change in pronunciation to reflect the change in meaning.
In both cases, however, the current program creates the context meaning without the necessity of pronounced verbal clues. The current program uses the written text and context structure to indicate the differences in word definitions from fixed-word-order placement.
c) My uncle owns a still in West Virginia. When the tax people search for him, he stills the engines and hides under the still, lying very still; still, he runs his still next to the still-life store.
Any English speaker knows that context changes the meanings of the word "still". The present invention identifies these changes without dictionary definitions to choose from in a look-up dictionary.
d) 'Twas brillig, and the slithy toves
Did gyre and gimble in the wabe;
All mimsy were the borogoves,
And the mome raths outgrabe
The present invention organizes bigger blocks of meaning by defining the syntactical architecture which has given these famous, nonsensical lines from Lewis Carroll's "Jabberwocky" context-based meaning. Therefore, the individual words do not have lexical meaning but, on the contrary, are not nonsensical lines because they have context meaning.
Thus, the context-based architecture of the present invention creates a syntactical paradigm which supersedes literal, object-based programming.
FIG. 1 shows a diagram of the structures into which the present program categorizes natural language.
FIG. 2 shows a diagram of the various binary constituents of FIG. 1 and their flow to binary placements.
FIG. 3 is a table of binary constituents which describe the constituent elements and comparatively align their elements.
FIG. 4 is a flip chart illustrating how the computer determines options to move or Flip data from one pattern to another and then how to place these patterns once structured.
FIG. 5 is a diagram showing a type of data compression.
The present invention is a computer-implemented parser for natural languages which supersedes traditional, generative, and object-based analysis by creating a binary architecture which defines context and parts of speech by word order in any use of language text. As such, it may be implemented on a computer, for example, within a software program or coordinated with a voice-synthesizing program or product. Provided a requisite amount of memory is available, it may also be implemented directly in a hard-wired system, such as in ROM chips. Moreover, while reference will be given to specific steps, it is clear to one skilled in the art that a number of different types of code may be used to implement these steps.
An embodiment of the present invention was first written in the language C and was enhanced in C++. A system of 4 MB would be sufficient to run the program. Please refer to the attached example 2 for a detailed description of the hardware specifications.
A first step in the program is the inputting of a stream of alphanumeric data delimited by a punctuation mark. More than likely, in the typical text to be analyzed, there will be a number of punctuation marks indicating a number of different sentences. Generally, in the what may be termed the "group mode", an entire file of text is analyzed at once or the user may choose to select only a part of a file for analysis. This text may come from a number of sources. For example, it may be imported from an ASCII text file, or it may come from a text file having a predetermined format, such as that from a standard word processor program. Alternatively, the user may create an original document within the program itself. Additionally, it may be imported from a magnetic tape, or from a data file stored in a semiconductor memory. Of course, numerous other methods may be used to store the data file as is well-known in the art.
The menu and tool bar options are described in detail in example 2. The screen colors serve as an integral aspect of the actual presentation and working architecture of the logic which presents the results of the analysis to the user. In no way can the colors on the screen be considered as merely secondary or decorative graphics, but instead they serve as an integral element of the program's operating system as the source of context meaning for the user. The colors and their significance are described more fully later. The program of the present invention can be run by a typical CPU with a standard keyboard used to effect choices of menu or tool bar commands.
The program operates on the premise that the verb is the "axis." As such, the location of the possible verbs must be located first. This is done in conjunction with a dictionary look-up system in which the dictionary comprises words categorized as one or more parts of speech. This dictionary contains no lexical definitions. Accompanying the dictionary look-up system is a word-order analyzer which separates the possible verbs from other words by word order and morphology, to be explained later with reference to a specific embodiment. In this way, the definition and isolation of the verb is performed by the system.
The Help system is not a tutorial in the embodiment described of the program; it serves as an actual interface between the logic of the system and the user's needs. The user makes changes in meaning from among the options offered by the Help system while, within the working of the internal programs, the order of the searching logic describes and re-describes the constituent elements of the user's text. Example 2 contains a detailed description of the mechanism and order of the searching logic.
A similar search and word-order analysis is performed in this respect in separate searching steps regarding the remaining parts of speech, including, and preferably in this order: noun, pronoun, adjective, adverb, conjunction, preposition, interjection.
After the word-isolation step is completed, the program moves on to analyze the constituent elements, i.e., the word patterns, of the text. These constituent elements are of four basic types: Primary Image, Conditional Image, Process Detail, and Background Detail. Prior to discussing these constituent elements in more detail, however, it is important to note that a word search is also conducted to determine the presence and location of certain words which function as Triggers for Conditional Images, Background Details and Process Details, and between Primary Images. Conditional Triggers are words which signal Conditional word group patterns for conditional relationships among Images. Conditional Triggers include subordinating conjunctions and relative pronouns and any combinations of words which serve as Conditional Triggers and are listed as such. Of course, for purposes of the program, they comprise a data file filled with an exhaustive list of these types of words. Background Triggers are words and combinations of words which signal Background word group patterns for background data added to other constituent-element word groups. Background Triggers include prepositions and are listed as such.
Background Triggers signal background: specifically, of time, place, or grouping.
Another type of Trigger for which the program searches is the Equalizing Trigger which signals Equalized and paired relationships among words and groups of words. These include conjunctive adverbs and coordinating conjunctions and are listed as such in Appendix 1.
The four basic types of word patterns, i.e., constituent elements, determined by the program, are the Primary Image, the Conditional Image, the Process Detail, and the Background Detail. These word patterns, or constituent elements, are shown in FIG. 1.
Example:
Primary Image: The system failed.
Conditional Image: Although the system failed
Although the original data does not change, the program recognizes that the flow of the thought automatically alters simply by the addition of the Conditional Trigger which creates a circumstance and a contrast, along with creating the expectation in a software user of additional meaning. On the other hand, the use of a different Conditional Trigger will replicate these changes while creating entirely different meaning:
For example, Unless the system failed
Process Detail To fail
To fail, the system must by-pass two defaults.
failed
The failed system had backups which saved the data.
failing
The system's failing rightened the client.
Background Detail: in the budget department
The system failed in the budget department.
The present program identifies a Primary Image whenever it finds a function of Nounness followed by a verb, where Nounness is defined as any Conditional Image, Process Detail, Background Detail, Primary Image in quotation marks, single-word nouns or multiple single-word nouns, personal pronouns, indefinite pronouns, or demonstrative pronouns. Any part of speech or word construct which functions in one of the binary Nounness locations, by formula, serves as Nounness as defined within the computer program. When followed by a verb, Nounness is denoted an agent/doer, and classifies as the subject of a clause, i.e., an Image. The present program will find a Primary Image rather than a Conditional one when the agent/doer, followed by a verb, is not preceded by a Conditional Trigger. If it is preceded by a Conditional Trigger, a Conditional Image will be identified by the program.
______________________________________ Conditional Image What she said inspired me. Process Detail Her saying that inspired me. Background Detail At issue is her job. Primary Image in quotation "I like him" is what she marks said. single-word nouns The speech surprised me. multiple single-word nouns Her speech and manner surprised me. personal pronouns She spoke. definite pronouns Someone spoke. demonstrative pronouns That is what she said. ______________________________________
Nounness has two placements in the program: doer or receiver. All of the examples above demonstrate the doer position but can be "Flipped" by the program to the other, receiver position. In the program, the receiver position follows a verb in tense in either a Conditional or Primary Image. Moving the above underlined examples to the other side of the verb would change the Nounness to a receiver position and therefore change the meaning of the sentence significantly. In the program, the doer position of Nounness occurs only in the two Image constituents. Both Image constituent elements contain the receiver position and both Detail constituent elements can contain receivers within the method of the invention. In the program, the Background Detail contains the receiver as an integral part of its formula but the Process Detail does not have a receiver when the Process Detail has an intransitive property. Also the -ed form of Process Detail never has a receiver within the method of the invention. Within the program, the only shift in position which would require a change in wording for the constituent element is when the personal pronoun must change its case from nominative to objective, in the terms of conventional grammar. The so-called nominative case pronouns function within the method of this invention solely as the doer or following a linking verb in tense, just as the objective case pronoun, as it is known, serves solely in a receiver position. Although these pronouns exist for that purpose, variations may occur because individual use in a dialect structure does not change the context and because fixed-word order has more meaning than do the words chosen to be put in those positions. The patterned positions of fixed word order, not vocabulary, control meaning, e.g., Him and me went to the store, or I want to know who he chose.
A typical document may say a great deal more in context than vocabulary alone. The present invention allows a simple, effective, and powerful contextual insight not otherwise accessible. Examples in this description will illustrate the ease of insight through the color-coded text analysis. However, in the present application, such colors will be denoted by boldface, italics, underlining, small text, and normal text.
One illustration of Nounness, which also showcases the program's nesting function and the inter-relationship of writing processes, follows. Wording from the first sentence will illustrate. The passage, "for understanding and forecasting the structure, variability and dynamic interaction," combines the Process Detail with a Background Detail. The word "for" Triggers the Background Detail context which the present invention would color orange. The phrase, "understanding and forecasting the ocean's structure, variability and dynamic interactions," uses the Process Detail formula to create the context of a process (done by the agency in the document, NOAA) and places it as the receiver of the Background Detail. The present invention colors the Process Detail green, superimposing it on an orange background to illustrate the Nesting, or Nounness, function. The full text of the example follows and a full analysis is given in Appendix 5.
Our present source of information for understanding and forecasting the ocean's structure, variability and dynamic interactions with the atmosphere is a loosely organized network of global satellite and conventional observations. This ocean network is a diverse composite of operational and research measurement systems and platforms operated by an equally varied group of responsible agencies, including NOAA, each with different missions and objectives. The measurements from this network serve often unrelated objectives that are dictated principally by the different missions of each operating agency and program. Consequently, the network is a costly, poorly coordinated mixture of observing systems with little commonality among them. Thus, we experienced the situation where some ocean regions have no observational coverage at all while other areas may have a significant amount of data being collected. Furthermore, incomparabilities in platform type and location, measuring instruments, data formats, quality control procedures and communication links, combined with inadequate operational data assimilation capability precludes full exploitation of the information that is available. This is a costly, inefficient, unacceptable situation for operational ocean forecasting. The problem is exacerbated by the inability to handle both the increasing volume of data from satellites (and real time automated conventional systems) and the expected increase in observations of all kinds needs to satisfy the requirements of fine resolution models for increased forecast accuracy.
In the present invention, Primary Images themselves must be joined together by Equalizers including conjunctive adverbs or coordinating conjunctions or a semicolon. Conditional Images must be joined to other Images and to each other in a similar fashion, these Conditional Images being initiated and joined by Conditional Triggers, including subordinating conjunctions and relative pronouns. As used herein, the term "join" may refer to adjacent constituent elements of Image in the case of Qualifier; or the term "join" may refer to nonadjacent, that is, stacked, or Nested, Images used as Nounness in which the Conditional Trigger introduces a Conditional Image which itself operates as part of another constituent element.
Using the above text example, the joining of concepts with Equalizers and Conditional Triggers provides one of the few clues to the intended meaning of the author. The following sentence therein illustrates Equalizers. "Consequently, the network is a costly, poorly coordinated mixture of observing systems with little commonality among them".
The Equalizer, "consequently," ties the prior Primary image to the Image following "consequently". This Equalized relationship creates a causal meaning that would not otherwise be understood.
In the text, two Conditional Triggers specify where the data has relevance.
The following examples illustrate Conditional Triggers" " . . . where some ocean regions have no observational coverage at all while other areas may have a significant amount of data being collected".
Within the inventive program, the Conditional Triggers, "where" and "while" specify particular conditions applying to the identified ocean regions, preventing generalization.
The Image creates the basic unit of an idea in the present program. The movement, or verb, within the Image creates the core of that idea. As used herein and in the program, the word `movement` refers to any verb in tense which expresses mental, physical, or emotional "doing" initiated by a "doer", that is, a form of Nounness fulfilled by a word or constituent element. Without the movement, the program recognizes no Image. In fact, this movement unit serves as the "axis" of the Image. Even an Image with only the movement stated has a doer understood by the computer, such as in commands. The verb is the axis of the Image, and every sentence in the program must have at least one Image.
Conventional grammar describes verbs as action words. The modern user, therefore, can feel lost or confused if verbs do not suggest, visually, physical action. From a lexicon of 155,000 words, the text corpus of a collegiate dictionary, for example, contains approximately 9400 verbs. A significant proportion of these verbs do not in any way express physical, visible movement. Verbs may express movement but without a physical embodiment. Even thoughts and emotions constitute a type of movement. Similarly, existence constitutes a "movement" in time and space. Therefore, the present program identifies "Verbness," whether as a form of tense in Image or as a Process Detail, to be movement. Examples of such verb movement can be physical, mental, and emotional.
Physical: walk, run, see, take, speak, paint, talk. . . Mental: think, meditate, consider, seem, appear, remain . . . Emotional: feel, love, hate, resist, wish . . .
English has a lexicon of 17 so-called linking verbs. They serve to show equality between the doer and the receiver or a receiver and some prior referent. This lexicon contains a few verbs which distinctly mean "existence" or some shade of existence.
Existence: to be, is, are, were, was, were, am, been, being, seem, become, appear, remain, feel . . .
The term "axis" refers to the position of verb as it occurs in tense in Images within the program because the sentence, or multiple clauses within sentences, cannot exist in English without the verb placed in near proximity to its Doer. At its simplest level, a verb which serves as a command can serve in one word within the program as a full sentence because its Doer is a silent and universally understood "you".
Example: go.
What you are is what you eat.
The verb in tense in this example is the word "is". Contrary to conventional descriptions of English and the current usage of those in the field of linguistics, the verb "is" does not allow for completed meaning in a simple sentence form by any stretch of the imagination. Within the inventive program "is" serves as the axis between the Doer and the Receiver and, thereby, the present invention would color "is" blue to indicate that the Primary Image axis of this sentence example serves to support the relationship between the doer and receiver and cannot, thereby, serve as meaning by itself. In linguistic terms, the verb in tense initiates the verb phrase. However, in the present invention, the verb in tense does not serve as a Trigger for forthcoming meaning. Rather, the verb in tense is an axis around which the meaning revolves for each Image.
To summarize Images, the combination of Nounness and Verbness creates a Primary Image. If the Primary Image is preceded by a Conditional Trigger, however, a Conditional Image is created.
______________________________________ Primary Image The system failed. Conditional Image Although the system failed . . . ______________________________________
Thus, the Image is a combination of Nounness and Verbness. Of course, it is important to note, as the prior examples illustrate, that certain varieties of verb forms, Process Details, such as infinitives, participles, and gerunds, may function as Nounness.
Besides Images, the other set of word patterns is the category of Details. The program categorizes Details in two different sub-categories: Process Details and Background Details. In contrast to Images, Details are not combinations of Nounness and Verbness. The program recognizes two binary elements in English, Nounness and Verbness. Together they form an Image and separately, they form the two Detail patterns.
Process Detail . . . failing . . .
Background Detail . . . in the system . . .
In the above Background Detail, the program would note that the Nounness "system" operates as the receiver of the Background Trigger but does not combine with Verbness. In the above Process Detail, the Verbness "failing" operates within the program as a Process Detail without connection to nounness/doer. Therefore, the two, binary, constituent elements of Nounness and Verbness comprise the two Detail structures when they operate separately from each other.
For example, a Background Trigger, which is conventionally a preposition, typically comprises a Background Trigger preceding an Agent/Receiver. That is, a Background Trigger precedes Nounness. Please see Appendix 4 for a list of those Triggers.
By contrast, a Process Detail within the meaning of the program has more of the qualities of Verbness. Although visually, the Process Detail looks like a verb, the Verbness here does not create an event or idea, as does an Image. The present program notes that the Process Detail has the qualities of a verb including one of the transitive, intransitive, or linking properties as well as the expression of movement; but the Process Detail never has tense. For example, the combination of the word "to" and a verb, with an optional agent/receiver, may create a Process Detail. Similarly, verbs ending in "-ing", with an optional agent/receiver, verbs ending in -"ed", as well as verbs ending in certain specific, irregular endings (-n, -ng, -t, -d, -k), typically create Process Details and would be colored appropriately in the present program. Process Details have no agent/doer. They may have agent/receivers, however.
Furthermore, Process Details in and of themselves function as either agent/doers or agent/receivers or as Qualifiers. Please see Appendix 3 for a list of these.
Auxiliary verbs are also considered. These verbs operate in conjunction with verbs in tense without changing the tense meanings. They add possibility, or conjecture, but do not, in this program, in any way alter the patterning of tense use. Please see Appendix 8 for a list of these.
The program's identification of the various types of word patterns and their locations, as described above, determines the meaning of a sentence. Since context meaning in an English sentence is determined by word and word-pattern location, binary units can be created which allow meaning to be changed by moving words and word patterns from one location to another, called "Flipping" in this embodiment of the invention. FIG. 4 shows the allowable Flipping options.
The reader's understanding drives the meaning as determined by the computer. The user of the present program can either write what seems to fulfill the intended meaning, then vary the placement to choose optional meaning, or, the user can read documents from other people to determine by context either the intention of the text or how to respond in like manner, contextually. The following sentences illustrate from among the various allowable placements, in fixed word order context, available for the arrangements of constituent elements in thought patterns.
Flipping options. In other words, changing the pattern, or meaning.
What she said inspired me.
I love what she said.
Although influenced by what she said, I made another commitment.
Wanting what she said to influence the staff, I raved about her and her speech.
Although what she said inspired me, I had to try my own plans first.
The user may choose to simply Flip the Verbness in an Image to a Process Detail:
Saying that influenced me. The Process Detail "saying that" can be Flipped to any qualifier or nounness position:
I resent her saying that.
To suggest saying that constitutes the worst type of irresponsibility.
By saying that, you have supported me.
From the above sentences, a word or phrase used in a Nounness position may be Flipped to another Nounness position which would change the meaning significantly.
Example: My commitment changed significantly after what she said. Our original friendship grew from shared commitment. Wanting her commitment, I pressed her for her opinion. Although I wanted her commitment, I accepted her promise.
The Flipping function allows the user to re-format data and thereby to access different contextual relationships and allowing the user multiple varied options for connecting thought.
The data in the constituent element may, as in this case, stay in exactly the same form from one sentence placement to another. However, the meaning that the information flow creates from the way the constituent element connects to another constituent element by changing placement completely alters the knowledge imparted.
The present program uses the fact that word pattern location determines meaning to present multiple perspectives to the user while the user analyzes a particular sentence or text. These alternatives are combinations of Images and Details or the data in them in different orders. Obviously, many different meanings can be conveyed by many different constituent element orders. In fact, once a user changes the pattern, the meaning is usually changed also. The user can choose among sentencing options and thereby weigh emphasis and focus, and thereby locate potential ambiguities and create the meaning intended. In this way, the user can choose the least ambiguous combination of words or uncover a completely different interpretation for the same data.
It should further be noted that the division of word patterns in the method of the present invention into Primary Images, Conditional Images, Background Details, and Process Details may be continued by the categorization of three of these types of word patterns into two further categories of placement as Nounness and Qualifiers. Primary Images can be forced into these placements if the Primary Image is set first in quotation marks. This is because Conditional Images, Background Details and Process Details must be used in and of themselves to create either Nounness or Qualifier. Primary Images do not ordinarily function as Nounness or Qualifier. FIGS. 2-4 all illustrate these binary placements. The sentence examples given later represent the allowable Nounness and Qualifier placements.
In addition to the foregoing, Images and Details in and of themselves may be used to show the user the underlying, contextual meanings being communicated in the text. For this purpose, the present system advantageously colors the Images and Details on the computer screen in such a way as to communicate what kind of Image or Detail is present in any particular text. The present method of the program uses a consistent coloring scheme to communicate the borders of the constituent element word patterns. These colors consistently communicate the type of context which the constituent elements symbolize. This specific color scheme is explained later in this paragraph. Just as the program categorizes English wording into one of these four patterns symbolized by colors, nothing that appears on the screen remains uncolored. If the present invention leaves any wording uncolored, that wording qualifies for inclusion in more than one pattern and thereby indicates to the user ambiguity in the writing. In Example 3 given later, the word "needs" can operate either as a verb in tense which would combine with the prior wording "and the expected increase . . . " to create a blue Primary image; or, the same wording "and the expected increase . . . " ties to a prior "both . . . " and therefore "needs" either requires "of" to be written before it or the writer neglected to complete the Image. Furthermore, depending on the location, these coloring schemes may overlap. Primary Images are colored blue; Conditional Images are colored pink; Background Details are colored orange; and Process Details are colored green. Thus, when looking at the coloring on the screen, the user is able to see which word patterns are occurring and in what orders.
The type of information communicated to a user by these colors is as follows. A blue Primary Image provides the skeleton of a sentence. A number of blue-colored patterns show equality of strength among the different Primary Images. Primary Images may serve as central or supporting points, however, but not necessarily as the only point. Primary Images occasionally but only rarely Nest or Qualify, whereas they usually serve as a support structure for Nesting and compressing meaning. A dominance of Primary Images may signal being too happy or literal. In this case, for example, the user may not be showing enough shades of meaning.
For example, The system failed.
The present invention would color this group of words blue.
It should be noted, however, that different types of writing require different types of word patterns. Thus, to say that a dominance of Primary Images tends to make the text too choppy or literal, is purely subjective. In the method and device of the present invention, the user has complete control over the written product. Therefore, commenting on the strength or weakness of the user's text is not a function of the present invention. This invention does not espouse a particular, or universal, style nor suggest repudiation.
The color pink, indicating a Conditional Image, reveals shades of meaning. For example, Conditional Images communicate causes or time constraints. They show contingency. They expand or limit events. They create idea relationships, and dimensions of time and place; they provide supporting information of quantity, quality, rank, degree, cause, contrast, exclusion.
For example, Although the system failed, the default function prevented damage.
The present program would color "Although the system failed" pink.
A phrase colored green indicates a Process Detail which compacts or compresses ideas, or layers meaning. The user may thus decide to more explicitly identify agents or causes or sources of action. The use of too many Process Details, shown by an abundance of green on the screen, may indicate to the user that they are overemphasizing how something is done rather than who or what caused the events. On the other hand, the user may choose to compact multiple sentences into more economical wording with Process Details.
The system's failing triggered two defaults.
The present invention would color "the system's failing" green.
A phrase colored orange indicates Background Detail which helps the user make the text more specific. A drawback of the extended use of Background Detail is that the user may obscure the overall picture. On the other hand, the audience or substance of the text may demand details of time, place, and grouping.
The system failed in the budget department.
The present program would color "in the budget department" orange.
The present program parses natural languages. The identification and location of various word patterns allow the user to arrange them at will according to the meaning actually intended. This contrasts with previous systems which are organized around and rely upon individual word meanings so as to make the problem of natural language analysis lexical and not context-based. The program is context-based. That is, the program analyzes language at a higher structural level than prior object-based systems, and is consequently more effective. The present invention maximizes as opposed to limits meaning choices. As a secondary rather than a primary point of focus, the present invention also correctly, and perhaps more accurately than other systems, parses traditional parts of speech. An example of how accurately context-, rather than object-based analysis identifies multiple word uses is illustrated in the following sentence:
Authors pen lines.
This sentence contains three words, all of which a dictionary would list as verbs, nouns, and adjectives. Only the context-based arrangement accurately identifies which usage applies.
Those skilled in the art will understand that the various optional features of the method and device may be combined in any number of ways without departing from the scope of the present invention, and that this scope is to be limited only by the claims appended hereto.
The present invention utilizes the screen to demonstrate, by color, the context patterns and interrelationships among them in natural text. In conjunction with the screen display, a handbook accompanies the program to help the user choose options of thought in the wording of that text. To make the process easier, both the handbook and the help system include reminder descriptions of how each pattern operates. Users may use, e.g., a set of highlighter markers with the same abbreviated descriptions for reminders and reinforcement. These descriptions also have a use symbolically in conjunction with developing technology for voice activation, machine translation, natural language programming and artificial intelligence. Because the tool bar and the message lines in both the DOS and Windows versions utilized these abbreviations to direct the user, this example elucidates and repeats the same abbreviated definitions. The abbreviated descriptions include PI=A/D+M+(A/R), CI=CT+A/D+M+(A/R), PD=(to+)M(+ing)+(A/R)(+ed)(4), and BD=BT+A/R.
The abbreviated descriptions represent the following constituent elements word patterns:
Primary Image=Agent/Doer+Movement+(Agent/Receiver)
______________________________________ Conditional Image = Conditional Trigger + Agent/Doer + Movement + (Agent/Receiver) Process Detail = to + movement + (Agent/Receiver) = movement + ing + (Agent/Receiver) = movement + ed or the irregular movement spellings symbolized by thenumber 4 for fourth column of verb-movement spelling ______________________________________
Background Detail=Background Trigger+Agent/Receiver
An example follows below for identifying any Conditional pattern. The Example, incidentally, may use a convenient data compression technique as indicated in FIG. 5 which allows the easy look-up of words. Within the industry of current computer technology, telecommunications and natural language processing, a need has arisen to efficiently symbolize bigger blocks of meaning than simply words. The transmission, for example, of voice synthesized communication via optic fiber requires compressing data into units or blocks of meaning which are as small or compact as possible. The constituent word patterns of the present program serve as such units. In other words, all Conditional Triggers, for example, could by symbolized with one type of computer code having symbolic variations to present all of the various Conditional Triggers.
The program runs on an IBM compatible system but may also run on many other platforms, e.g., Apple Macintosh and UNIX. 4 MB of RAM, or more, are recommended. There are approximately 50,000 lines of code, and 110 subprograms combined in the DOS version and 10,000 lines of code plus the zApp software in the windows version. The language is C; however, some of the code has been written in C++. The data files consist of a dictionary, approximately 750K, a translation file, 1K, and a configuration file. The dictionary contains no definitions, only syntactic information listing all the parts of speech a particular word can be. The translation file decodes the encrypted dictionary. The software in this example can highlight a 1000 sentence text corpus; however, as fully implemented, the software can handle any size document.
First, the software opens up a configuration file which tells certain hardware whether, for example, to spell check or not, to use black and white or color, and to write to the screen using fast video memory. Depending upon what the configuration file tells it, it sets color attributes that will be used. Then the user is prompted for a file name. The file is located, which fills up the display memory with that file; or the user may create a new file; or the user may input a file by any other means.
The user, at this point, has several options; for example, Alt+S will begin sentence analysis. The software parses one character at a time from the beginning until it locates the ending punctuation for sentences, including {?, |, .). It fills up the internal memory structures, identifies the beginning and end of each sentence relative to the display memory. It analyzes what kind of sentence each one is, i.e., whether it is simple, compound, complex, or compound-complex. Whichever sentence the cursor is on, is highlighted and centered on the screen.
The user may press W to identify each word's syntactic function. It parses one sentence at a time from the first character to the end and locates the end of word punctuation including {space, comma, semi-colon, quotation marks, or/and period). Then it fills another internal structure called word structure which contains the beginning and end of each word relative to display memory and contains information about the word type from the dictionary and the word's use, e.g., whether a noun is a subject noun or a receiver. As it locates each word, it checks the dictionary to find the first occurrence of the word in the dictionary, then stops when * finds a hit and marks the word with a syntactic part of speech label. Then it goes to the next word. After it identifies all the words in the sentence, it goes through a series of other tests to determine if there are any impossible situations if the words are marked as is. For example, if three words occur adjacently without any intervening commas or coordinating conjunctions that can all be nouns, when at least two of them cannot be nouns in that context, e.g. "Authors pen lines." In this case the computer would check the dictionary to see if the words in question could be any other part of speech and changes the syntactic label accordingly. The program makes approximately 50 scans per sentence. A sample listing of rules followed by the program is given in Appendix 6.
The user may opt to highlight Images in their appropriate colors. The user chooses "I". It parses the word structure from the first word to the end of the sentence, looking for Conditional Triggers and Equalizing and Joining Triggers.
It reads one word at a time, filling up a clause structure which contains the beginning and end of each clause in a sentence relative to the display memory and at the same time, the clause type, in both traditional terms and in the terms of the present program, i.e., whether the Images are Primary, Conditional, or Nested. It analyzes what kind of sentence each one is, i.e., whether it is simple, compound, complex, or compound-complex.
The user may also opt to highlight Details. The Detail logic is nearly the same as the clause logic in that the program searches for Detail Triggers as it scans the words from beginning to end, filling up the phrase structure relative to the display memory. Likewise, it identifies Detail type and highlights each Detail on the screen in the appropriate colors.
Of course, a user may opt to examine many sentences simultaneously to look for structure patterns. The user may select G for group mode and drag the cursor to the desired amount of text. The highlighting occurs identically to the sentence by sentence process except that it runs the logic repeatedly and displays colors for a body of text rather than just one sentence.
If a user opts to begin with Image mode or Phrase mode without first going into Word mode, the program automatically executes the word logic behind the scenes because the word logic is required for the Image and Detail logic. An embodiment of the word logic is demonstrated in Appendix 7.
Sentence variation for re-organizing thought serves as a central point for the user of the present program. Context patterns change their "weight" of substance by the simple additional change of Triggers and patterns. Without the color on the screen to illustrate these differences, the user may so struggle with the concepts that the present invention would be less than easily used. The color, therefore, provides real and substantive meaning beyond the labeling and parsing in the message lines. Furthermore, conventional grammatical labels focus primarily upon single-word use and part-of-speech designation with an incidental nod to group patterns which, although named, seem to have minor significance in the process of communicating. Contrarily, the group patterns, in fact, have more significance that the labels suggest because their interrelationships actually create an ordered meaning which the labels do not well explain. The following example illustrates the contribution that the color coding makes toward the understanding of natural text. The prior Detailed Description contains extensive examples of sentence variations for accessing multiple arrangements of meaning; therefore, those examples are not repeated herein in order to demonstrate the importance of the powerful information created in context and represented in the present invention by consistent color coding, but represented here with the following meaning.
Boldface type represents the "blue-colored" Primary Image.
Underlined text represents the "pink-colored" Conditional Image.
Italicized text represents the "green-colored" Process Detail.
Capitalized text represents the "orange-colored" Background Detail.
______________________________________ Although wanting to know! what you said!! tempted me pink green green pink pink to change what I did! to you, I left. green pink orange blue Although wanting to know! what you said!! tempted me to change what I did! TO YOU, I left. What you said! inspired me in spite of being what you called inattentive. pink blue orange green pink What you said inspired me IN SPITE OF BEING WHAT YOU CALLED! INATTENTIVE!. ______________________________________ ##SPC1##
Claims (17)
1. A method for context-based parsing of natural language text, said method comprising the steps of:
receiving a stream of alphanumeric data representing a plurality of words delimited by a punctuation mark;
associating a semantic label with each of said plurality of words by using a look-up table to identify one or more potential semantic labels for a word, and selecting one of said potential semantic labels based on an order of said word within said plurality of words; and
assigning each of said plurality of words to one of two binary context patterns based on said semantic label associated therewith and an order of said word with respect to one or more others of said plurality of words.
2. The method of claim 1, wherein each of said semantic labels is selected from a group consisting of a function of nounness, a function of verbness, and a qualifier.
3. The method of claim 2, wherein said two binary context patterns comprise an image context pattern and a detail context pattern.
4. The method of claim 3, wherein each of said two binary context patterns includes two binary context sub-patterns, respectively comprising a primary image context pattern, a conditional image context pattern, a process detail context pattern, and a background detail context pattern, said assignment step further comprising assigning each of said plurality of words to one of said binary context sub-patterns.
5. The method of claim 4, wherein said assignment step further comprises identifying a primary image upon detection of a word associated with a verbness semantic label in proximity to a word associated with a nounness semantic label.
6. The method of claim 1, wherein said assignment step further comprises searching said plurality of words for a predetermined trigger word associated with a binary context pattern.
7. The method of claim 6, wherein said predetermined trigger word is of a type selected from a group including a conditional trigger, a background trigger, an equalizing trigger, and a process detail trigger.
8. The method of claim 1, wherein said assignment step further comprises comparing a contiguous portion of said plurality of words and their associated semantic labels to a formula associated with a binary context pattern, said formula comprising a plurality of semantic labels in a predetermined order.
9. The method of claim 1, wherein said step of associating a semantic label with each of said plurality of words further comprises first identifying a potential verb in said plurality of words.
10. The method of claim 9, wherein said step of associating a semantic label with each of said plurality of words further comprises identifying a potential noun in said plurality of words.
11. The method of claim 10, wherein said step of associating a semantic label with each of said plurality of words further comprises identifying a potential modifier in said plurality of words.
12. The method of claim 1, wherein said assignment step further comprises searching said plurality of words for two or more contiguous words associated with a like semantic label.
13. A set of instructions stored on a medium for parsing a natural language text, said set of instructions comprising instructions for:
receiving a stream of alphanumeric data representing a plurality of words delimited by a punctuation mark;
associating a semantic label with each of said plurality of words by using a look-up table to identify one or more potential semantic labels for a word, and selecting one of said potential semantic labels based on an order of said word within said plurality of words; and
assigning each of said plurality of words to one of two binary context patterns based on said semantic label associated therewith and an order of said word with respect to one or more others of said plurality of words.
14. The set of instructions of claim 13, wherein said instructions for assigning said plurality of words to a binary context pattern further comprise instructions for comparing a contiguous portion of said plurality of words and their associated semantic labels to a formula associated with a context pattern, said formula comprising a plurality of semantic labels in a predetermined order.
15. An apparatus for analyzing a natural language text including a plurality of words delimited by a punctuation mark, said apparatus comprising:
a processor;
a memory coupled to said processor, said memory including a look-up table with a plurality of entries, wherein each of said entries associates a semantic label with a word;
a text parser controlled to said processor and coupled to said memory, said text parser comprising a set of executable instructions for assigning a selected one of said plurality of words to a binary context pattern based on a semantic label associated with said selected word and an order of said selected word within said plurality of words.
16. The apparatus of claim 15, further comprising a text reformatter coupled to said text parser, said text reformatter comprising a set of executable instructions enabling a user to alter a meaning of said plurality of words by reordering said binary context patterns.
17. The apparatus of claim 16, further comprising a help system controlled by said processor, said help system including executable instructions for providing a user with guidance related to said text reformatter.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/487,263 US5721938A (en) | 1995-06-07 | 1995-06-07 | Method and device for parsing and analyzing natural language sentences and text |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/487,263 US5721938A (en) | 1995-06-07 | 1995-06-07 | Method and device for parsing and analyzing natural language sentences and text |
Publications (1)
Publication Number | Publication Date |
---|---|
US5721938A true US5721938A (en) | 1998-02-24 |
Family
ID=23935029
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US08/487,263 Expired - Lifetime US5721938A (en) | 1995-06-07 | 1995-06-07 | Method and device for parsing and analyzing natural language sentences and text |
Country Status (1)
Country | Link |
---|---|
US (1) | US5721938A (en) |
Cited By (119)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5841895A (en) * | 1996-10-25 | 1998-11-24 | Pricewaterhousecoopers, Llp | Method for learning local syntactic relationships for use in example-based information-extraction-pattern learning |
US5878385A (en) * | 1996-09-16 | 1999-03-02 | Ergo Linguistic Technologies | Method and apparatus for universal parsing of language |
US5974557A (en) * | 1994-06-20 | 1999-10-26 | Thomas; C. Douglass | Method and system for performing thermal and power management for a computer |
US5999896A (en) * | 1996-06-25 | 1999-12-07 | Microsoft Corporation | Method and system for identifying and resolving commonly confused words in a natural language parser |
US6014616A (en) * | 1996-11-13 | 2000-01-11 | Samsung Electronics Co., Ltd. | Method for monitoring the language used for character generation by an operating system |
US6081772A (en) * | 1998-03-26 | 2000-06-27 | International Business Machines Corporation | Proofreading aid based on closed-class vocabulary |
US6181909B1 (en) * | 1997-07-22 | 2001-01-30 | Educational Testing Service | System and method for computer-based automatic essay scoring |
US20010056350A1 (en) * | 2000-06-08 | 2001-12-27 | Theodore Calderone | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US6346892B1 (en) * | 1999-05-07 | 2002-02-12 | Honeywell International Inc. | Method and apparatus for aircraft systems management |
US20020076112A1 (en) * | 2000-12-18 | 2002-06-20 | Philips Electronics North America Corporation | Apparatus and method of program classification based on syntax of transcript information |
US20020129066A1 (en) * | 2000-12-28 | 2002-09-12 | Milward David R. | Computer implemented method for reformatting logically complex clauses in an electronic text-based document |
US6513063B1 (en) | 1999-01-05 | 2003-01-28 | Sri International | Accessing network-based electronic information through scripted online interfaces using spoken input |
US20030023642A1 (en) * | 2001-07-30 | 2003-01-30 | Spragins James S-B | Method and system for marking writings online |
US20030028368A1 (en) * | 2001-08-01 | 2003-02-06 | Grandy Woodrow W. | Method for entering, recording, distributing and reporting data |
US6523061B1 (en) | 1999-01-05 | 2003-02-18 | Sri International, Inc. | System, method, and article of manufacture for agent-based navigation in a speech-based data navigation system |
US20030074188A1 (en) * | 2001-10-12 | 2003-04-17 | Tohgo Murata | Method and apparatus for language instruction |
US6598021B1 (en) * | 2000-07-13 | 2003-07-22 | Craig R. Shambaugh | Method of modifying speech to provide a user selectable dialect |
US6615168B1 (en) * | 1996-07-26 | 2003-09-02 | Sun Microsystems, Inc. | Multilingual agent for use in computer systems |
US6625608B1 (en) * | 1999-08-24 | 2003-09-23 | International Business Machines Corporation | Method and apparatus for displaying natural language structure |
US6675159B1 (en) | 2000-07-27 | 2004-01-06 | Science Applic Int Corp | Concept-based search and retrieval system |
US6691151B1 (en) | 1999-01-05 | 2004-02-10 | Sri International | Unified messaging methods and systems for communication and cooperation among distributed agents in a computing environment |
US20040044516A1 (en) * | 2002-06-03 | 2004-03-04 | Kennewick Robert A. | Systems and methods for responding to natural language speech utterance |
US6742021B1 (en) | 1999-01-05 | 2004-05-25 | Sri International, Inc. | Navigating network-based electronic information using spoken input with multimodal error feedback |
US20040117189A1 (en) * | 1999-11-12 | 2004-06-17 | Bennett Ian M. | Query engine for processing voice based queries including semantic decoding |
US6757718B1 (en) | 1999-01-05 | 2004-06-29 | Sri International | Mobile navigation of network-based electronic information using spoken input |
US6766316B2 (en) | 2001-01-18 | 2004-07-20 | Science Applications International Corporation | Method and system of ranking and clustering for document indexing and retrieval |
US20040167884A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Methods and products for producing role related information from free text sources |
US20040193420A1 (en) * | 2002-07-15 | 2004-09-30 | Kennewick Robert A. | Mobile systems and methods for responding to natural language speech utterance |
US20040193399A1 (en) * | 2003-03-31 | 2004-09-30 | Microsoft Corporation | System and method for word analysis |
US6810392B1 (en) * | 1998-07-31 | 2004-10-26 | Northrop Grumman Corporation | Method and apparatus for estimating computer software development effort |
US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050010553A1 (en) * | 2000-10-30 | 2005-01-13 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20050033581A1 (en) * | 2001-02-16 | 2005-02-10 | Foster Mark J. | Dual compression voice recordation non-repudiation system |
US20050080625A1 (en) * | 1999-11-12 | 2005-04-14 | Bennett Ian M. | Distributed real time speech recognition system |
US20050091038A1 (en) * | 2003-10-22 | 2005-04-28 | Jeonghee Yi | Method and system for extracting opinions from text documents |
US20050119874A1 (en) * | 1997-03-04 | 2005-06-02 | Hiroshi Ishikura | Pivot translation method and system |
US20050165781A1 (en) * | 2004-01-26 | 2005-07-28 | Reiner Kraft | Method, system, and program for handling anchor text |
US20050165838A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Architecture for an indexer |
US20050165718A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Pipelined architecture for global analysis and index building |
US20050165800A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Method, system, and program for handling redirects in a search engine |
US6941513B2 (en) | 2000-06-15 | 2005-09-06 | Cognisphere, Inc. | System and method for text structuring and text generation |
US20060010138A1 (en) * | 2004-07-09 | 2006-01-12 | International Business Machines Corporation | Method and system for efficient representation, manipulation, communication, and search of hierarchical composite named entities |
US7020326B1 (en) * | 1998-11-13 | 2006-03-28 | Hsu Shin-Yi | System for guiding users to formulate and use object extraction rules |
US20060074668A1 (en) * | 2002-11-28 | 2006-04-06 | Koninklijke Philips Electronics N.V. | Method to assign word class information |
US20060074962A1 (en) * | 2004-09-24 | 2006-04-06 | Fontoura Marcus F | Method, system, and program for searching documents for ranges of numeric values |
US7027974B1 (en) | 2000-10-27 | 2006-04-11 | Science Applications International Corporation | Ontology-based parser for natural language processing |
US7036128B1 (en) | 1999-01-05 | 2006-04-25 | Sri International Offices | Using a community of distributed electronic agents to support a highly mobile, ambient computing environment |
US20060117039A1 (en) * | 2002-01-07 | 2006-06-01 | Hintz Kenneth J | Lexicon-based new idea detector |
US20070010992A1 (en) * | 2005-07-08 | 2007-01-11 | Microsoft Corporation | Processing collocation mistakes in documents |
US20070016583A1 (en) * | 2005-07-14 | 2007-01-18 | Ronny Lempel | Enforcing native access control to indexed documents |
US20070033005A1 (en) * | 2005-08-05 | 2007-02-08 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20070038436A1 (en) * | 2005-08-10 | 2007-02-15 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US20070055525A1 (en) * | 2005-08-31 | 2007-03-08 | Kennewick Robert A | Dynamic speech sharpening |
US20070073678A1 (en) * | 2005-09-23 | 2007-03-29 | Applied Linguistics, Llc | Semantic document profiling |
US20070073745A1 (en) * | 2005-09-23 | 2007-03-29 | Applied Linguistics, Llc | Similarity metric for semantic profiling |
US20070179789A1 (en) * | 1999-11-12 | 2007-08-02 | Bennett Ian M | Speech Recognition System With Support For Variable Portable Devices |
US20070213985A1 (en) * | 2006-03-13 | 2007-09-13 | Corwin Daniel W | Self-Annotating Identifiers |
US20080052078A1 (en) * | 1999-11-12 | 2008-02-28 | Bennett Ian M | Statistical Language Model Trained With Semantic Variants |
US20080086299A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between languages |
US20080086298A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between langauges |
US20080086300A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between languages |
US20080091406A1 (en) * | 2006-10-16 | 2008-04-17 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US7379862B1 (en) * | 1999-11-19 | 2008-05-27 | Microsoft Corporation | Method and apparatus for analyzing and debugging natural language parses |
US20080189110A1 (en) * | 2007-02-06 | 2008-08-07 | Tom Freeman | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US20080228467A1 (en) * | 2004-01-06 | 2008-09-18 | Neuric Technologies, Llc | Natural language parsing method to provide conceptual flow |
US20090070099A1 (en) * | 2006-10-10 | 2009-03-12 | Konstantin Anisimovich | Method for translating documents from one language into another using a database of translations, a terminology dictionary, a translation dictionary, and a machine translation system |
US20090150156A1 (en) * | 2007-12-11 | 2009-06-11 | Kennewick Michael R | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US20090182549A1 (en) * | 2006-10-10 | 2009-07-16 | Konstantin Anisimovich | Deep Model Statistics Method for Machine Translation |
US20090299745A1 (en) * | 2008-05-27 | 2009-12-03 | Kennewick Robert A | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US20100063795A1 (en) * | 2005-08-04 | 2010-03-11 | Nec Corporation | Data processing device, data processing method, and data processing program |
US20100082333A1 (en) * | 2008-05-30 | 2010-04-01 | Eiman Tamah Al-Shammari | Lemmatizing, stemming, and query expansion method and system |
US20100161311A1 (en) * | 2008-12-19 | 2010-06-24 | Massuh Lucas A | Method, apparatus and system for location assisted translation |
US20100185437A1 (en) * | 2005-01-06 | 2010-07-22 | Neuric Technologies, Llc | Process of dialogue and discussion |
US20100217604A1 (en) * | 2009-02-20 | 2010-08-26 | Voicebox Technologies, Inc. | System and method for processing multi-modal device interactions in a natural language voice services environment |
US20110060733A1 (en) * | 2009-09-04 | 2011-03-10 | Alibaba Group Holding Limited | Information retrieval based on semantic patterns of queries |
US20110112827A1 (en) * | 2009-11-10 | 2011-05-12 | Kennewick Robert A | System and method for hybrid processing in a natural language voice services environment |
US7949529B2 (en) | 2005-08-29 | 2011-05-24 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US20110307252A1 (en) * | 2010-06-15 | 2011-12-15 | Microsoft Corporation | Using Utterance Classification in Telephony and Speech Recognition Applications |
US20120158703A1 (en) * | 2010-12-16 | 2012-06-21 | Microsoft Corporation | Search lexicon expansion |
US20120281011A1 (en) * | 2011-03-07 | 2012-11-08 | Oliver Reichenstein | Method of displaying text in a text editor |
US8909595B2 (en) | 2001-08-01 | 2014-12-09 | T-System, Inc. | Method for entering, recording, distributing and reporting data |
US8959011B2 (en) | 2007-03-22 | 2015-02-17 | Abbyy Infopoisk Llc | Indicating and correcting errors in machine translation systems |
US8971630B2 (en) | 2012-04-27 | 2015-03-03 | Abbyy Development Llc | Fast CJK character recognition |
US8989485B2 (en) | 2012-04-27 | 2015-03-24 | Abbyy Development Llc | Detecting a junction in a text line of CJK characters |
US20150142825A1 (en) * | 2013-11-15 | 2015-05-21 | Tata Consultancy Services Limited | Converting procedural text to an actionable knowledge form |
US9047275B2 (en) | 2006-10-10 | 2015-06-02 | Abbyy Infopoisk Llc | Methods and systems for alignment of parallel text corpora |
US9082403B2 (en) | 2011-12-15 | 2015-07-14 | Microsoft Technology Licensing, Llc | Spoken utterance classification training for a speech recognition system |
CN105144149A (en) * | 2013-05-29 | 2015-12-09 | 国立研究开发法人情报通信研究机构 | Translation word order information output device, translation word order information output method, and recording medium |
US9213936B2 (en) | 2004-01-06 | 2015-12-15 | Neuric, Llc | Electronic brain model with neuron tables |
US9235573B2 (en) | 2006-10-10 | 2016-01-12 | Abbyy Infopoisk Llc | Universal difference measure |
US9239826B2 (en) | 2007-06-27 | 2016-01-19 | Abbyy Infopoisk Llc | Method and system for generating new entries in natural language dictionary |
US9262409B2 (en) | 2008-08-06 | 2016-02-16 | Abbyy Infopoisk Llc | Translation of a selected text fragment of a screen |
US9305548B2 (en) | 2008-05-27 | 2016-04-05 | Voicebox Technologies Corporation | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
WO2016149918A1 (en) * | 2015-03-25 | 2016-09-29 | 北京旷视科技有限公司 | Determining of geographical position of user |
US9502025B2 (en) | 2009-11-10 | 2016-11-22 | Voicebox Technologies Corporation | System and method for providing a natural language content dedication service |
US20170011119A1 (en) * | 2015-07-06 | 2017-01-12 | Rima Ghannam | System for Natural Language Understanding |
US9626353B2 (en) | 2014-01-15 | 2017-04-18 | Abbyy Infopoisk Llc | Arc filtering in a syntactic graph |
US9626703B2 (en) | 2014-09-16 | 2017-04-18 | Voicebox Technologies Corporation | Voice commerce |
US9626358B2 (en) | 2014-11-26 | 2017-04-18 | Abbyy Infopoisk Llc | Creating ontologies by analyzing natural language texts |
US9633005B2 (en) | 2006-10-10 | 2017-04-25 | Abbyy Infopoisk Llc | Exhaustive automatic processing of textual information |
US9645993B2 (en) | 2006-10-10 | 2017-05-09 | Abbyy Infopoisk Llc | Method and system for semantic searching |
US9710450B2 (en) | 2015-04-10 | 2017-07-18 | International Business Machines Corporation | Recombining incorrectly separated tokens in natural language processing |
US9740682B2 (en) | 2013-12-19 | 2017-08-22 | Abbyy Infopoisk Llc | Semantic disambiguation using a statistical analysis |
US9747896B2 (en) | 2014-10-15 | 2017-08-29 | Voicebox Technologies Corporation | System and method for providing follow-up responses to prior natural language inputs of a user |
TWI610185B (en) * | 2014-12-22 | 2018-01-01 | 晨星半導體股份有限公司 | Related information displaying method and electronic device that can automatically display related information |
US9858506B2 (en) | 2014-09-02 | 2018-01-02 | Abbyy Development Llc | Methods and systems for processing of images of mathematical expressions |
US9898459B2 (en) | 2014-09-16 | 2018-02-20 | Voicebox Technologies Corporation | Integration of domain information into state transitions of a finite state transducer for natural language processing |
US9977826B2 (en) | 2015-10-21 | 2018-05-22 | Cloudera, Inc. | Computerized method of generating and analytically evaluating multiple instances of natural language-generated text |
US9984071B2 (en) | 2006-10-10 | 2018-05-29 | Abbyy Production Llc | Language ambiguity detection of text |
US9992745B2 (en) | 2011-11-01 | 2018-06-05 | Qualcomm Incorporated | Extraction and analysis of buffered audio data using multiple codec rates each greater than a low-power processor rate |
US10331784B2 (en) | 2016-07-29 | 2019-06-25 | Voicebox Technologies Corporation | System and method of disambiguating natural language processing requests |
US10381007B2 (en) | 2011-12-07 | 2019-08-13 | Qualcomm Incorporated | Low power integrated circuit to analyze a digitized audio stream |
US10431214B2 (en) | 2014-11-26 | 2019-10-01 | Voicebox Technologies Corporation | System and method of determining a domain and/or an action related to a natural language input |
US10515138B2 (en) | 2014-04-25 | 2019-12-24 | Mayo Foundation For Medical Education And Research | Enhancing reading accuracy, efficiency and retention |
US10521497B2 (en) * | 2017-10-10 | 2019-12-31 | Adobe Inc. | Maintaining semantic information in document conversion |
US10592738B2 (en) | 2017-12-01 | 2020-03-17 | International Business Machines Corporation | Cognitive document image digitalization |
US10614799B2 (en) | 2014-11-26 | 2020-04-07 | Voicebox Technologies Corporation | System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance |
US10671801B2 (en) * | 2017-02-28 | 2020-06-02 | Microsoft Technology Licensing, Llc | Markup code generator |
CN114282530A (en) * | 2021-12-24 | 2022-04-05 | 厦门大学 | A sentiment analysis method for complex sentences triggered by grammatical structure and connection information |
Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4641264A (en) * | 1981-09-04 | 1987-02-03 | Hitachi, Ltd. | Method for automatic translation between natural languages |
US4731021A (en) * | 1984-09-14 | 1988-03-15 | Chan See F | Classification language and method |
US4773009A (en) * | 1986-06-06 | 1988-09-20 | Houghton Mifflin Company | Method and apparatus for text analysis |
US4864502A (en) * | 1987-10-07 | 1989-09-05 | Houghton Mifflin Company | Sentence analyzer |
US4887212A (en) * | 1986-10-29 | 1989-12-12 | International Business Machines Corporation | Parser for natural language text |
US4907971A (en) * | 1988-10-26 | 1990-03-13 | Tucker Ruth L | System for analyzing the syntactical structure of a sentence |
US5029085A (en) * | 1989-05-18 | 1991-07-02 | Ricoh Company, Ltd. | Conversational-type natural language analysis apparatus |
US5056021A (en) * | 1989-06-08 | 1991-10-08 | Carolyn Ausborn | Method and apparatus for abstracting concepts from natural language |
US5060155A (en) * | 1989-02-01 | 1991-10-22 | Bso/Buro Voor Systeemontwikkeling B.V. | Method and system for the representation of multiple analyses in dependency grammar and parser for generating such representation |
US5099425A (en) * | 1988-12-13 | 1992-03-24 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for analyzing the semantics and syntax of a sentence or a phrase |
US5101349A (en) * | 1989-03-14 | 1992-03-31 | Canon Kabushiki Kaisha | Natural language processing system |
US5113342A (en) * | 1989-04-26 | 1992-05-12 | International Business Machines Corporation | Computer method for executing transformation rules |
US5146405A (en) * | 1988-02-05 | 1992-09-08 | At&T Bell Laboratories | Methods for part-of-speech determination and usage |
US5146406A (en) * | 1989-08-16 | 1992-09-08 | International Business Machines Corporation | Computer method for identifying predicate-argument structures in natural language text |
US5225991A (en) * | 1991-04-11 | 1993-07-06 | International Business Machines Corporation | Optimized automated macro embedding for standard cell blocks |
US5237502A (en) * | 1990-09-04 | 1993-08-17 | International Business Machines Corporation | Method and apparatus for paraphrasing information contained in logical forms |
US5265065A (en) * | 1991-10-08 | 1993-11-23 | West Publishing Company | Method and apparatus for information retrieval from a database by replacing domain specific stemmed phases in a natural language to create a search query |
US5297040A (en) * | 1991-10-23 | 1994-03-22 | Franklin T. Hu | Molecular natural language processing system |
US5299124A (en) * | 1991-01-10 | 1994-03-29 | Sharp Kabushiki Kaisha | Translation machine capable of analyzing syntactic of sentence |
US5317510A (en) * | 1989-03-15 | 1994-05-31 | Kabushiki Kaisha Toshiba | Method and apparatus for generating sentences from conceptual structures |
US5321607A (en) * | 1992-05-25 | 1994-06-14 | Sharp Kabushiki Kaisha | Automatic translating machine |
US5331556A (en) * | 1993-06-28 | 1994-07-19 | General Electric Company | Method for natural language data processing using morphological and part-of-speech information |
US5337232A (en) * | 1989-03-02 | 1994-08-09 | Nec Corporation | Morpheme analysis device |
US5338976A (en) * | 1991-06-20 | 1994-08-16 | Ricoh Company, Ltd. | Interactive language conversion system |
US5559693A (en) * | 1991-06-28 | 1996-09-24 | Digital Equipment Corporation | Method and apparatus for efficient morphological text analysis using a high-level language for compact specification of inflectional paradigms |
-
1995
- 1995-06-07 US US08/487,263 patent/US5721938A/en not_active Expired - Lifetime
Patent Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4641264A (en) * | 1981-09-04 | 1987-02-03 | Hitachi, Ltd. | Method for automatic translation between natural languages |
US4731021A (en) * | 1984-09-14 | 1988-03-15 | Chan See F | Classification language and method |
US4773009A (en) * | 1986-06-06 | 1988-09-20 | Houghton Mifflin Company | Method and apparatus for text analysis |
US4887212A (en) * | 1986-10-29 | 1989-12-12 | International Business Machines Corporation | Parser for natural language text |
US4864502A (en) * | 1987-10-07 | 1989-09-05 | Houghton Mifflin Company | Sentence analyzer |
US5146405A (en) * | 1988-02-05 | 1992-09-08 | At&T Bell Laboratories | Methods for part-of-speech determination and usage |
US4907971A (en) * | 1988-10-26 | 1990-03-13 | Tucker Ruth L | System for analyzing the syntactical structure of a sentence |
US5099425A (en) * | 1988-12-13 | 1992-03-24 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for analyzing the semantics and syntax of a sentence or a phrase |
US5060155A (en) * | 1989-02-01 | 1991-10-22 | Bso/Buro Voor Systeemontwikkeling B.V. | Method and system for the representation of multiple analyses in dependency grammar and parser for generating such representation |
US5337232A (en) * | 1989-03-02 | 1994-08-09 | Nec Corporation | Morpheme analysis device |
US5101349A (en) * | 1989-03-14 | 1992-03-31 | Canon Kabushiki Kaisha | Natural language processing system |
US5317510A (en) * | 1989-03-15 | 1994-05-31 | Kabushiki Kaisha Toshiba | Method and apparatus for generating sentences from conceptual structures |
US5113342A (en) * | 1989-04-26 | 1992-05-12 | International Business Machines Corporation | Computer method for executing transformation rules |
US5029085A (en) * | 1989-05-18 | 1991-07-02 | Ricoh Company, Ltd. | Conversational-type natural language analysis apparatus |
US5056021A (en) * | 1989-06-08 | 1991-10-08 | Carolyn Ausborn | Method and apparatus for abstracting concepts from natural language |
US5146406A (en) * | 1989-08-16 | 1992-09-08 | International Business Machines Corporation | Computer method for identifying predicate-argument structures in natural language text |
US5237502A (en) * | 1990-09-04 | 1993-08-17 | International Business Machines Corporation | Method and apparatus for paraphrasing information contained in logical forms |
US5299124A (en) * | 1991-01-10 | 1994-03-29 | Sharp Kabushiki Kaisha | Translation machine capable of analyzing syntactic of sentence |
US5225991A (en) * | 1991-04-11 | 1993-07-06 | International Business Machines Corporation | Optimized automated macro embedding for standard cell blocks |
US5338976A (en) * | 1991-06-20 | 1994-08-16 | Ricoh Company, Ltd. | Interactive language conversion system |
US5559693A (en) * | 1991-06-28 | 1996-09-24 | Digital Equipment Corporation | Method and apparatus for efficient morphological text analysis using a high-level language for compact specification of inflectional paradigms |
US5265065A (en) * | 1991-10-08 | 1993-11-23 | West Publishing Company | Method and apparatus for information retrieval from a database by replacing domain specific stemmed phases in a natural language to create a search query |
US5297040A (en) * | 1991-10-23 | 1994-03-22 | Franklin T. Hu | Molecular natural language processing system |
US5321607A (en) * | 1992-05-25 | 1994-06-14 | Sharp Kabushiki Kaisha | Automatic translating machine |
US5331556A (en) * | 1993-06-28 | 1994-07-19 | General Electric Company | Method for natural language data processing using morphological and part-of-speech information |
Cited By (315)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5974557A (en) * | 1994-06-20 | 1999-10-26 | Thomas; C. Douglass | Method and system for performing thermal and power management for a computer |
US6487668B2 (en) * | 1994-06-20 | 2002-11-26 | C. Douglass Thomas | Thermal and power management to computer systems |
US6216235B1 (en) | 1994-06-20 | 2001-04-10 | C. Douglass Thomas | Thermal and power management for computer systems |
US5999896A (en) * | 1996-06-25 | 1999-12-07 | Microsoft Corporation | Method and system for identifying and resolving commonly confused words in a natural language parser |
US6615168B1 (en) * | 1996-07-26 | 2003-09-02 | Sun Microsystems, Inc. | Multilingual agent for use in computer systems |
US5878385A (en) * | 1996-09-16 | 1999-03-02 | Ergo Linguistic Technologies | Method and apparatus for universal parsing of language |
US5841895A (en) * | 1996-10-25 | 1998-11-24 | Pricewaterhousecoopers, Llp | Method for learning local syntactic relationships for use in example-based information-extraction-pattern learning |
US6014616A (en) * | 1996-11-13 | 2000-01-11 | Samsung Electronics Co., Ltd. | Method for monitoring the language used for character generation by an operating system |
US7672829B2 (en) * | 1997-03-04 | 2010-03-02 | Hiroshi Ishikura | Pivot translation method and system |
US20050119874A1 (en) * | 1997-03-04 | 2005-06-02 | Hiroshi Ishikura | Pivot translation method and system |
US6181909B1 (en) * | 1997-07-22 | 2001-01-30 | Educational Testing Service | System and method for computer-based automatic essay scoring |
US6081772A (en) * | 1998-03-26 | 2000-06-27 | International Business Machines Corporation | Proofreading aid based on closed-class vocabulary |
US6810392B1 (en) * | 1998-07-31 | 2004-10-26 | Northrop Grumman Corporation | Method and apparatus for estimating computer software development effort |
US7020326B1 (en) * | 1998-11-13 | 2006-03-28 | Hsu Shin-Yi | System for guiding users to formulate and use object extraction rules |
US6859931B1 (en) | 1999-01-05 | 2005-02-22 | Sri International | Extensible software-based architecture for communication and cooperation within and between communities of distributed agents and distributed objects |
US6523061B1 (en) | 1999-01-05 | 2003-02-18 | Sri International, Inc. | System, method, and article of manufacture for agent-based navigation in a speech-based data navigation system |
US6513063B1 (en) | 1999-01-05 | 2003-01-28 | Sri International | Accessing network-based electronic information through scripted online interfaces using spoken input |
US6742021B1 (en) | 1999-01-05 | 2004-05-25 | Sri International, Inc. | Navigating network-based electronic information using spoken input with multimodal error feedback |
US7036128B1 (en) | 1999-01-05 | 2006-04-25 | Sri International Offices | Using a community of distributed electronic agents to support a highly mobile, ambient computing environment |
US7069560B1 (en) | 1999-01-05 | 2006-06-27 | Sri International | Highly scalable software-based architecture for communication and cooperation among distributed electronic agents |
US6691151B1 (en) | 1999-01-05 | 2004-02-10 | Sri International | Unified messaging methods and systems for communication and cooperation among distributed agents in a computing environment |
US6757718B1 (en) | 1999-01-05 | 2004-06-29 | Sri International | Mobile navigation of network-based electronic information using spoken input |
US6346892B1 (en) * | 1999-05-07 | 2002-02-12 | Honeywell International Inc. | Method and apparatus for aircraft systems management |
US6625608B1 (en) * | 1999-08-24 | 2003-09-23 | International Business Machines Corporation | Method and apparatus for displaying natural language structure |
US8352277B2 (en) | 1999-11-12 | 2013-01-08 | Phoenix Solutions, Inc. | Method of interacting through speech with a web-connected server |
US20080255845A1 (en) * | 1999-11-12 | 2008-10-16 | Bennett Ian M | Speech Based Query System Using Semantic Decoding |
US7729904B2 (en) | 1999-11-12 | 2010-06-01 | Phoenix Solutions, Inc. | Partial speech processing device and method for use in distributed systems |
US7725320B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Internet based speech recognition system with dynamic grammars |
US7725321B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Speech based query system using semantic decoding |
US7725307B2 (en) | 1999-11-12 | 2010-05-25 | Phoenix Solutions, Inc. | Query engine for processing voice based queries including semantic decoding |
US7702508B2 (en) | 1999-11-12 | 2010-04-20 | Phoenix Solutions, Inc. | System and method for natural language processing of query answers |
US7698131B2 (en) | 1999-11-12 | 2010-04-13 | Phoenix Solutions, Inc. | Speech recognition system for client devices having differing computing capabilities |
US9190063B2 (en) | 1999-11-12 | 2015-11-17 | Nuance Communications, Inc. | Multi-language speech recognition system |
US20070094032A1 (en) * | 1999-11-12 | 2007-04-26 | Bennett Ian M | Adjustable resource based speech recognition system |
US7672841B2 (en) | 1999-11-12 | 2010-03-02 | Phoenix Solutions, Inc. | Method for processing speech data for a distributed recognition system |
US20070179789A1 (en) * | 1999-11-12 | 2007-08-02 | Bennett Ian M | Speech Recognition System With Support For Variable Portable Devices |
US7657424B2 (en) | 1999-11-12 | 2010-02-02 | Phoenix Solutions, Inc. | System and method for processing sentence based queries |
US7647225B2 (en) | 1999-11-12 | 2010-01-12 | Phoenix Solutions, Inc. | Adjustable resource based speech recognition system |
US7912702B2 (en) | 1999-11-12 | 2011-03-22 | Phoenix Solutions, Inc. | Statistical language model trained with semantic variants |
US7624007B2 (en) | 1999-11-12 | 2009-11-24 | Phoenix Solutions, Inc. | System and method for natural language processing of sentence based queries |
US7831426B2 (en) | 1999-11-12 | 2010-11-09 | Phoenix Solutions, Inc. | Network based interactive speech recognition system |
US7555431B2 (en) | 1999-11-12 | 2009-06-30 | Phoenix Solutions, Inc. | Method for processing speech using dynamic grammars |
US20040236580A1 (en) * | 1999-11-12 | 2004-11-25 | Bennett Ian M. | Method for processing speech using dynamic grammars |
US20090157401A1 (en) * | 1999-11-12 | 2009-06-18 | Bennett Ian M | Semantic Decoding of User Queries |
US8762152B2 (en) | 1999-11-12 | 2014-06-24 | Nuance Communications, Inc. | Speech recognition system interactive agent |
US20080300878A1 (en) * | 1999-11-12 | 2008-12-04 | Bennett Ian M | Method For Transporting Speech Data For A Distributed Recognition System |
US8229734B2 (en) | 1999-11-12 | 2012-07-24 | Phoenix Solutions, Inc. | Semantic decoding of user queries |
US20040117189A1 (en) * | 1999-11-12 | 2004-06-17 | Bennett Ian M. | Query engine for processing voice based queries including semantic decoding |
US20050080625A1 (en) * | 1999-11-12 | 2005-04-14 | Bennett Ian M. | Distributed real time speech recognition system |
US20050086049A1 (en) * | 1999-11-12 | 2005-04-21 | Bennett Ian M. | System & method for processing sentence based queries |
US20050086046A1 (en) * | 1999-11-12 | 2005-04-21 | Bennett Ian M. | System & method for natural language processing of sentence based queries |
US20070185716A1 (en) * | 1999-11-12 | 2007-08-09 | Bennett Ian M | Internet based speech recognition system with dynamic grammars |
US20080059153A1 (en) * | 1999-11-12 | 2008-03-06 | Bennett Ian M | Natural Language Speech Lattice Containing Semantic Variants |
US20080052077A1 (en) * | 1999-11-12 | 2008-02-28 | Bennett Ian M | Multi-language speech recognition system |
US7873519B2 (en) | 1999-11-12 | 2011-01-18 | Phoenix Solutions, Inc. | Natural language speech lattice containing semantic variants |
US20080052063A1 (en) * | 1999-11-12 | 2008-02-28 | Bennett Ian M | Multi-language speech recognition system |
US20080052078A1 (en) * | 1999-11-12 | 2008-02-28 | Bennett Ian M | Statistical Language Model Trained With Semantic Variants |
US20080021708A1 (en) * | 1999-11-12 | 2008-01-24 | Bennett Ian M | Speech recognition system interactive agent |
US9076448B2 (en) | 1999-11-12 | 2015-07-07 | Nuance Communications, Inc. | Distributed real time speech recognition system |
US7379862B1 (en) * | 1999-11-19 | 2008-05-27 | Microsoft Corporation | Method and apparatus for analyzing and debugging natural language parses |
USRE44326E1 (en) | 2000-06-08 | 2013-06-25 | Promptu Systems Corporation | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US20060080106A1 (en) * | 2000-06-08 | 2006-04-13 | Theodore Calderone | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US20010056350A1 (en) * | 2000-06-08 | 2001-12-27 | Theodore Calderone | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US7047196B2 (en) | 2000-06-08 | 2006-05-16 | Agiletv Corporation | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US7685523B2 (en) | 2000-06-08 | 2010-03-23 | Agiletv Corporation | System and method of voice recognition near a wireline node of network supporting cable television and/or video delivery |
US6941513B2 (en) | 2000-06-15 | 2005-09-06 | Cognisphere, Inc. | System and method for text structuring and text generation |
US6598021B1 (en) * | 2000-07-13 | 2003-07-22 | Craig R. Shambaugh | Method of modifying speech to provide a user selectable dialect |
US6675159B1 (en) | 2000-07-27 | 2004-01-06 | Science Applic Int Corp | Concept-based search and retrieval system |
US7027974B1 (en) | 2000-10-27 | 2006-04-11 | Science Applications International Corporation | Ontology-based parser for natural language processing |
US7627556B2 (en) | 2000-10-30 | 2009-12-01 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US7099860B1 (en) * | 2000-10-30 | 2006-08-29 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US7349895B2 (en) | 2000-10-30 | 2008-03-25 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20050055344A1 (en) * | 2000-10-30 | 2005-03-10 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050114325A1 (en) * | 2000-10-30 | 2005-05-26 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US7499916B2 (en) | 2000-10-30 | 2009-03-03 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US7529732B2 (en) | 2000-10-30 | 2009-05-05 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US20050010553A1 (en) * | 2000-10-30 | 2005-01-13 | Microsoft Corporation | Semi-automatic annotation of multimedia objects |
US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
US6798912B2 (en) * | 2000-12-18 | 2004-09-28 | Koninklijke Philips Electronics N.V. | Apparatus and method of program classification based on syntax of transcript information |
US20020076112A1 (en) * | 2000-12-18 | 2002-06-20 | Philips Electronics North America Corporation | Apparatus and method of program classification based on syntax of transcript information |
US20020129066A1 (en) * | 2000-12-28 | 2002-09-12 | Milward David R. | Computer implemented method for reformatting logically complex clauses in an electronic text-based document |
US6766316B2 (en) | 2001-01-18 | 2004-07-20 | Science Applications International Corporation | Method and system of ranking and clustering for document indexing and retrieval |
US7496561B2 (en) | 2001-01-18 | 2009-02-24 | Science Applications International Corporation | Method and system of ranking and clustering for document indexing and retrieval |
US8095370B2 (en) | 2001-02-16 | 2012-01-10 | Agiletv Corporation | Dual compression voice recordation non-repudiation system |
US20050033581A1 (en) * | 2001-02-16 | 2005-02-10 | Foster Mark J. | Dual compression voice recordation non-repudiation system |
US20030023642A1 (en) * | 2001-07-30 | 2003-01-30 | Spragins James S-B | Method and system for marking writings online |
US20030028368A1 (en) * | 2001-08-01 | 2003-02-06 | Grandy Woodrow W. | Method for entering, recording, distributing and reporting data |
US8909595B2 (en) | 2001-08-01 | 2014-12-09 | T-System, Inc. | Method for entering, recording, distributing and reporting data |
US8898106B2 (en) * | 2001-08-01 | 2014-11-25 | T-System, Inc. | Method for entering, recording, distributing and reporting data |
US20030074188A1 (en) * | 2001-10-12 | 2003-04-17 | Tohgo Murata | Method and apparatus for language instruction |
US20060117039A1 (en) * | 2002-01-07 | 2006-06-01 | Hintz Kenneth J | Lexicon-based new idea detector |
US7823065B2 (en) * | 2002-01-07 | 2010-10-26 | Kenneth James Hintz | Lexicon-based new idea detector |
US8731929B2 (en) | 2002-06-03 | 2014-05-20 | Voicebox Technologies Corporation | Agent architecture for determining meanings of natural language utterances |
US20080319751A1 (en) * | 2002-06-03 | 2008-12-25 | Kennewick Robert A | Systems and methods for responding to natural language speech utterance |
US7809570B2 (en) | 2002-06-03 | 2010-10-05 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20100204986A1 (en) * | 2002-06-03 | 2010-08-12 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US7398209B2 (en) | 2002-06-03 | 2008-07-08 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20100204994A1 (en) * | 2002-06-03 | 2010-08-12 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US8112275B2 (en) | 2002-06-03 | 2012-02-07 | Voicebox Technologies, Inc. | System and method for user-specific speech recognition |
US8015006B2 (en) | 2002-06-03 | 2011-09-06 | Voicebox Technologies, Inc. | Systems and methods for processing natural language speech utterances with context-specific domain agents |
US20040044516A1 (en) * | 2002-06-03 | 2004-03-04 | Kennewick Robert A. | Systems and methods for responding to natural language speech utterance |
US8140327B2 (en) | 2002-06-03 | 2012-03-20 | Voicebox Technologies, Inc. | System and method for filtering and eliminating noise from natural language utterances to improve speech recognition and parsing |
US8155962B2 (en) | 2002-06-03 | 2012-04-10 | Voicebox Technologies, Inc. | Method and system for asynchronously processing natural language utterances |
US20090171664A1 (en) * | 2002-06-03 | 2009-07-02 | Kennewick Robert A | Systems and methods for responding to natural language speech utterance |
US20070265850A1 (en) * | 2002-06-03 | 2007-11-15 | Kennewick Robert A | Systems and methods for responding to natural language speech utterance |
US7502738B2 (en) | 2002-06-03 | 2009-03-10 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20040193420A1 (en) * | 2002-07-15 | 2004-09-30 | Kennewick Robert A. | Mobile systems and methods for responding to natural language speech utterance |
US7693720B2 (en) | 2002-07-15 | 2010-04-06 | Voicebox Technologies, Inc. | Mobile systems and methods for responding to natural language speech utterance |
US9031845B2 (en) | 2002-07-15 | 2015-05-12 | Nuance Communications, Inc. | Mobile systems and methods for responding to natural language speech utterance |
US10923219B2 (en) | 2002-11-28 | 2021-02-16 | Nuance Communications, Inc. | Method to assign word class information |
US20060074668A1 (en) * | 2002-11-28 | 2006-04-06 | Koninklijke Philips Electronics N.V. | Method to assign word class information |
US9996675B2 (en) | 2002-11-28 | 2018-06-12 | Nuance Communications, Inc. | Method to assign word class information |
US10515719B2 (en) | 2002-11-28 | 2019-12-24 | Nuance Communications, Inc. | Method to assign world class information |
US8032358B2 (en) * | 2002-11-28 | 2011-10-04 | Nuance Communications Austria Gmbh | Classifying text via topical analysis, for applications to speech recognition |
US8965753B2 (en) | 2002-11-28 | 2015-02-24 | Nuance Communications, Inc. | Method to assign word class information |
US8612209B2 (en) | 2002-11-28 | 2013-12-17 | Nuance Communications, Inc. | Classifying text via topical analysis, for applications to speech recognition |
US20040167910A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Integrated data products of processes of integrating mixed format data |
US20040167911A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Methods and products for integrating mixed format data including the extraction of relational facts from free text |
US20040167887A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Integration of structured data with relational facts from free text for data mining |
US20040167883A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Methods and systems for providing a service for producing structured data elements from free text sources |
US20040167885A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Data products of processes of extracting role related information from free text sources |
US20050108256A1 (en) * | 2002-12-06 | 2005-05-19 | Attensity Corporation | Visualization of integrated structured and unstructured data |
US20040167884A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Methods and products for producing role related information from free text sources |
US20040215634A1 (en) * | 2002-12-06 | 2004-10-28 | Attensity Corporation | Methods and products for merging codes and notes into an integrated relational database |
US20040167870A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Systems and methods for providing a mixed data integration service |
US20040167908A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Integration of structured data with free text for data mining |
US20040167886A1 (en) * | 2002-12-06 | 2004-08-26 | Attensity Corporation | Production of role related information from free text sources utilizing thematic caseframes |
US20040193399A1 (en) * | 2003-03-31 | 2004-09-30 | Microsoft Corporation | System and method for word analysis |
US8200477B2 (en) * | 2003-10-22 | 2012-06-12 | International Business Machines Corporation | Method and system for extracting opinions from text documents |
US20050091038A1 (en) * | 2003-10-22 | 2005-04-28 | Jeonghee Yi | Method and system for extracting opinions from text documents |
US20080228467A1 (en) * | 2004-01-06 | 2008-09-18 | Neuric Technologies, Llc | Natural language parsing method to provide conceptual flow |
US9213936B2 (en) | 2004-01-06 | 2015-12-15 | Neuric, Llc | Electronic brain model with neuron tables |
US8725493B2 (en) * | 2004-01-06 | 2014-05-13 | Neuric Llc | Natural language parsing method to provide conceptual flow |
US8296304B2 (en) | 2004-01-26 | 2012-10-23 | International Business Machines Corporation | Method, system, and program for handling redirects in a search engine |
US8285724B2 (en) | 2004-01-26 | 2012-10-09 | International Business Machines Corporation | System and program for handling anchor text |
US7424467B2 (en) | 2004-01-26 | 2008-09-09 | International Business Machines Corporation | Architecture for an indexer with fixed width sort and variable width sort |
US20050165781A1 (en) * | 2004-01-26 | 2005-07-28 | Reiner Kraft | Method, system, and program for handling anchor text |
US20050165838A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Architecture for an indexer |
US7499913B2 (en) | 2004-01-26 | 2009-03-03 | International Business Machines Corporation | Method for handling anchor text |
US20050165718A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Pipelined architecture for global analysis and index building |
US7783626B2 (en) | 2004-01-26 | 2010-08-24 | International Business Machines Corporation | Pipelined architecture for global analysis and index building |
US20050165800A1 (en) * | 2004-01-26 | 2005-07-28 | Fontoura Marcus F. | Method, system, and program for handling redirects in a search engine |
US7743060B2 (en) | 2004-01-26 | 2010-06-22 | International Business Machines Corporation | Architecture for an indexer |
US7293005B2 (en) | 2004-01-26 | 2007-11-06 | International Business Machines Corporation | Pipelined architecture for global analysis and index building |
US20090083270A1 (en) * | 2004-01-26 | 2009-03-26 | International Business Machines Corporation | System and program for handling anchor text |
US20070282829A1 (en) * | 2004-01-26 | 2007-12-06 | International Business Machines Corporation | Pipelined architecture for global analysis and index building |
US8768969B2 (en) * | 2004-07-09 | 2014-07-01 | Nuance Communications, Inc. | Method and system for efficient representation, manipulation, communication, and search of hierarchical composite named entities |
US20060010138A1 (en) * | 2004-07-09 | 2006-01-12 | International Business Machines Corporation | Method and system for efficient representation, manipulation, communication, and search of hierarchical composite named entities |
US7461064B2 (en) | 2004-09-24 | 2008-12-02 | International Buiness Machines Corporation | Method for searching documents for ranges of numeric values |
US20080301130A1 (en) * | 2004-09-24 | 2008-12-04 | International Business Machines Corporation | Method, system and article of manufacture for searching documents for ranges of numeric values |
US8271498B2 (en) | 2004-09-24 | 2012-09-18 | International Business Machines Corporation | Searching documents for ranges of numeric values |
US20060074962A1 (en) * | 2004-09-24 | 2006-04-06 | Fontoura Marcus F | Method, system, and program for searching documents for ranges of numeric values |
US8655888B2 (en) | 2004-09-24 | 2014-02-18 | International Business Machines Corporation | Searching documents for ranges of numeric values |
US20080294634A1 (en) * | 2004-09-24 | 2008-11-27 | International Business Machines Corporation | System and article of manufacture for searching documents for ranges of numeric values |
US8346759B2 (en) | 2004-09-24 | 2013-01-01 | International Business Machines Corporation | Searching documents for ranges of numeric values |
US8473449B2 (en) | 2005-01-06 | 2013-06-25 | Neuric Technologies, Llc | Process of dialogue and discussion |
US20100185437A1 (en) * | 2005-01-06 | 2010-07-22 | Neuric Technologies, Llc | Process of dialogue and discussion |
US7574348B2 (en) | 2005-07-08 | 2009-08-11 | Microsoft Corporation | Processing collocation mistakes in documents |
US20070010992A1 (en) * | 2005-07-08 | 2007-01-11 | Microsoft Corporation | Processing collocation mistakes in documents |
US8417693B2 (en) | 2005-07-14 | 2013-04-09 | International Business Machines Corporation | Enforcing native access control to indexed documents |
US20070016583A1 (en) * | 2005-07-14 | 2007-01-18 | Ronny Lempel | Enforcing native access control to indexed documents |
US20100063795A1 (en) * | 2005-08-04 | 2010-03-11 | Nec Corporation | Data processing device, data processing method, and data processing program |
US8775158B2 (en) * | 2005-08-04 | 2014-07-08 | Nec Corporation | Data processing device, data processing method, and data processing program |
US7917367B2 (en) | 2005-08-05 | 2011-03-29 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US8326634B2 (en) | 2005-08-05 | 2012-12-04 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20100057443A1 (en) * | 2005-08-05 | 2010-03-04 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20070033005A1 (en) * | 2005-08-05 | 2007-02-08 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US8849670B2 (en) | 2005-08-05 | 2014-09-30 | Voicebox Technologies Corporation | Systems and methods for responding to natural language speech utterance |
US9263039B2 (en) | 2005-08-05 | 2016-02-16 | Nuance Communications, Inc. | Systems and methods for responding to natural language speech utterance |
US7640160B2 (en) | 2005-08-05 | 2009-12-29 | Voicebox Technologies, Inc. | Systems and methods for responding to natural language speech utterance |
US20110131036A1 (en) * | 2005-08-10 | 2011-06-02 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US9626959B2 (en) | 2005-08-10 | 2017-04-18 | Nuance Communications, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US8620659B2 (en) | 2005-08-10 | 2013-12-31 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US20070038436A1 (en) * | 2005-08-10 | 2007-02-15 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US7620549B2 (en) | 2005-08-10 | 2009-11-17 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US20100023320A1 (en) * | 2005-08-10 | 2010-01-28 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition in conversational speech |
US8332224B2 (en) | 2005-08-10 | 2012-12-11 | Voicebox Technologies, Inc. | System and method of supporting adaptive misrecognition conversational speech |
US7949529B2 (en) | 2005-08-29 | 2011-05-24 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US20110231182A1 (en) * | 2005-08-29 | 2011-09-22 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US8195468B2 (en) | 2005-08-29 | 2012-06-05 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US8447607B2 (en) | 2005-08-29 | 2013-05-21 | Voicebox Technologies, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US8849652B2 (en) | 2005-08-29 | 2014-09-30 | Voicebox Technologies Corporation | Mobile systems and methods of supporting natural language human-machine interactions |
US9495957B2 (en) | 2005-08-29 | 2016-11-15 | Nuance Communications, Inc. | Mobile systems and methods of supporting natural language human-machine interactions |
US20100049514A1 (en) * | 2005-08-31 | 2010-02-25 | Voicebox Technologies, Inc. | Dynamic speech sharpening |
US7634409B2 (en) | 2005-08-31 | 2009-12-15 | Voicebox Technologies, Inc. | Dynamic speech sharpening |
US8150694B2 (en) | 2005-08-31 | 2012-04-03 | Voicebox Technologies, Inc. | System and method for providing an acoustic grammar to dynamically sharpen speech interpretation |
US8069046B2 (en) | 2005-08-31 | 2011-11-29 | Voicebox Technologies, Inc. | Dynamic speech sharpening |
US20070055525A1 (en) * | 2005-08-31 | 2007-03-08 | Kennewick Robert A | Dynamic speech sharpening |
US7983917B2 (en) | 2005-08-31 | 2011-07-19 | Voicebox Technologies, Inc. | Dynamic speech sharpening |
US20070073745A1 (en) * | 2005-09-23 | 2007-03-29 | Applied Linguistics, Llc | Similarity metric for semantic profiling |
US20070073678A1 (en) * | 2005-09-23 | 2007-03-29 | Applied Linguistics, Llc | Semantic document profiling |
US20070213985A1 (en) * | 2006-03-13 | 2007-09-13 | Corwin Daniel W | Self-Annotating Identifiers |
US7962328B2 (en) * | 2006-03-13 | 2011-06-14 | Lexikos Corporation | Method and apparatus for generating a compact data structure to identify the meaning of a symbol |
US9047275B2 (en) | 2006-10-10 | 2015-06-02 | Abbyy Infopoisk Llc | Methods and systems for alignment of parallel text corpora |
US8145473B2 (en) | 2006-10-10 | 2012-03-27 | Abbyy Software Ltd. | Deep model statistics method for machine translation |
US9633005B2 (en) | 2006-10-10 | 2017-04-25 | Abbyy Infopoisk Llc | Exhaustive automatic processing of textual information |
US8442810B2 (en) | 2006-10-10 | 2013-05-14 | Abbyy Software Ltd. | Deep model statistics method for machine translation |
US9645993B2 (en) | 2006-10-10 | 2017-05-09 | Abbyy Infopoisk Llc | Method and system for semantic searching |
US9323747B2 (en) | 2006-10-10 | 2016-04-26 | Abbyy Infopoisk Llc | Deep model statistics method for machine translation |
US9235573B2 (en) | 2006-10-10 | 2016-01-12 | Abbyy Infopoisk Llc | Universal difference measure |
US9817818B2 (en) | 2006-10-10 | 2017-11-14 | Abbyy Production Llc | Method and system for translating sentence between languages based on semantic structure of the sentence |
US8412513B2 (en) | 2006-10-10 | 2013-04-02 | Abbyy Software Ltd. | Deep model statistics method for machine translation |
US9984071B2 (en) | 2006-10-10 | 2018-05-29 | Abbyy Production Llc | Language ambiguity detection of text |
US8195447B2 (en) | 2006-10-10 | 2012-06-05 | Abbyy Software Ltd. | Translating sentences between languages using language-independent semantic structures and ratings of syntactic constructions |
US8548795B2 (en) | 2006-10-10 | 2013-10-01 | Abbyy Software Ltd. | Method for translating documents from one language into another using a database of translations, a terminology dictionary, a translation dictionary, and a machine translation system |
US20080086299A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between languages |
US20080086298A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between langauges |
US8918309B2 (en) | 2006-10-10 | 2014-12-23 | Abbyy Infopoisk Llc | Deep model statistics method for machine translation |
US20080086300A1 (en) * | 2006-10-10 | 2008-04-10 | Anisimovich Konstantin | Method and system for translating sentences between languages |
US8892418B2 (en) | 2006-10-10 | 2014-11-18 | Abbyy Infopoisk Llc | Translating sentences between languages |
US8214199B2 (en) | 2006-10-10 | 2012-07-03 | Abbyy Software, Ltd. | Systems for translating sentences between languages using language-independent semantic structures and ratings of syntactic constructions |
US8805676B2 (en) | 2006-10-10 | 2014-08-12 | Abbyy Infopoisk Llc | Deep model statistics method for machine translation |
US20090182549A1 (en) * | 2006-10-10 | 2009-07-16 | Konstantin Anisimovich | Deep Model Statistics Method for Machine Translation |
US20090070099A1 (en) * | 2006-10-10 | 2009-03-12 | Konstantin Anisimovich | Method for translating documents from one language into another using a database of translations, a terminology dictionary, a translation dictionary, and a machine translation system |
US20080091406A1 (en) * | 2006-10-16 | 2008-04-17 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US10510341B1 (en) | 2006-10-16 | 2019-12-17 | Vb Assets, Llc | System and method for a cooperative conversational voice user interface |
US10515628B2 (en) | 2006-10-16 | 2019-12-24 | Vb Assets, Llc | System and method for a cooperative conversational voice user interface |
US8515765B2 (en) | 2006-10-16 | 2013-08-20 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US11222626B2 (en) | 2006-10-16 | 2022-01-11 | Vb Assets, Llc | System and method for a cooperative conversational voice user interface |
US8073681B2 (en) | 2006-10-16 | 2011-12-06 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US10755699B2 (en) | 2006-10-16 | 2020-08-25 | Vb Assets, Llc | System and method for a cooperative conversational voice user interface |
US9015049B2 (en) | 2006-10-16 | 2015-04-21 | Voicebox Technologies Corporation | System and method for a cooperative conversational voice user interface |
US10297249B2 (en) | 2006-10-16 | 2019-05-21 | Vb Assets, Llc | System and method for a cooperative conversational voice user interface |
US10134060B2 (en) | 2007-02-06 | 2018-11-20 | Vb Assets, Llc | System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements |
US11080758B2 (en) | 2007-02-06 | 2021-08-03 | Vb Assets, Llc | System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements |
US12236456B2 (en) | 2007-02-06 | 2025-02-25 | Vb Assets, Llc | System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements |
US7818176B2 (en) | 2007-02-06 | 2010-10-19 | Voicebox Technologies, Inc. | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US8527274B2 (en) | 2007-02-06 | 2013-09-03 | Voicebox Technologies, Inc. | System and method for delivering targeted advertisements and tracking advertisement interactions in voice recognition contexts |
US9406078B2 (en) | 2007-02-06 | 2016-08-02 | Voicebox Technologies Corporation | System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements |
US20080189110A1 (en) * | 2007-02-06 | 2008-08-07 | Tom Freeman | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US9269097B2 (en) | 2007-02-06 | 2016-02-23 | Voicebox Technologies Corporation | System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements |
US8886536B2 (en) | 2007-02-06 | 2014-11-11 | Voicebox Technologies Corporation | System and method for delivering targeted advertisements and tracking advertisement interactions in voice recognition contexts |
US20100299142A1 (en) * | 2007-02-06 | 2010-11-25 | Voicebox Technologies, Inc. | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US8145489B2 (en) | 2007-02-06 | 2012-03-27 | Voicebox Technologies, Inc. | System and method for selecting and presenting advertisements based on natural language processing of voice-based input |
US9772998B2 (en) | 2007-03-22 | 2017-09-26 | Abbyy Production Llc | Indicating and correcting errors in machine translation systems |
US8959011B2 (en) | 2007-03-22 | 2015-02-17 | Abbyy Infopoisk Llc | Indicating and correcting errors in machine translation systems |
US9239826B2 (en) | 2007-06-27 | 2016-01-19 | Abbyy Infopoisk Llc | Method and system for generating new entries in natural language dictionary |
US8983839B2 (en) | 2007-12-11 | 2015-03-17 | Voicebox Technologies Corporation | System and method for dynamically generating a recognition grammar in an integrated voice navigation services environment |
US10347248B2 (en) | 2007-12-11 | 2019-07-09 | Voicebox Technologies Corporation | System and method for providing in-vehicle services via a natural language voice user interface |
US8326627B2 (en) | 2007-12-11 | 2012-12-04 | Voicebox Technologies, Inc. | System and method for dynamically generating a recognition grammar in an integrated voice navigation services environment |
US20090150156A1 (en) * | 2007-12-11 | 2009-06-11 | Kennewick Michael R | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US8140335B2 (en) | 2007-12-11 | 2012-03-20 | Voicebox Technologies, Inc. | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US8719026B2 (en) | 2007-12-11 | 2014-05-06 | Voicebox Technologies Corporation | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US8370147B2 (en) | 2007-12-11 | 2013-02-05 | Voicebox Technologies, Inc. | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US8452598B2 (en) | 2007-12-11 | 2013-05-28 | Voicebox Technologies, Inc. | System and method for providing advertisements in an integrated voice navigation services environment |
US9620113B2 (en) | 2007-12-11 | 2017-04-11 | Voicebox Technologies Corporation | System and method for providing a natural language voice user interface |
US20090299745A1 (en) * | 2008-05-27 | 2009-12-03 | Kennewick Robert A | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US9305548B2 (en) | 2008-05-27 | 2016-04-05 | Voicebox Technologies Corporation | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US10089984B2 (en) | 2008-05-27 | 2018-10-02 | Vb Assets, Llc | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US8589161B2 (en) | 2008-05-27 | 2013-11-19 | Voicebox Technologies, Inc. | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US9711143B2 (en) | 2008-05-27 | 2017-07-18 | Voicebox Technologies Corporation | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US10553216B2 (en) | 2008-05-27 | 2020-02-04 | Oracle International Corporation | System and method for an integrated, multi-modal, multi-device natural language voice services environment |
US8473279B2 (en) * | 2008-05-30 | 2013-06-25 | Eiman Al-Shammari | Lemmatizing, stemming, and query expansion method and system |
US20100082333A1 (en) * | 2008-05-30 | 2010-04-01 | Eiman Tamah Al-Shammari | Lemmatizing, stemming, and query expansion method and system |
US9262409B2 (en) | 2008-08-06 | 2016-02-16 | Abbyy Infopoisk Llc | Translation of a selected text fragment of a screen |
US9323854B2 (en) * | 2008-12-19 | 2016-04-26 | Intel Corporation | Method, apparatus and system for location assisted translation |
US20100161311A1 (en) * | 2008-12-19 | 2010-06-24 | Massuh Lucas A | Method, apparatus and system for location assisted translation |
US9105266B2 (en) | 2009-02-20 | 2015-08-11 | Voicebox Technologies Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US10553213B2 (en) | 2009-02-20 | 2020-02-04 | Oracle International Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US9953649B2 (en) | 2009-02-20 | 2018-04-24 | Voicebox Technologies Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US8719009B2 (en) | 2009-02-20 | 2014-05-06 | Voicebox Technologies Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US9570070B2 (en) | 2009-02-20 | 2017-02-14 | Voicebox Technologies Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US8326637B2 (en) | 2009-02-20 | 2012-12-04 | Voicebox Technologies, Inc. | System and method for processing multi-modal device interactions in a natural language voice services environment |
US20100217604A1 (en) * | 2009-02-20 | 2010-08-26 | Voicebox Technologies, Inc. | System and method for processing multi-modal device interactions in a natural language voice services environment |
US8738380B2 (en) | 2009-02-20 | 2014-05-27 | Voicebox Technologies Corporation | System and method for processing multi-modal device interactions in a natural language voice services environment |
US20110060733A1 (en) * | 2009-09-04 | 2011-03-10 | Alibaba Group Holding Limited | Information retrieval based on semantic patterns of queries |
US8799275B2 (en) * | 2009-09-04 | 2014-08-05 | Alibaba Group Holding Limited | Information retrieval based on semantic patterns of queries |
US20110112827A1 (en) * | 2009-11-10 | 2011-05-12 | Kennewick Robert A | System and method for hybrid processing in a natural language voice services environment |
US9502025B2 (en) | 2009-11-10 | 2016-11-22 | Voicebox Technologies Corporation | System and method for providing a natural language content dedication service |
US9171541B2 (en) | 2009-11-10 | 2015-10-27 | Voicebox Technologies Corporation | System and method for hybrid processing in a natural language voice services environment |
US20110307252A1 (en) * | 2010-06-15 | 2011-12-15 | Microsoft Corporation | Using Utterance Classification in Telephony and Speech Recognition Applications |
US9928296B2 (en) * | 2010-12-16 | 2018-03-27 | Microsoft Technology Licensing, Llc | Search lexicon expansion |
US20120158703A1 (en) * | 2010-12-16 | 2012-06-21 | Microsoft Corporation | Search lexicon expansion |
US20120281011A1 (en) * | 2011-03-07 | 2012-11-08 | Oliver Reichenstein | Method of displaying text in a text editor |
US9992745B2 (en) | 2011-11-01 | 2018-06-05 | Qualcomm Incorporated | Extraction and analysis of buffered audio data using multiple codec rates each greater than a low-power processor rate |
US10381007B2 (en) | 2011-12-07 | 2019-08-13 | Qualcomm Incorporated | Low power integrated circuit to analyze a digitized audio stream |
US11810569B2 (en) | 2011-12-07 | 2023-11-07 | Qualcomm Incorporated | Low power integrated circuit to analyze a digitized audio stream |
US11069360B2 (en) | 2011-12-07 | 2021-07-20 | Qualcomm Incorporated | Low power integrated circuit to analyze a digitized audio stream |
US9082403B2 (en) | 2011-12-15 | 2015-07-14 | Microsoft Technology Licensing, Llc | Spoken utterance classification training for a speech recognition system |
US8989485B2 (en) | 2012-04-27 | 2015-03-24 | Abbyy Development Llc | Detecting a junction in a text line of CJK characters |
US8971630B2 (en) | 2012-04-27 | 2015-03-03 | Abbyy Development Llc | Fast CJK character recognition |
CN105144149A (en) * | 2013-05-29 | 2015-12-09 | 国立研究开发法人情报通信研究机构 | Translation word order information output device, translation word order information output method, and recording medium |
CN105144149B (en) * | 2013-05-29 | 2017-10-27 | 国立研究开发法人情报通信研究机构 | Translate word order information output apparatus, translation word order information output method and storage medium |
US10922614B2 (en) * | 2013-10-15 | 2021-02-16 | Tata Consultancy Services Limited | Converting procedural text to an actionable knowledge form |
US20150142825A1 (en) * | 2013-11-15 | 2015-05-21 | Tata Consultancy Services Limited | Converting procedural text to an actionable knowledge form |
US9740682B2 (en) | 2013-12-19 | 2017-08-22 | Abbyy Infopoisk Llc | Semantic disambiguation using a statistical analysis |
US9626353B2 (en) | 2014-01-15 | 2017-04-18 | Abbyy Infopoisk Llc | Arc filtering in a syntactic graph |
US11093688B2 (en) | 2014-04-25 | 2021-08-17 | Mayo Foundation For Medical Education And Research | Enhancing reading accuracy, efficiency and retention |
US10515138B2 (en) | 2014-04-25 | 2019-12-24 | Mayo Foundation For Medical Education And Research | Enhancing reading accuracy, efficiency and retention |
US11531804B2 (en) | 2014-04-25 | 2022-12-20 | Mayo Foundation For Medical Education And Research | Enhancing reading accuracy, efficiency and retention |
US9858506B2 (en) | 2014-09-02 | 2018-01-02 | Abbyy Development Llc | Methods and systems for processing of images of mathematical expressions |
US10430863B2 (en) | 2014-09-16 | 2019-10-01 | Vb Assets, Llc | Voice commerce |
US9626703B2 (en) | 2014-09-16 | 2017-04-18 | Voicebox Technologies Corporation | Voice commerce |
US11087385B2 (en) | 2014-09-16 | 2021-08-10 | Vb Assets, Llc | Voice commerce |
US9898459B2 (en) | 2014-09-16 | 2018-02-20 | Voicebox Technologies Corporation | Integration of domain information into state transitions of a finite state transducer for natural language processing |
US10216725B2 (en) | 2014-09-16 | 2019-02-26 | Voicebox Technologies Corporation | Integration of domain information into state transitions of a finite state transducer for natural language processing |
US9747896B2 (en) | 2014-10-15 | 2017-08-29 | Voicebox Technologies Corporation | System and method for providing follow-up responses to prior natural language inputs of a user |
US10229673B2 (en) | 2014-10-15 | 2019-03-12 | Voicebox Technologies Corporation | System and method for providing follow-up responses to prior natural language inputs of a user |
US10431214B2 (en) | 2014-11-26 | 2019-10-01 | Voicebox Technologies Corporation | System and method of determining a domain and/or an action related to a natural language input |
US9626358B2 (en) | 2014-11-26 | 2017-04-18 | Abbyy Infopoisk Llc | Creating ontologies by analyzing natural language texts |
US10614799B2 (en) | 2014-11-26 | 2020-04-07 | Voicebox Technologies Corporation | System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance |
TWI610185B (en) * | 2014-12-22 | 2018-01-01 | 晨星半導體股份有限公司 | Related information displaying method and electronic device that can automatically display related information |
US10657669B2 (en) | 2015-03-25 | 2020-05-19 | Beijing Kuangshi Technology Co., Ltd. | Determination of a geographical location of a user |
WO2016149918A1 (en) * | 2015-03-25 | 2016-09-29 | 北京旷视科技有限公司 | Determining of geographical position of user |
US9710450B2 (en) | 2015-04-10 | 2017-07-18 | International Business Machines Corporation | Recombining incorrectly separated tokens in natural language processing |
US10503769B2 (en) * | 2015-07-06 | 2019-12-10 | Rima Ghannam | System for natural language understanding |
US20170011119A1 (en) * | 2015-07-06 | 2017-01-12 | Rima Ghannam | System for Natural Language Understanding |
US9977826B2 (en) | 2015-10-21 | 2018-05-22 | Cloudera, Inc. | Computerized method of generating and analytically evaluating multiple instances of natural language-generated text |
US10331784B2 (en) | 2016-07-29 | 2019-06-25 | Voicebox Technologies Corporation | System and method of disambiguating natural language processing requests |
US10671801B2 (en) * | 2017-02-28 | 2020-06-02 | Microsoft Technology Licensing, Llc | Markup code generator |
US10885266B2 (en) | 2017-10-10 | 2021-01-05 | Adobe Inc. | Preserving semantic information in document conversion via color codes |
US10521497B2 (en) * | 2017-10-10 | 2019-12-31 | Adobe Inc. | Maintaining semantic information in document conversion |
US10592738B2 (en) | 2017-12-01 | 2020-03-17 | International Business Machines Corporation | Cognitive document image digitalization |
CN114282530A (en) * | 2021-12-24 | 2022-04-05 | 厦门大学 | A sentiment analysis method for complex sentences triggered by grammatical structure and connection information |
CN114282530B (en) * | 2021-12-24 | 2024-06-07 | 厦门大学 | Complex sentence emotion analysis method based on grammar structure and connection information trigger |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5721938A (en) | Method and device for parsing and analyzing natural language sentences and text | |
US20220198135A1 (en) | Systems and methods for semantic paraphrasing | |
US11250842B2 (en) | Multi-dimensional parsing method and system for natural language processing | |
RU2136038C1 (en) | Computer system and method for preparing texts in source language and their translation into foreign languages | |
Grosz et al. | TEAM: An experiment in the design of transportable natural-language interfaces | |
US5083268A (en) | System and method for parsing natural language by unifying lexical features of words | |
US6269189B1 (en) | Finding selected character strings in text and providing information relating to the selected character strings | |
US6139201A (en) | Integrated authoring and translation system | |
US5146406A (en) | Computer method for identifying predicate-argument structures in natural language text | |
AU753108B2 (en) | Data input and retrieval apparatus | |
Van Halteren et al. | Linguistic Exploitation of Syntactic Databases: The Use of the Nijmegen LDB Program | |
JP2012520528A (en) | System and method for automatic semantic labeling of natural language text | |
Wauchope | Eucalyptus: Integrating natural language input with a graphical user interface | |
Li et al. | Natural language interfaces to databases | |
JP2997469B2 (en) | Natural language understanding method and information retrieval device | |
Hanane et al. | A model of a generic Arabic language interface for multimodel database | |
Narita | Constructing a Tagged EJ Parallel Corpus for Assisting Japanese Software Engineers in Writing English Abstracts. | |
Monte-Serrat et al. | NEW UNIVERSAL LABELING STRATEGY FOR MEANING REPRESENTATION IN NLP. | |
Nijholt | Linguistic engineering: a survey | |
Balcha et al. | Design and Development of Sentence Parser for Afan Oromo Language | |
Reitter et al. | Hybrid natural language processing in a customer-care environment | |
Qabbaah et al. | Using Visual Analytics and K-Means Clustering for Monetising Logistics Data, a Case Study with Multiple E-Commerce Companies | |
Daybelge | Improving the precision of example-based machine translation by learning from user feedback | |
Sumam et al. | Design and Development of an Adaptable Frame-Based System for Dravidian Language Processing | |
Hornick et al. | A natural language query system for Hubble Space Telescope proposal selection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
REMI | Maintenance fee reminder mailed | ||
FPAY | Fee payment |
Year of fee payment: 8 |
|
SULP | Surcharge for late payment |
Year of fee payment: 7 |
|
REMI | Maintenance fee reminder mailed | ||
FPAY | Fee payment |
Year of fee payment: 12 |
|
SULP | Surcharge for late payment |
Year of fee payment: 11 |