1. Review. 1.1.1. Introduction. 1.1.1. Background, target. 1.1.1.1. Language is the fact that thinking is an important tool for interpersonal communication. Knowledge in the form of language and text in human history, accounting for more than 80% of the total knowledge. As far as the computer is applied, it is based on only 10% of the mathematical calculation. It is used for less than 5% of the process control, and the remaining 85% of the information processing is used for language text. Under such social needs, natural language understanding as a high-level important direction of language information processing technology, has always been one of the core topics of the artificial intelligence. 2. Since the creation and use of natural languages are human high intelligent performance, the research on natural language is also helps to uncover the mystery of human intelligence, deepen our understanding of language skills and thinking. .1.1.1.2. What is computational linguistics refers to such a discipline, which analyzes, processing natural languages by establishing a formal mathematical model, and implements a natural language and implemented in computer. And the process of processing, thereby achieving the purpose of simulating the human and even the language of the machine. Computational Linguistics (Computational Linguistics), sometimes called quantitative linguistics (Quantitative Linguistics), mathematical linguistics (Mathematical Linguistics), natural language understanding (Natural Language Understanding), NLP (Natural Language Processing), human language technology (Human Language TECHNOLOGY). . 1.1.1.3. Tulex test In the field of artificial intelligence, or language information processing, the language information processing is generally believed to use the famous 1950-descriptive Test to judge whether the computer "understands" has a natural language. . .1.1.1.3.1. T 图 模 游戏) l: Male Trial, Female Try, Observer, 3 in 3 different rooms, the room number is X, Y, OL rules: Observer Communicate with the electrically dozen typewriters and the subjects, male quilt deceived observers, female subjects to help observers. l Objective: The observer should judge the gender of the trial in the X room.
.1.1.1.3.2. Turing test L Scene: Trial, computer, observer 3 in 3 different rooms, the room number is X, Y, OL rules: observers "Some Methods "and the testist and computer communication computer deceived the observer, the test person helps the observer L target: the observer should judge the subject in that room. 1.1.1.3.3. Full TURING TEST L Scenario: Trial object (people or computer), observer, observer can see the subject L rules: Observer can communicate with the subject to the subject: The observer is to judge that the subject is a person or a computer. 1.1.1.3.4. Reference 1. AM TURING, Computing Machinery and Intelligence, http://cogsci.ucsd.edu/~asaygin/tt/ttest.html connection http://www.oxy.edu/departments/cog-sci/courses/1998/cs101/ Texts / computing-machinery.html 2. Cao Shugen, "AI History and Problem", the Chinese Academy of Sciences calculates 3. Roland Hausser, Springer, 1999.1.1.2. Research history. 1.1.2.1. In the 1950s, NLP began in the United States early in the 1950s. At that time, the United States was afraid to defeat in the space competition. Technical literature, so that developing machine translation systems, especially Russian-English machine translation systems, the practice is to use the words to translate. Since the cost is high, the efficiency is low, and the financial support is gradually annoyed. .1.1.2.2. Natural language understanding of the 1990s, mostly there is no real grammatical analysis, mainly relying on keyword matching techniques to identify the meaning of the input sentence. In these systems, the designer stores a large number of modes containing certain keywords in advance, each mode corresponds to one or more interpretation (also called response). The system matches the current input sentence one by one, and once the match is successfully explained, this sentence is immediately explained, and no longer consider what impact on the meaning of the ingredients that do not belong to keywords. Sirsir (Semantic Information Retrieval) was 1968 B. Raphael is done, this is part of his work in the University of Massachusetts Institute of Technology. The system is programmed in Lisp language. This is a prototype that understands the machine, because it can remember the user through English, and then answers the questions raised by users by interpretation of these facts. SIR has an ability to accept a restricted subset of English, which matches the input sentences with the following types of 24 keyword patterns: * Is ** is part of * is * *? How much * does * have? What is the the the * Of *? When the symbol "*" matches one of the nouns in the input sentence, the noun is allowed to have a modifier such as A, THE, EVERY, EACH, etc., quantifiers or numbers. Whenever matching a mode, the corresponding action will be triggered in the program. STUDENT1968 Database & Sli Duo Duo Research Bobrow completed another mode-matched natural language understanding system Studen Ding. The system can understand and solve the middle school generation. Eliza 1968, J. Weizenbaum designed in the US MIT, perhaps these most famous natural language systems based on mode matching. The system simulates a psychotherapy doctor (machine) Talk to the same patient (user).
TGNOAM Chomsky created the Generative Transformational Grammar. Start using syntax analysis in machine translation. .1.1.2.3. After the 1970s, a batch of natural language understanding systems with syntax-semantic analysis technology stood out, in terms of the depth and difficulty in language analysis, have a great progress than the early system. The representative of this period is Lunar, SHRDLU and Margie System. Lunarlunar is the first person who allows the use of ordinary English with computer database dialogue, which is W. BBN, US BBN, 1972. Woods is responsible for design. The system is used to assist geologists to find, compare and evaluate the chemical analysis of the moon rocks and soil specimens brought back by Apollo -11 spacecraft. The SHRDLU SHRDLU system was designed in 1972, which is his Ph.D. research in the US MIT. SHRDLU is a natural language understanding system for English conversations in "Building Block". The system simulates a robotic arm capable of manipulating some toys on the table, and users use people - machine dialogue to command robots to knead the block blocks, the system gives an answer and displays the corresponding scenarios of the site. This system is to explain that it is possible to make computer understanding of language; MEANING Analysis, response generation, and lnference on eng1ish is R. Schank and its students have established a system in the artificial intelligence laboratory of Stanford University, USA, in order to provide a sense of intuitive model of natural language understanding. .1.1.2.4. The biggest feature of the natural language understanding system in the 1980s in the 1980s is practical and engineering. Its important logo is a batch of commercial natural language people ---- machine interface and machine translation system appeared in the international market. The famous English people - machine interface system produced by American Artificial Intelligence Co., Ltd. Intellect, Frei Company produced by Frey - the ASK interface developed by the US California Institute; the European Community is Based on the Motor Translation System, the University of Georgetown, the University of Georgetown, successfully conducted a machine translation of the English, Mr., Germany, West, Italian, and Portuguese. System TAUM-Mete0, Japan Fujitsu company developed Atlas English Japan, Japan Hitachi, Japan Hitachi, Japan, Japan, Japan, Japanese Translation System, etc. During the "Seventh Five-Year" period, the "translation star" developed by China Software Corporation is also an example of this. Corpus Linguistics "Corpus Linguistics" is a new branch discipline of a computational linguistic in the 1980s. It studies the collection, storage, retrieval, and statistics of machine readable natural language text. Syntax labeling, syntax semantic analysis, and the use of the above-mentioned function in language quantitative analysis, dictionary compilation, work style analysis, natural language understanding, and machine translation. " Corpus Linguistics began to rise. First, it complies with the needs of large-scale real text processing, and proposes new ideas based on computer-speaking basis and new ideas for natural language processing. This school insists that the true source of linguistics is a large-scale living corpus. The task of calculating the linguist workers is to automatically obtain various knowledge required to understand language from the large-scale spending, they must objectively The ground is not subjectively described in the language facts of the inventory. .1.1.2.5. In August 1990, at the 13th International Computational Linguistic Conference held in Helsinki, the organizer of the General Assembly put forward the strategic goal of handling large-scale real text, and organized before the meeting. " The role of large-scale corpus in the construction of natural language systems "," Dictionary knowledge and representation "and" electronic dictionary ", which preacted a new historical stage of language information processing.
.1.1.2.6. 21.1.2.7. 21st century. 1.1.2.8. Reference 1) Shijiyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principles", Tsinghua University Press 2) Chris Manning and Hinrich Schutze, Foundations Of Statistical Natural Language Processing, http://www-nlp.stanford.edu/fsnlp/3) Weighing, "Based on Corpus and Contemporary Natural Language Processing Technology", http: // www. Icl.pku.edu.cn/research/papers/chinese/collection-2/zqlw6.htm .1.1.3. Research content. 1.1.3.1. From the calculated perspective to study the nature of the language so-called language from the calculated perspective The nature is to present the understanding of the structure of the language with accurate, formalized, calculated manner, rather than in the statement of the language of the language, as in other linguistics. Expression form. .1.1.3.2. As a corresponding algorithm for calculation objects, the language is used to study the corresponding algorithm to study the language as a computational object. It is a process of studying how to handle the language object (mainly natural language). Objects, of course, may also be a formal language object), including a language disconnect (such as phrase, sentence or chapter) identification, the structure and meaning of the language disconnection (natural language understanding), and how to generate a language Disclosure to express the meaning of determination (natural language generation), and so on. 1.1.4. Different levels of language analysis. 1.1.4.1. Language-constituenation level. 1.1.4.1.1. Vocabulary .1.1.4.1.2. Phrase. .1.4.1.3. Sentence. 1.1.4.1.4. Paragraph. 1.1.4.1.5. Chapter .1.1.4.2. Relationship between the rhyme word and its pronunciation based on the language characteristics. .1.1.4.2.2. How to form a word, such as Friend-LY. .1.1.4.2.3. Syntax. 1.1.4.2.4. Semantic .1.1.4.2.5. Pragmatic .1.1.5. Application arena. 1.1.5.1. Machine Translation and Machine Translation .1.1. 5.2. Speech Recognition .1.1.5.3. Speech Synthesis .1.1.5.4. Text Classification .1.5.5. Information Retrieval .1.1.5.6. Information Extraction (Information Extraction) and Automatic Summarizing .1.1.5.7. Human-Machine Interface .1.1.5.8. Story Comprehension and Q & A System
.1.1.6. Related disciplines. 1.1.6.1. Cross. 1.1.6.2. Philosophy of a word and a sentence make sense, how to specify the object in the world. What is the belief, goal, and meaning, what is the relationship with the language. Through the intuitive intuition, the natural language is expanded; .1.1.6.3. Mathematics. 1.1.6.3.1. Mathematical logic. 1.1.6.3.2. Figure .1.1.6.3.3. Probability .1.1.6.4. Linguistics The structure of the language, how the word forms a phrase, how the phrase forms a sentence, what is the meaning of a sentence, etc. Tools for research: humans' intuition for appropriate grammar and meaning, and some mathematical tools such as form language theory, model theory semantics. .1.1.6.5. Psychology research the process of human language generation and understanding, how to identify the correct structure of the sentence, when to determine the correct meaning of a word, and how to understand the process. The method of research is to measure experimental techniques for human object implementation, and statistical analysis of observations. .1.1.6.6. Computer science. 1.1.6.6.1. Artificial intelligence. 1..1.6.6.2. Machine study. 1.1.6.6.3. Mode recognition. 1.1.6.7. Information Science .1.6.7.1. Database. 1.1.6.7.2. Data Mining. 1.1.6.7.3. Data Warehouse .1.1.6.7.4. Information Extraction .1.1.6.7.5. Automatic Abstract .1.1.6.7.6. Information Categories .1.1.6.7.7 Information Retrieval. 1.1.6.7.8. Information Filtering. 1.2. Features of English. 1.3. Chinese Features
2. Volume 3. Lessment 4. Syntax .4.1. How to form a phrase, words and phrases to form a correct sentence, and every word role in the institution in the sentence. .4.1.1. Task of syntactic analysis For the analysis of natural language, syntactic analysis has the following two main tasks: 1. Identify sentence of a language and determine the structure of the input sentence given the language method G and the language L of the grammatical description, (1) give a string S, determine if S is to L; (2) give a string S, if S belcomes L, gives the tree structure of the S, 3. Standardization of Syntactic Structure If we can map a large number of possible input structures to a fewer structures, then subsequent processing (e.g., semantic analysis) is simplified. Below is an example of several structural standardization: (1) Some ingredients can be omitted or "zero" in the sentence; (2) Various conversions can link the synthesis of the surface structure, such as active tone and passive tone; 3) Normal word sequence and so-called split structure: That I Like Wine is evident. It is evident this ike Wine. (4) Nouncing structure and Verbic Structure: The Barbarians'Destruction of Rome The Barbarians Destroyed Rome, etc. Such a class of conversion makes a subsequent processing only with a small number of structures. .4.1.2. Different types of syntax analysis. Traditional Non-Probability Analysis Method Probability Method (PCFG) 2. PARTIAL PARSING / SHALLLOW PARSING 3. TOP-DOWN syntax analysis Predicative Parserbottom-UP syntax analysis Shift-reduuce Parser4. Deterministic Parser Analysis of Non-deterministic Parser.4.1.3. Form Syntax Camp 1) TG, GB, MP, ... 2) LFG, GPSG, HPSG, ... 3) PATR-II, DCG , Fug, ... 4) Tree neighborly (TAG) 5) Link Grammar 6) Categorialgrammar 7) Dependency grammar 8) Word syntax (Word grammar) ....4.1.4 Classification of contemporary form grammar theoretical system
.4.1.5. Evolution history of form syntax theory. 4.2. Theory. 4.2.1. Form language and automaton. 4.2.1.1. Basic concept. 4.2.1.1.1. Basic concept. 4.2.1.1.1.1. Letter Table is a non-empty limited set of elements. We refer to the elements in the alphabet to symbols, so the alphabet is also called symbolic sets. .4.2.1.1.1.2. Word (also called string, symbol string) and empty characters (also called empty strings) a poor sequence composed of elements in alphabets. In the symbol string, the order in which the symbol is important. If there is M symbols in a symbol string x, it is called M. Expressed as | x | = m. The sequence of any character is not included, it is ε. | ε | = 0. All words in the alphabet σ are σ *. Σ * is called a symbol string collection on alphabet σ. .4.2.1.1.1.3. Empty set does not contain any elements of any elements, remember to φ. . Also known as the N times power of V) is recorded as VN = VV ... V, V is n specified V0 = {ε}. Let V * = V0 ∪v1∪v2∪v3∪ ... called V * is a V-closure. Remember V = V v *, called V is positive (then) closing package. Obviously, εx = xε = x, x is a symbol string; or {ε} x = x {ε} = x, x is a symbol string collection. .4.2.1.1.2. Regular and formal set below: Regularity and formal set recursive definitions: 1. Ε and φ are all regular forms of σ, which are {ε} and φ; 2. Any a ∈σ, a is a regular shape on the σ, which is {A}; 3. Assume that u and v are correct, and the regular sets they represent are L (U) and L (V), then (u | v), (uv) and (u) * are also Regular formula, the regular sets represent are L (u) ∪L (V), L (U) L (V) (Connection Set) and (L (U)) * (closed). The expression defined only by the following steps is only a normal set of σ on the σ, which is only the formal set of σ only by these regularly. If the formal set represented by two formal u and V is the same, it is considered to be U and V equivalence, and it is written as u = v. .4.2.1.2. Automata. 4.2.1.2.1. State conversion diagram status conversion diagram is a finite direction map. In the state conversion diagram, the node represents the state, represented by a circle. The state is connected with an arrow arc. The mark (symbol or symbol string) on the arc arc represents the input symbol or symbol string that may appear in the state of the knot (ie, arrow arc). A state transition diagram only contains limited state, some of which are referred to as an initial state, and some are referred to as a final (represented by a double circle).
The use of state conversion graphs can construct a lexical and syntax analysis program. However, in order to analyze the automatic generation of the program, it is necessary to form a state transition graph. This creates automatic machine theory. . 4.2.1.2.2. Ε-Closed Packet and A Arc Conversion 1) Status Collection I ε - Closed Pack, is ε-Closure (i), is a state set: a) If s ∈i, s ∈ε-Closure (i); b) If s∈i, the state s' ∈ε-Closure (i) that can be reached from the arc of the arccable epsterer. 2) The A arc conversion of the status set i, is represented as Move (i, a), which is a state set: order j = move (i, a), then J is all those that can be obtained from one of I The whole state of the status arrives at an arc. For status set I and arc A, we define IA = ε-closure (j), where j = move (i, a) is IA is an ε-closure of the A arc converted of the status set I. . 4.2.1.2.3. Determining a finite automation (DFA) one determination finite automation (DFA) m is a five-yuan m = (S, σ, f, s0, z), 1. S is a limited set, and each element is called a state; 2. Σ is a poor alphabet, and each element is called an input character. Therefore, σ is called an input symbol alphabet; 3. f is a portion mapped from S * σ to S (single value). f (s, a) = s 'means: When the current state is S, when the input character is A, it will be converted to the next state S'. We call S 'a subsequent state of S; 4. S0 is an element in S, is the only initial state, also known as the start state. 4. Z is a subset of s and is a final set (blank). The final state is also called acceptable state or end state.
Determining a finite automation (DFA) can represent a (determined) state conversion diagram. .4.2.1.2.4. Undetermination of finite automation (NFA) A non-determination finite automation (NFA) m is a five-yuan m = (S, σ, f, s0, z), 1. S is a limited set, and each element is called a state; 2. Σ is a poor alphabet, and each element is called an input character. Therefore, σ is called an input symbol alphabet; 3. F is a mapping from S * * to S. That is, F: S * σ * à2S4. S0 is a subset of s and is a non-empty primitive set. 5. Z is a subset of s and is a final set (blank).
Non-determination of finite automation (DFA) can be represented as a (non-determined) state conversion diagram.
DFA is a special case of NFA. However, there is a DFA m 'for each NFA m to make L (m) = L (m'). . 4.2.1.2.5. Determining the simplification of a limited self-motivation so-called a determination of limited self-motivation M is simplified: looking for DFA M 'of a state ratio m, making L (m) = L (m) ). We said that a poor self-motivation is simplified, ie, it doesn't have a lot of state and there are no two in its state. A poor self-motivation can be converted into a smallest and equivalent of a poor automation by eliminating excess states and consolidated equivalents. The excess state of the so-called "self-motivation refers to the state of such a state: from the start state of the automaton, any input string cannot be reached. It is assumed that S and T are two different states of DFA m, and we call S and T are equivalent: if you can read a word α from the status S, then, then, from t, from T. After reading the same word α, it is turned on; contrary, if you can read a word α from the status t, then the same word α can be read from the Siya word α and stopped from S. state. If the two states S and T of the DFA M, these two states are distinguished. We introduce a method called the "Segmentation Law" to divide a status of a DFA M (excluded) into some non-intersecting subsets, so that any different two subset is different, and the same child Any two status of the concentration is equivalent. Steps for diving the status set S of the DFA M separately separate S the terminal and non-tetheral separation, divided into two subsets, forming a substantially differentiation π. 2) Assume that π has a m subset of π, π = {i (1), i (2), ..., i (m)}, and the state belonging to different subsets is different. Then check that each I in π can be further diverted. For a certain I (i), let I (i) = {S1, S2, ..., SK}, if there is an input character A such that IA (i) is not included in a subset I (J) of the current π The I (i) is divided into two: i (i1) and i (i2) such that the state in the status and i (i2) in i (i1) is different, so that new points are formed. Π. 3) Repeat 2) until the number of subsets contained in π is no longer growing, get the final scratch π, for each subset in this π, we choose one state in the subset representing other status, which is obtained DFA M 'and the original DFA m are equivalent. .4.2.1.2.6. The conversion theorem of Nfaàdfa: Sets L for a collection of uncertainty. There is a poor self-motivation that accepts L determined. Subcommination: A algorithm for converting NFA into DFA receiving the same language. The following detailed description: Basic ideas: each state of the DFA corresponds to a set of states of NFA. This DFA uses its state to record all states that may be reached after the NFA reads an input symbol. That is, after reading the input symbol string A1A2 ... AN, the DFA is in such a state, and the state indicates a subset t in the state of the NFA, T is from the start state of the NFA along a certain labeled A1A2. ... The path of the AN can arrive. Algorithm: For an NFA Mn = (SN, σN, FN, S0N, Zn), we constructed an MD = (SD, σd, fd, s0d, zd) in accordance with the following method, so that L (Mn) = L (MD) : 1) The status set SD of the MD is composed of some subsets of the SN (the algorithm of these subsets of Sn will be given later).
We use [SD1, SD2, ..., SDJ] to represent any of the SD, where SD1, SD2, ..., SDJ are SN state. Further, the state SD1, SD2, ..., SDJ are arranged in a certain rule, that is, the state of the SD is {SD1, SD2}; 2) MD and MN input alphabetic table. The same, 即 = =N; 3) The conversion function FD is defined in this: FD ([SD1, SD2, ..., SDJ], A) = ε-closure (Move ([SD1, SD2, ..., SDJ] , A)); 4) S0D = ε-closure (S0N); 5) zd = {[SDP, SDQ, ..., SDR] | [SDP, SDQ, ..., SDR] ∈SD & {SDP, SDQ, ..., SDR} ∩ZN! = Φ}
Algorithm for the subset of state SNs constructed in the NPA Mn is given below. It is assumed that the structured subsets are c, ie c = (T1, T2, ..., Ti), wherein T1, T2, ..., Ti is a subset of state Sn: 1. Start, let ε-Closure (S0N) are unique members in C, and it is not marked; 2. WHILE (a subset of subsets that has not been marked in C) DO {tag T, for each input letter A (a! = ε) Do {u: = ε-closure (Move (t, a)); if u is not C in Tken will be used as a subset of unmarked subsets in C;}}
For example: convert the NFA represented by the following figure into DFA. .4.2.1.3. Text Law. 4.2.1.3.1. Rules also known as the REWRITING RULE, Production Rule, or the generating formula, is like α} or α :: = β (α, β Ordered. Where α is a symbol in the positive closure V of a certain alphabet V, β is a symbol in V *. α is called the left, beta called the rightmost portion of the rule. .4.2.1.3.2. A grammatical g of literacy is defined as a quadritical group (VT, VN, S, R), where VT is a terminal symbol set, is a non-empty limited set; the terminator is a basic symbol of the composition. VN is a set of non-final symbols (or syntax entities, or variables), is a non-empty finite set; non-terminator is used to represent grammatical categories; vt∩vn = φ. S is called the identifier symbol or start symbol. It is a non-end symbol, at least in a rule as the left; R is a collection of generated (also known as rules), each generating type α}, α, β ∈ (vt∪vn) *, and α There must be at least one non-final, and cannot be empty characters; at least one of the genes in R is active as s. Usually vt∪vn, V is called alphabet or went sheet of grammar G. For example: g = (vt = {0, 1}, vn = {s}, s, r = {sà0s1, sà01})
Three effects: 1) Generate: Generate all the sentences in language L; 2) Determine: Whether a string belongs to language L; 3) Analysis: Get the structure of the sentence of the sentence;. 4.2.1.4. Language. 4.2.1.4.1. Direct derived / derivation / can be derived from the grammatical g = (Vt, Vn, S, R), we call αa] directly derive γβ, αaβ ==> αγ β is only R in the R A production type, and α, β ∈ (vt∪vn) *. If α1 ==> α2 ==> ... ==> αn, this sequence is called a derived from α1 to αn. If there is a derived from α1 to αn, α1 can be said to be derived. Use α1 = => αn: From α1, one or several steps can be derived; αn can be derived; α1 = * => αn is represented by α1: from α1, it can be derived from 0 step or several steps. 4.2.1.4.2. The left deduction / right-derived derived derivation: Any one-step α ==> β is replaced with the least left non-ending in α. Right derivation: Any one-step α ==> β is replaced with the right non-ending in α. In form languages, the rightmost derivation is often referred to as a standard derivation. The sentence pattern derived from the specification is called a standard sentence. .4.2.1.4.3. Sentence / sentence / language For grammatal g = (VT, Vn, S, R), if s = * => α, α is called a sentence pattern. The typographical type containing only a good sentence is a sentence. The whole sentence of the sentence G is a language that will be recorded as L (g). L (g) = {α | s = => α & α∈vt *} For grams g1, g2, if L (G1) = L (G2), the literal methods G1 and G2 are equivalent. . Sentence, this language is recursive. Recursively Enumerable Language If you can write-part programs, make it possible to output (ie enumerate) one language in some order, saying that this language is recursive. . .4.2.1.5. The formal language Jumsky (Chomsky) established a form language in 1956. Jumsky divided the grammar into four types, namely, type 1, 2, and 3. The difference in these types of grammar is to apply different limits to the generating type. For gelegial g = (Vt, Vn, S, R), 0) If each of the generated α} is satisfied: α∈ (vt∪vn) * and contains at least one non-final, and β∈ (vt∪vn ), Then g is a 0 type text method (PSG). 0 type text method is also known as phrase structure grammars. A very important theoretical result is that the ability of the 0 type literary law is equivalent to Turning. Alternatively, any 0 language is recursive and can be enumerated; in turn, recursive enumeration set must be a 0 type. But some languages are not recursive.
1) Set G is a 0 type text method, if each of the generated α} is satisfied | α | <= | β |, only Sà ε is except for Sà ε, but S must not appear in the right part of the generated, then the gramatory G is a 1 type Or the context-related literary law (CSG). One equivalent definition: set G is a 0-type text method, if each of the G is αaβ ==> αγβ, a∈vn, and γ is not ε, α, β, γ ∈ (vt∪vn) * If the text method G is a 1-type text method or a context. This definition shows that only A appears in the context of α and β, γ is allowed to replace A. 2) Set G is a 0-type text method, if each of the G is a à beta, a∈vn, β∈ (vt∪vn) *, the grammatial g is a 2-type text method or context-free (CFG), also known as BNF paradigm (Backus-Naur Form or Backus Normal Form). This definition shows that the replacement of non-finals can not consider context. The context does not have a copy of the self-motivation corresponding to the non-determination. 3) Set G is a 0 type text method, if each of the G is AàαB or A, A, B α∈vt *, a, b∈vn, the grammatial method G is a 3-type text method or a formal grammar (RG) or right linear grammar. 3 models or formal grammar (RG) Another definition is: set G is a 0-type text method, if G's G is Aà Bα or Aàα, α∈vt *, a, b∈vn, then the literacy G is 3 types or formal grammar (RG) or left-line grammar. Obviously, an NFA can be designed for any 3-type text method G, which can only recognize the language of G. The definition of the four grammar is gradually increasing, so each formal literary law is unrelated to the context, and each context is related to the context, and each context-related literary law is a 0 type. The language generated by the 0 type text method is 0 type. The context-related grammar, the language of the context-free level method and the formal grammap generation is called context, respectively, and the context-independent language is formal.
Difficulty in determination of various types of grammatics: 1) PSG: Semi-semi-determination for a sentence L belonging to GTYPE0, can always determine "Yes" in the determination; but for a sentence L 'that does not belong to GTYPE0, there is no algorithm, You can determine "NO" within the determination step. 2) CSG: can be determined, complexity: NP is complete. 3) CFG: can be determined, complexity: polynomial. 4) RG: can be determined, complexity: linear. . 4.2.1.6. Equivalence and equivalence of normal and finite automators and the equivalence of the self-motivation are described below: 1. For NFA m on σ, a normal formar R routing on σ can be constructed such that L (r) = L (m); 2. For each of the normal raride rs on σ, an NFA m on a σ can be constructed such that L (m) = L (r).
Proof: 1) The corresponding regular R constructed of NFA m on σ. We broaden the concept of the status conversion graph to make each arc can be marked with a regular formal. In the first step, two knots are added to the state transition diagram of M, one is an X node, one is Y node. Use the ε arc to all of the initial nodes of M, and connect from the total end of the M full to the Y node with an ε arc. Form a M ', M' only one initial X and a final Y. In the second step, you will gradually eliminate all nodes in M 'until there are only x and y nodes. During the dispensing process, the arc is gradually marked. The rules of the decline are as follows: The label in the last X and Y nodes is a regular R. 2) A NFA M on any of the normal R constructs from σ. The method of l (m) = l (r) is made. a) Represents the normal Reaver Reated conversion diagram: R or when R = φ is:
b) Step by transforming this figure by splitting and adding a knot to R, each arc is marked as σ, a character or ε. Its conversion rules are the degree of decline in 1). For example: R = (A | B) * ABB constructs NFA N such that L (n) = L (r). . 4.2.1.7. Formal grammar is equivalent to formal, formal language is equivalent to formal set. 4.2.1.7.1. Formal grammar is equivalent to the formal formal grammar, there is a formal form of definition of the same language: Conversely, there is a formal grammap that generates the same language for each formal shape. Certificate: 1) A regular conversion of a formal conversion to grammar g = (Vt, Vn, S, R). Let the VT = σ, determine the elements of the generated and VN, as follows: a) Select a non-terminal Sàr to generate a generating Sàr for any formal formula R, and set S to the identification symbol of g. B) If x and y are regular, the generation of Aàxy is generated, rewritten into: A-XB, B yY two generation, where B is the newly selected non-end, ie B∈VN. C) Written as a àx * y in the conversion-converted grammatism, is written as a àxb aày bàxb bày, where B is a new non-end. D) For the generation of Aàx | Y, rewriting is: aàx, aày constantly uses the above rules to change until each generation is up to one end. 2) Convert formal grammatism into a regular basis. Basically, the reverse process of the above process. Finally, only one start symbol definition is left, and the right portion of the generating type does not contain a non-end. The formal grammar to the formal conversion rule is listed in the table: the formal formal regular rule 1 rule 2 rules 3 aàxb, bàyaàxa | yaàx, aày a = xya = x * Ya = x | Y,: 1) R = a (a | d) * Convert to the corresponding formal grammar; 2) will grammar g = (vt =}, r = {s, a}, s, r = {sàaa, sàa, aàaa, aàda, aàa , Aàd}) Converts to the corresponding formal grammar; .4.2.1.7.2. Regular language is equivalent to formal set. 4.2.1.8. Sormal grammar and finite automatic conversion. 4.2.1.8.1. Qualal grammar to limited automatic The conversion of the machine is directly constructed from formal grammatics G, so that L (m) = l (g): a) alphabet is the same; b) for each non-terminator in G Generate a state of M, (may disable the same name). The start symbol S of G is the start state S. c) Increase a new state Z, as the final state of NFA; D) a conversion of M-shaped in the form of a g of a in g, and the end of the endon or ε, a and b is a non-finalizer) Function f (a, t) = b; the generation of G is in the form of Aàt. Constructs a conversion function f (a, t) = z.
.4.2.1.8.2. Conversion of finite automation to formal grammar directly constructs a formal text method G from a poor self-motivation NFA m, so that L (g) = L (m): a) The alphabet of the poor self-motivation is The end symbol set of the grammar; b) The initial state of the poor self-motivation correspondence; C) The conversion rules of the poor automatic motivation are very simple: conversion functions f (a, t) = B, can write a generating style: A-TB adds a generating type: zà ε.4.2.1.9. The main task of lexical analysis is to scan the input string from left to with a character string, generate a word sequence for grammar analysis. Regularity is used to illustrate the structure of the word is very simple and convenient. Then, a formal compiled (or call) is an NFA to convert to the corresponding DFA, which is an identifier that identifies the sentence of the language represented by the formality. .4.2.1.10. The syntax analysis of the context of unrelated grammar context has sufficient ability to describe the grammatical structure of today's programming language. Currently, the context of context is unlicensed as a description tool for programming language syntax. .4.2.1.10.1. Syntax Tree Law Tree also known as the derived tree, it is an intuitive method for describing the sentence pattern derivation of the context. For the context without any sentence type type type G = (VT, Vn, S, R) can construct the grammar tree associated with it, this tree meets the following four conditions: 1. Each node has a tag that is a symbol of (vt∪vn) *; 2. The tag of the root is S; 3. If a node n has at least one of its own descendants, and there is marker A, then A is definitely in VN; 4. If the direct descendants of n (marked as A), from left to right order is nodes N1, N2, N3, ..., NK, which are marked as A1, A2, A3, ..., AK, then Aàa1, A2, A3, ..., AK must be a generating type in R. Example: For grammatism g = ({s, A}, {a, b}, s, r), where R is (1) sàaas (2) aàsba (3) aàsss (4) sàa (5) aàba construct sentence Aabbaa Grammar tree. .4.2.1.10.2. Erlining of the grammar If a sentence has a sentence corresponding to two different grammar trees, it is said that this grammat is secondary. Or, if there is a sentence in a literary method, it has two different left (or right) derivation, and this grammat is secondary. Theorem: The two-meaning problem is not determined. That is, there is no algorithm that can be defined in a limited step in a limited step. .4.2.1.10.3. Syntax analysis method From left to right analysis: ie, always identify input symbol strings from left to right, first identify the leftmost symbol in the symbol string, and then identify one symbol on the right. From the right direction to the left: ie, the input symbol string is always identified from right to left, first identify the rightmost symbol in the symbol string, and identify one symbol on the left. Since the top, the analysis method is also known as the top-down analysis method, and the target-oriented analysis method. Departure from the beginning of the grammat, repeatedly use a variety of generated, looking for "matching" in the input symbol string. The top-down analysis can be divided into two types of identified and uncertain, and the determined analytical method needs to be limited to the literary law, but due to the simple, intuitive, intuitive, intuitive, easy manual construction or automatic generation of syntax, thus It is still one of the currently common methods. Uncertain methods, that is, the backtracking analysis method, this method is actually an exhaustive test method, so the efficiency is low, the cost is high, and thus is minimal. Self-analysis: starting from the input symbol string. Gradually "detriment" until the beginning of the cultural law
The way from grammar tree can understand the difference between these two types. Since the uppermost method is starting from the grammatical symbol, it is the root of the syntax tree, and gradually establishes a grammatical tree down, so that the end node symbol string of the grammar tree is just an input symbol string; from bottom to bottom The input symbol string begins, with the end node symbol string of the syntax tree, constructed the syntax tree up to bottom. The following discussed is from left to right analysis. . 4.2.1.10.4. Since the top and the analysis method. 4.2.1.10.4.1. Problem In the top and down analysis method, it is assumed that the least left non-end symbol to be converted is V, and there is N rules: Vàα1 | α2 | ... | αn how to determine which right to replace V? There is a solution to randomly select one from a variety of possible choices, and hope it is correct. If it is wrong later, we must return back, try another choice, this is called back. Obviously this cost is extremely high, the efficiency is very low, which is the problem we need to solve. .4.2.1.10.4.2. Recursive drop analysis. 4.2.1.10.5. Self-analysis. 4.2.1.10.5.1. Basic concept. 4.2.1.10.5.1.1. Phrase / direct phrase / handle g It is a text method, and S is the beginning of the grammar, and αβδ is a sentence pattern of cultural law G. If there is: s = * => αaδ and A = => β is referred to as a phrase αβδ relative to the non-finals A. In particular, if A ==> β is called β, it is a direct phrase (also known as simple phrase) relative to rule a à beta. The leftmost direct phrase of a sentence pattern is called the handle of the sentence. .4.2.1.10.5.1.2. Normative regulations. 4.2.1.10.5.2. Problem In the bottom-up analysis method, every step of the analysis program works from the current string, will Arrived to a non-end symbol, we temporarily refer to this substring as "can be detrimental." The problem is how each step is how to determine this "can be contracted", that is, how to choose a string in each step, so that it can be contracted, not unable to destination. . 4.2.1.10.5.3. The operator priority analysis method. 4.2.1.10.5.4. LR analyzer. 4.2.1.11. Reference 1) Compilation of "Programming Language Compilation", National Defense Industry Press 2) Principles, Tsinghua University computer series textbook. 4.2.2. State transfer network. 4.2.2.2.1. Finite state transfer network (FSTN) A finite state transition network (Finite State Transition Network) by a set of status (ie nodes) and one An arc (used to put a state connection to another state): 1) One of the states is designated as the starting state; 2) Labeling the syntax in each arc (including the word Or words, etc.). It indicates that such a word must be found in the input sentence, and the transfer specified by this arc can be performed; 3) There is a subset of the end state in the state set. If the header of the input sentence (or phrase) starts from the starting state, after a series of transfer, the end of the sentence just reaches the end state, say this sentence (or phrase) is accepted (or identified) by this transfer network. Description: The finite state transfer network can only be used to generate or identify the formal (ie 3) language. For example: The following figure identifies the "Dong Yong likes the seven fairy" FSTN. .4.2.2.2. State transition table (State Transition Table) Status Transfer Arc N V Q0 Q1 Q1 Q1 Q2 Q2 Q1
.4.2.2.3. Recursive Transfer Network (RTN) Recursive Transition Networks, referred to as RTN) is an extension of a limited state transition network, and the label of each arc in the RTN can not only be a finalist (word or Word class, etc.), and can be a non-finalizer used to indicate another network name. Description: (1) Any subscriber in RTN can call any other network included in the network. (2) From the generation capability, the recursive transfer network is equivalent to the context. For example: .4.2.2.3.1. RTN algorithm (TOP-DOWN). 4.2.2.2.3.1.1. Algorithm Description ---- Basic Concept Subnet Name: S, NP, VP and other status nodes: Q0, Q1, Q2; Outside: From the current state, the arc is transferred to the next state; W1W2W3 ... (the idea is Dong Yong thinks); recording stack: record from which subnet, and return to the upper subnet State; current status: , the string pointer points to the first character of the string to be analyzed, and the stack is empty. 2. 2.1. If the current status node is not a termination state: Pointing the current state node out of the pointer to the first out, 2.1.1. If the currently outline is labeled as a terminal, compare the tag and the current string move pointer finger 19.1.1.1. If equally, the prediction is verified, the subtree is constructed, and the current state node is set to the subsequent state node of the currently outbound, and the current outline; 2.1.1.2 is modified. If you don't wait, the front edge pointer points to the next one. If there is no next outbound and there is a retrospect point, it will be traced, otherwise the analysis failed. 2..1.2. If the currently outline is marked as the non-final, the subsequent state of the current subnet name and the current node, and set the current status node to the start state of the current subnet, and modify the current outbound and subsequent state. 2.1.3 If there is multiple options now, you need to set the recovery breakpoint, save the rendering stack, to analyze the string, the current ATN state, and the full content of the outbound list status to use. 2.2. If the current state node is the termination state but not the termination state of the subnet S: The rendered stack is retired, and the current state is set to the state of the current stack top, and the transfer step 2 continues. 2.3. If the current state node is the termination state and is the termination state of the subnet S: 2.3.1. If the reboot is empty and the string to be analyzed is empty, the analysis is successful, ended; 2.3.2. Otherwise, if there is a retrospect point, backtrack. 2.3.3. If there is no retrospect point, the analysis failed.
.4.2.2.3.1.3. Example analysis "I am a county magistrate": .4.2.2.4. Augment Transition Network, ATN) Expansion Transition Networks, referred to as ATN) This formal system It is 1970 W. Woods is proposed and successfully applied to his famous Lunar system. .4.2.2.4.1. Basic idea ATN syntax belongs to an enhanced context without syntax, its basic idea is to continue to use the context-free syntax to describe the composition structure of the sentence; but add some functions to individual generation in the grammar Mainly describe some necessary grammar restrictions and the deep structure of the sentence. .4.2.2.4.2. ATN's expansion ATN of RTN is extended and enhanced on RTN in the following three: (1) Add a set of registers to store intermediate results obtained during the analysis (such as local) Syntax tree) and related information (as number of noun phrases, semantic characteristics, etc. of some components). But setting which registers are fully dependent on the needs of syntax analysis, and there is no hard regulation; (2) In addition to the scope of the syntactic category (such as the words and phrase tags), any test (TEST) can only This arc can be passed after this test is successful; (3) Actions can also be attached to each arc. When passing a arc, the corresponding action is sequentially executed, these actions are mainly used. To set or modify the contents of the register.
. 4.2.2.4.3. ATN form system The following is an ATN form system defined with BNF:
(2) The JUMP arc is different from the CAT arc, which does not consume the current word in the input string, so the transfer of the JUMP arc can occur without any processing of the input string without any processing. (3) A Push arc means a network inverse return, its second element
In ATN, each subnet has its own register table. SETR sets the contents of one specified register in this layer register table to
.4.2.2.5. FSTN: Transducer. 4.2.2.2.6. Reference 1) Shi Junyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principle", Tsinghua University Press 2) Zhan Weidong, "Computational Linguistics Introduction "Sectation, Peking University Chinese Department. 4.2.3. Characteristic Structure and Complex. 4.2.3.1. Feature Structure and Complex Feature Structure 1) A Feature Structure (FS) is a feature description binary group
Description: 1. From the above definition, this is a recursive definition. Since the value V and V 'itself can also be a complex feature set, α and β can only be combined only when V and V' can be combined. 2. If the natural language is regarded as a message delivery system, and recognizes the synthesis of the natural language, that is, whether it is a syntactic component or semantic ingredients, it is gradually combined by small to a large manner. Basic operations as syntactics and semantic analysis are ideal. .4.2.4. * FUG function combined with the function of M.KAY is proposed in 1979, he first introduced an operation into grammar theory. .4.2.4.1. Basic Features a) The biggest feature is that the terms definition, syntact rules, semantic information, and structural functions of the sentence can be represented by complex feature sets, and complex feature set in FUG is called function description ( FD). b) Weakeline linear structure relationship; C) emphasize functional structure; d) Description of all language units uniformly adopted FD form; e) suitable for generating. .4.2.4.2. A combination of application formal formal grammar system 1) Generation rules: a basic scope of a language and its combination type 2) Combined equation: combination relationship between basic categories; 3) Tree Structure Feature Structure: Describe the structural relationship of the sentence, semantic relationship information. . 4.2.4.3. From a simple example, S = "one piece of clothes" S is the internal composition: MP NP, set-neutral structure, MP is attributive, NP is a central language; "one" with "clothes" It can be combined into a set of structural structures; the external function of S is a noun phrase (NP) as a whole, which can be derived with the subject of the subject, the object of the guest structure, the central language of the center; Structure of the complement of the structure, the idioms of the design, the contrams and the central language of the structure. .4.2.4.4. Rules && {r1} np-> mp! Np :: $. Internal structure = set, $. =% Mp, $. Central language =% np, $. Dingyu = No, ..., 1 % NP. Quantity name = Yes, ..., 2IF% MP. Quantifier class = individual THEN% NP. Identh quantifier =% MP. Original Endif, ... 3 && {r2} MP -> M! Q :: $. internal structure = =% M, $. 定 =% =% q, $. Dingyu = Yes, ..., && {r3} np->! N :: $. Internal structure = word note: 1 Description NP internal Structural, attributive, central language, and functional characteristics, etc .; 2 Describe the constraints (independent conditions) of the central language NP; 3 describes the mutual constraints between the attributive MP and the central language NP (group with conditions). .4.2.4.5. Dictionary [Words: M, Number: Q, Quantifier Class: Individual, Number: Number] Clothes [Words: N, Nouns] NA, Quantity Name : Yes, individual quantities: pieces | sets, ...] cardiac [Words: N, Nouns] NE, Quantity Name: No, ...]. 4.2.4.6. Analysis results of "a piece of clothes" results Clothes word internal structure Word phrase Type Central Word Word Internal Structure Phrase Type The Split Structure Phrase Type Internal Structure Phrase Type
.4.2.4.7. Reference f) Martin Kay, 1979, Functional Grammar, In Proceedings of the 4thAnnual Meeting of the Berkeley Linguistics Society.g) Martin Kay, 1985, Parsing in Functional Grammar, In D.Dowty, L.Karttunen, And A.zwicky EDS, Natural Language Parsing, Cambridge University Press, Cambridge, 1985..4.2.5. * LFG vocabulary function syntax (Lexical Functional Grammar) is J. BRESNAN and R. M. Kaplan was proposed in 1982. . 4.2.5.1. Basic characteristic basic idea: Relying on the tree structure of the phrase structure syntax, spread the various information loaded by the word remittance by the bottom-up (Bottom-Up) layer, bring together to the upper node The complete structural information and functional information describing a sentence are finally formed. Analysis: Constituent Structure: The syntax tree description of the sentence; functional structure: The function description of each component in the sentence; .4.2.5.2. Rewriting Rulesà NP VP with "modified" (↑ Subj) = ↓ ↑ = ↓ vpà v np ↑ = ↓ (↑ obj) = ↓ vpà v ↑ = ↓ npà n ↑ = ↓ Note: ↑ Represents a father's day; indicates the current node; ↑ Subj means the Subj of the father's day Characteristics; = represents the one. .4.2.5.3. Vocabulary information description (Characteristics) Dong Yong N (↑ lex) = 'Dong Yong' (↑ SEM) = Human (↑ PERS) = 3 seven fairy ni like V (↑ PRED) = 'like < ↑ Subj), (↑ Obj)> '(↑ subj SEM) = human (↑ obj SEM) = human.4.2.5.4. C- structure. 4.2.5.5. From C-structure to F-structure
.4.2.5.6. Export function equation from the component structure 1. x1.Subj = x22. X1 = x33. X3.obj = x44. X2. o = "Dong Yong" 5. x2.sem = human6. X2.pers = 37 . x3.pred = "Like <(x3 subj), (x3 obj)>" 8. x3.subj.sem = human9. x3.obj.sem = human10. x4.lex = "seven fairy" 11. x4.sem = Human12. X4.pers = 3 Note: "Equivalent to" "
. 4.2.5.7. Solution functional structure from functional equation
X1 = x3 =
X2 = x4 =
"Dong Yong likes the seven fairy" functional structure:
FS (x1) =
.4.2.5.8. Standards for LFG check sentence legality 1) A property only allows a value (uniqueness); 2) There should be value (complete) each attribute; 3) The attribute that should not be there should not be Value (consistency). .4.2.5.9. Reference 1) Kaplan, R. And J. BRESNAN, 1982, Lexical-Functional Grammar: a formal system for grammatical representation, in j.bresnaned. The Mental Representation of Grammatical Relations, Mit Press 1982. 2) http://www-lfg.stanford.edu/lfg/bresnan/3)http://www.parc.xerox.com/istl/members/kaplan/.4.2.6. GPSG Generalized Phrase Structure Structure (Generalized Phrase Structure) Grammar) is proposed by Gazdar et al. In 1985. .4.2.6.1. Basic character syntax model: Basic rules -> Quality check -> Tree structure Basic Rules: Meta Rule Quality Check: Feature Co- occurrence Restriction) describe linear order (linear Precedence Statements) ...... .4.2.6.2. References 1) erald Gazdar, EwanKlein, GeofferyK. Pullum, Ivan A. Sag, 1985, Generalized Phrase Structure Grammar, Oxford, England, Blackwell Publishing and Cambridge, 1985.2) http://www.cogs.susx.ac.uk/LAB/nlp/gazdar/gazdar.html.4.2.7. HPSG center word-driven phrase structure syntax (also known as Nuclear-driven phrase structure syntax).
.4.2.7.1. Basic Features 1) Emphasizing the Center in the Foundation of the Center in the Phrase Structure Rules - Head -Complement Rule Central Words - Head -Specifierrule Central Words - Modification Rules (Head -Modifier rule) 2) Production rules feature structure combined operation 3) Based on the attribute feature transfer of central words (Head feature principle, ...) 4) Expresses syntax knowledge and semantic knowledge in the same form of formation. 4.2.7.2 References 1) Pollard, Carl andIvagA. Sag. 1987. Information Based Syntax and Semantics. CSLI Lecture Notes, No.13, The University of Chicago Press, Chicago.2) Pollard, Carl andIvagA. Sag. 1994. Head-Driven Phrase Structure Grammar The University of Chicago Press, Chicago.3) Sag, Ivan A. & ThomasWasow, 1999, Syntactic Theory:. A Formal Introduction, CSLI Publications, Stanford, California.4) http: //ling.ohio-state. EDU / Research / HPSG / 5) http://hpsg.stanford.edu/.4.2.8. PATR-II.4.2.8.1. Basic Features Introducing the feature structure and combined operation on the basis of independent CFG in rules Additional categories (such as wildcarders *.%) Introduced a function in a combination (such as Morph_Gen ()) introduced logical expression symbol enhancement rule expressions (such as multiple cleaning relationships). 4.2.8.2. Reference 1) Shieber, Stuart M., 1984, The Design of a Computer Language for Linguistic Information, In Proceedings of Coling84 (10th), Stanford University, Stanford, California.2) Shieber, Stuart M. , 1985, Using Restriction to Extend Parsing Algorithms for Complex-feature-based Formalisms, In Proceedings of the 22ndAnnual Meeting of the ACL, University of Chicago, Chicago, Illinois.3) Shieber, Stuart M., 1986, An Introduction to Unification Based Approaches To Grammar, CSli Lecture Notes Series, No.4, Stanford University.4) http://www.eecs.harvard.edu/~shieber/ .4.2.9. LSP and LST Language Stroke Analyzer (LINGUITISTIC STRING PARSER, Abbreviation LSP) is made from New York State University N. The research team led by Sager, the system design goal is to deal with a large number of scientific literature for intelligence retrieval services, so it emphasizes that the English grammar system it established should have a wide range of coverage.
This English syntax based on the language string of Z. Harris (N.Chomsky), including approximately 250 contexts, regulations and 200 restrictions. The system's dictionary income 9500 entry. This system has been used to analyze the medical records of hospitals and medicine and physiology. .4.2.9.1. Language string Theory (LST) (2) Language String Theory uses specific syntax categories (nouns, time verbs, etc.), a group of basic strings and rules to combine some basic strings into stringers; (3) Adjunct strings can be formed by inserting an additional string (Adjunct String) through the left or right side of one element of a basic string in the sentence, any string can form a more complex string; (4) sentence can also be inserted into the connection string (Conjunct String) ) To obtain an expansion; (5) String theory allows one element in the string to be replaced by a replacement string; (6) each word in the language is based on one or more words based on its syntax characteristics. According to this, each word will be accompanied by one or more of the same type. Language string Theory believes that each sentence of language is at least one word class is a stroke, ie it is built by a central string by addition, connection, and replacement. However, no combination of words selected from the correct words is inserted into a cassette to form a legal sentence. For example: a central string (Noun Tensed-Verb Noun): Programmers Write Code.è Add addition string: Programmers At Our Installation Write Length Code.è Insert Connection Skewers: Programmers At Our Installation WRTTE AND DEBUGTHY CODE.è Using Replacement Strings: Center After a string, a NOUN is replaced into a Noun Tensed-Verb adj, becomes Programmers Write Code That Proves Excellent..4.2.9.2. Reference 1) Shi Chunyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principles", Tsinghua University Publishing Society. 4.2.10. * TG conversion generating syntax (Transformational-Generative Grammar), was created by Chomsky later in the 1950s. Z. Harris proposed a language conversion theory. As the students of Harris, Chomsky has greatly developed the idea of converting grammar, and created the conversion generation school. .4.2.10.1. Reference 1) Shi Chunyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principles", Tsinghua University Press 2). 4.2.11. * GB.4.2.11.1. Reference. 4.2.12. MP .4.2.12.1. Reference
.4.2.13. DCG stator sentence syntax (Definite Clause grammar). In 1980, University of Edinburgh, University of Edinburgh, published the asis of the first stator sentence grammar (DCG). The paper proves that DCG is an enhanced context-free grammar (ACFG), which is at least not less than the famous ATN syntax. It is especially important that the syntax rules expressing the stator sentence itself is the logic program design. Language pro1OG executable program, in other words, the Pro1OG system can directly explain the syntax rules written in DCG form without having to design an syntactic analyzer (or rule interpreter) to complete this task. .4.2.13.1. References 1) Pereira, FCN, and DHD Warren, 1980, Definite Clause Grammars for Language Analysis -A Survey of the Formalism and a Comparison with Augmented Transition Networks, In Artificial Intelligence, Vol.13, pp231-278 1980.
.4.2.14. Lingware.4.2.14.1. References. 4.2.15. Tagtree Adjunct Grammar, tree neighborly. . 4.2.15.1 Reference 1) a. Joshi, L. Levy, & M. Takahashi, 1975, Tree Adjunct Grammar, Journal of Computer & System Science, 1975, 10 (1): PP136-163.2) Anneabeillé, OwenRambow, 2000, Tree Adjoining Grammars: Formalisms, Linguistic Analysis and processing, csli publications.3) http://www.cis.upenn.edu/~xtag/ .4.2.16. * LGLink Grammar, chain syntax. .4.2.16.1. Basic Features 2) Link grammar is not based on the tree structure, but the language knowledge is fully implemented on the vocabulary, and the sentence is analyzed by the link attribute of the word. . 3) The chain syntax method is shown in the analysis of the sentence as a link relationship between the words in the sentence, ie not a "tree" structure, but "Figure" structure. 4) Compared with other form grammar systems, the chain syntax is a formal grammar system that holds a strong vocabulary view. It does not emphasize the hierarchical relationship of language components, but from the local part of the vocabulary, try to reveal: whether there is contact between any two words in a sentence, and what is connected. .4.2.16.2. Words with links with link syntax to see sentences in the dictionary: sentences from the word link: .4.2.16.3. Basic concept a language of a language is the word in the language. Collection, and a link request is defined for each word (Linking Requirement). The link requirements of words can be specified by one or several link expressors (formula ofnector). .4.2.16.4. A chain language example word link expression Description A D D is a link crown (D of the name of the noun) The D represents the right link Cat D- & (O-OR S ) o is a link verb and Bimic (Object) Snake D- & (OR S ) S is a link verb and subject (Subject) chain chaline CHASED S-& O & Represents logical "and" relationship RAN S- - represents the left link Mary O-OR S OR represents a logical "or" relationship DOG {@A} & D- @A indicates that the word can have multiple A links
.4.2.16.5. Definitions of Link Expression (1) A link expression consists of a link subcutor, a binary operator (&, OR), and parentheses (specified in the priority order of combination). The link expression containing only one link is simply referred to as a link; (2) a link is composed of two parts of the chain name (Name) and the link direction; (3) The chain name is a symbol string for tag two The relationship between words; (4) there are two link directions, left (-) and right ( ); Eg Cat: D- & (O-OR S ). 4.2.16.6. Link expression inside Order If a link expression is combined by multiple links, there is a sequential requirement between the links of the side-by-column relationship, such as "CAT" link expressions are: D- & (O-OR S ) You cannot write: (o-or s ) & D-. If the latter is written, there will be an o-chain, and there is a case where there is a D chain. .4.2.16.7. Link sub-string a word in a word string If there is a link to the right, such as X , and the other word has a link to the left link X-, then the two links match each other ( Match), you can draw a X chain between these two words. At this time, we say that links sub-x or X-gets satisfied (Satisfication) or there is a link to satisfy the link sub-X or X-. .4.2.16.8. Link expressions Meeting Link Expression X & Y To be met, the link must meet the links X and Y. Link Expression X or Y is to be satisfied, the link must meet at least one of the chain connects x and y. .4.2.16.9. Use a chain language to check the synchronization of the sentence for a string (s) composed of words. If the link requirements of all words in S in a chain syntax (LG), each link requires It is only satisfied, and all links meet the requirements of the following 4 yuan rules (Planarity), the chain and chains do not cross each other; connectivity (connecture), all words should be chained together To form a communication map. Sequence (ORDERING), links in the link expressions are linked to the word closer to the word, and the linked linked sub-links in the link expressions are linked to the word far from the word. EXclusion, a pair of words cannot be two links at the same time. So S is the sentence in the language defined by LG. Making all the links of S legally referred to as a link set (LINKAGE), the link set is the result of the chain syntax analysis sentence, as the general CFG analysis result is like a syntact tree. .4.2.16.10. Application example
(1)
(2)
(1) legal; (2) illegal
.4.2.16.11. Reference 1) Chain syntax is proposed by Danielsleator, CMU (Carnegie Mellon University) Computer College, US COLUMBIA University, and the earliest article, the earliest article is a technical report in 1991. The topic is "Parsing English with a link grammar" (CMU-CS-91-196, October 1991) See: http://city91parsing.html 2) http://www.link.cs .cmu.edu / link / .4.2.17. * cgcategorial grammar, category syntax. .4.2.17.1. Basic idea corresponds to a "type" / "category" in a "type" / "category", and the construction process of the language structure corresponds to the calculation process between "Type" / "Category". .4.2.17.2. Basic concept 1) There are two basic categories: s and N. Roughly understanding, S is equivalent to "sentence", n is equivalent to "noun". 2) Category (type) of a language component is made of a basic category S, N plus a categorical expression constructor "/", "/", parentheses "(", and ")". 3) Categorical constructor "/" means "left deficiency"; "/" means "right deficiency", intuitive, can be envisaged as a division, accordingly, the construction of the category can be regarded as a direction with a direction Operation. Brackets indicate binding order. 4) When the relationship occurs between the two language components, they have the corresponding "multiplication" operation. The most critical operation in the operation is "Distribution." .4.2.17.3. Category of English language representation of the scope of the figures Skilis Scholar category n The basic category is not that the object vocabulary N / S left deficiency subject and matrix (N / S) / N or N / (S / N) Left Left Subject, Right Leadyman Adjectives (Fixed) N / N Right Less Central Language Adjectives (S / N) / s Leave Leave "Middle Sentence Sentence Sentence] Advantages) (N / S) / (n / s) Right Less Central Language Advance (Decondency) (N / S) / (N / S) Leave Leave Central Propono Word (Decoction) (( N / S) / (n / s)) / n Right less messenger object (resembling the post rations) (N / N) / n Right less common objects N / N right side less words (main body) S / (n / s) The right side is less than the matrix (Bing) (S / N) / s Leave less "screwdriver sentence"
.4.2.17.4. Category calculation 1) The construction process of the sentence is the calculation process of the scope corresponding to the language component. 2) The result of a single word string or S, or not s, the former is a legal sentence, the latter is a non-legal sentence. 3) The specific operation of the calculus is divided into two: a kind of application ", discounted to A; a kind of" composition ", a brief note C. The application is a complete point, that is, the denominator is approximately, only the molecule is left as a result; for example: S / NN-> S Nn / S -> S-in-the-in-one achievement is that the category expression is still binding. For example: s / (n / s) (n / s) / n -> s / n.4.2.17.5. Category calculation example HE is a clever boy.
1) Category Dictionary HE S / (N / S) IS (N / S) / N / N / NClever N / NBOY N2) Step HE IS A CLEVER BOY.S / (N / S) (N / S) / N N / N N / N N --------------------> CS / N ------------------> CS / N ---------------------> CS / N ----------------------- ---> AS.4.2.17.6. Linguistic hypothesis 1) Assume that all structures are loaded by vocabulary, so that the scope of the various superiocal structure components can be derived from the vocabulary category. 2) hypothesis All combinations must be the combination of adjacent components, and it is impossible to have a leapfrogging of the adjacent component, so that the approval calculation of the adjacent relationship is achieved; 3) Assuming a strict sequence relationship, so that it can be found in the determination direction Objects of the decree; 4) Suppose each structure must be complete, so that a s mark is finally obtained at the end. .4.2.17.7. Problems in the category 1) Category mark and the word class are not one by one, and it is necessary to determine the specific words in the specific language to determine the scope of the specified difficulty, and even more difficult to understand. 2) Do not load the structure above the word (such as the combined structure in Chinese, the structure, etc.), it is difficult to incorporate the expression in the framework of the syntax. 3) The "Wang Yan" and "Father" in the "Wang Yan" and "Father" are unable to establish a decidism in the framework of the speech syntax. 4) Language in Chinese in Chinese, flexible, fills often incomplete (omitted), using a category grammatical description, will encounter a lot of trouble problems. .4.2.17.8. Reference 1) Lesniewski, 1929, Ajdukiewicz, 1935Bar-Hillel, 1950, Lambek, 1958, 1961, Montague, 1970, Zou Chong, 1995, Jiang Yan & Pan Haica, 1998, Bai Shuo, 1998 2) http: //www.cs.man.ac.uk/ai/cg/ .4.2.18. * DGDependency grammar, dependence syntax. The price standard grammar and the word syntax are two development of the grammar. .4.2.18.1. Tesniere dependent the basic framework association (Connexion): The dependence between the sentence components. Jonction: Squi Sproducts. Translation: Transfer of the word function.
The verb is the center of the sentence, the verb domains other components, itself is not dominated; the famous word group and adaptation of the verb domain; the noun phrase is an Actant; the adverb words are the state of the verb (Circonstant) The number of action elements is the valence of the verb; 4.2.18.2. Verbal action Yuan - price zero-priced verb: Do not dominate any action elements, "," a price verb: dominate a action element, " Swim "two price verbs: domain two action yuan," purchase "three-price verb: dominate the three action yuan," give ". 4.2.18.3. Robinson proposed four axioms 2) A sentence in one sentence Independent; 3) In addition to independent ingredients, other ingredients in the sentence must be contained in a certain component; 4) No component in the sentence cannot be dependent on more than two other components; 5) If the A component belongs to B ingredients, The C component is between the A and B in the sentence, then C ingredients or from A, or from B, or from a component belonging to A, B. Obviously, like chain grammar, dependence syntax description is the direct relationship between the words and words in the sentence, and this relationship is a direction (dominance - dependent, generally referred to the word word called "subject", "From the word"). .4.2.18.4. Depending on the tree and phrase tree
Dependency relation Tree Phrase structure Tree.4.2.18.5. Analysis example
Beijing is the capital of China. .4.2.18.6. Reference 1) Survival grammar is also known from the relationship syntax, earliest is proposed by French linguist Tesniere, its main thinking reflects "structure" in 1959 after he passed away. Syntax foundation "(Element de Syntaxestructrua). But he has put forward this grammar idea in the paper published in 1934 "Comment Construire Une Syntax).
. 2) Haim Gaifman, 1965, Dependency systems and phrase-structure systems Information and Control, Vol.8, No.3, pp304-337.3) Jane J. Robinson, 1970, Dependency structures and transformational rules Language, 46:. 259- 285.4) Gert-Jan M. Kruijff, Dependency Grammar, http://www.coli.uni-sb.de/~gj /lectures/dg.sslli/sslli-heads dependents.pdf5) http://www.cs .bham.ac.uk / research / conferences / Esslli / Notes / Hudson / 6) http://ufal.mff.cuni.cz/dg.html7) Feng Zhiwei, 1983, "Trinkiyeer Gender Syntax", "Foreign Language" 1983 No. 1 8) Peter Hellwig, 1986, Dependency Unification Grammar, Proceedings of Coling'86, Bonn.9) Wu Sheng (1992), Zhou Ming, Huang Changning (1994) 10) Michael A. Covington , A Dependency Parser for Variable-Word-Order Languages (1990), http: //citeseer.nj.nec.com/covington90dependency.html11) Dependency Grammar and the Parsing of Chinese Sentences, http: //arxiv.org/PS_cache/ CMP-LG / PDF / 9412 / 9412001.PDF12) Converting Dependency Structures to phrase structures, http://hlt2001.org/papers/hlt2001-14.pdf13) Towards a Competitive Dependency Grammar FormAlism, HTTP : //www.coli.uni-sb.de/egk/talks/rd_sem_ss02.pdf14) Functional Constraints in Dependency Grammar, http: //titus.uni-frankfurt.de/curric/gldv99/paper/LAI/LAIX.PDF15 ) Dependency grammars, http://www.vinartus.com/spa/94g.pdf16) a Fundamental Algorithm for Dependency Parsing, http://webster.cs.uga.edu/~jam/ACM-SE /Review/referee/mc.pdf17) Dependency Parsing with an extended finite state approach, http://acl.ldc.upenn.edu/p/p99/p99-1033.pdf .4.2.19. * Vgvalency grammar, price grammar.
.4.2.19.1. References. 4.2.20. WGWORD grammar, word syntax. .4.2.20.1. Reference 1) Richard Hudson, 1984, 1990, Blackwell2) http://www.phon.ucl.ac.uk/home/dick/wg.htm3)
.4.2.21. Slmstatistical language model. .4.2.22. Semantic .5.1. Today, the meaning of each word, the meaning of the word combines the meaning of the sentence, mainly refers to the meaning of independent of the context. .5.1.1. Development and status quo of semantic research. 5.1.1.1. Semantic studies in the Period of the Philology period, the earliest work is to note the ancient books. Language research is treated as tools for understanding ancient books and customs, habits, systems, etc. In Europe, the 3rd century BC, a batch of scholars were gathered in Alexandria and Belgamus, which were specially engaged in the work of argument, comment, etc., especially to organize Homer. Epic Eriat (ILIAD) and Odyssey. Their attention is concentrated on the grammar issues of the classics, and the meaning of the words. The semantic research in my Chinese literary period is called training. During the Spring and Autumn Warring States, the "Spring and Autumn" and "Gu Liang Chuan" were commented from the "Spring and Autumn" during the Spring and Autumn Warring States Period. Han Dynasty training, Zheng Xuan's "Book of Songs", "Zhou Li", "Gifts", "Book", and highly respected. In order to explain the word meaning of the ancient book, the Han Dynasty has compiled several important tools: "Er", "dialect" and "release". Eastern Han Xu Shen's "Said Words" saved a lot of ancient ancient training and ancient characters. After the Jin Dynasty, the focus of Chinese literature was turned to sound rhyme. The study of the study and the meaning of the study and the word of the Qing Dynasty, broke the blind manual. The focus of Ancient Greece, Rome, and Ancient Indian language is grammar, and it is very important to voice. Semantics is not their focus. And my country's training focuses on the relationship between glyphs and words. . Linguistics began to become an independent, there is a science of their own theories and methods. This period of semantics research has become an important part of the vocabulary, mainly studying the semantic part of the word collection. Traditional semantics research and clarify the relationship between a series of issues: (1) the relationship between the word meaning, voice, objective things; (2) the relationship between words and concepts; (3) The color of the meaning; (4) polysemy words , Homophone, synonyms, antonyms; (5) evolution, especially expansion, reduction, and transfer in evolution, and more. After the 19th century (including the 19th century) language teaching (national language and foreign language), the dictionary compiled, translation, etc. have a great progress. Traditional semantic studies have 3 weaknesses: (1) It uses the meaning as a whole, and does not break it into a smaller factor; (2) did not use the meaning as a system to study, but only simple Study the word "(3) it has not been studied in a larger semantic unit (such as sentence), and has been lying on the meaning. .5.1.1.3. Modern semantics of modern semantics, the creation of individual geography, is the 20th century, and gradually launched after the 1960s. The following describes the main genre of modern semantics: .5.1.1.3.1. A major task that is referred to as a language is to clarify the link between words and the world. When we talk, always use words to refer to things. This connection between the words and things is referred to the referous relationship. The word is called referring to the name (referring expression, also known as REFERENTIAL Expression, and some language books are submitted as R-Expression or R-Expression). The things referred to in words are called the refernes (REFERENT).
. 5.1.1.3.2. The most important contribution of structural semantics is the theory of semantic field. . .5.1.1.3.4. Generative semantics semantics are also generated. .5.1.1.3.5. Semantic theoretical syntax of Phil Mo. .5.1.1.3.6. Theory of the semantic theory verb of the husband. .5.1.1.3.7. Demoniology. 5.1.1.3.8. Scenario semantics. 5.1.1.3.9. Conceptual theory. 5.1.1.3.10. Preferred semantics. 5.1.1.4. Reference 1. Jia Yande, "Chinese Distance", Peking University Press; 2. Yan Guang, "Theory of Modern Chinese Monet Semantic Calculation Theory", Peking University Press; 3. Xu Lee, "Semantic". 5.1.2. "Significance". 5.1.3. Significance of meaning .5.1.4. Semantic knowledge. 5.1.5. Content (type) of semantic knowledge (type) Its formation representation. 5.2. Logical semantics (formal semantics). 5.2.1. LFL logical language LFL (Logical Form Language) is made of US IBM Watson Research Center M. McCord proposed, it uses predicate --- the way it means to indicate the meaning of the sentence. The Watson Research Center has applied this semantic representation to a university document database quiz system and an Yingde machine translation system. LFL expressions can be represented by items in the ProLog language and generated by the ProLog program. In applications such as database query, the LFL expression can be regarded as the target of ProLog and is executed directly by the ProLog system to generate the response required by the user. .5.2.1.1. Formation rules of logical (ie, LFL expressions) logical (ie, LFL expressions): 1. If P is an N eye predicate word of LFL, the variables x1, ..., XN is constant, variables, or logic, and p (x1, ..., xn) is a logic; 2. If P and Q are logical, P & Q is also logical; 3. If P is logical, e is a variable, then p: e (read "P" through E index) is also logical. Description: 1) In Rule 3, ":" is called an index operator. If p: e is a part of the logical Q, and E is referenced everywhere in Q, then E can be considered as a representative of P and its "speech". The speech here includes the alleged time and place for the relevant time and place in the natural language. For example, it is assumed that P represents event see (john, mary), then uses E to align the event in the entire logic through P: E. 2) There are two predicates of LFL: one is based on the meaning of the natural language, and the other is a non-word predicate for information such as the timeliness and tone of the sentence. For example: John Believes That Each Man Knows Bill. Èbelieve1 (John, Each (MAN1 (X), Know2 (X, Bill)). à Words predicate, Believe1 is a romance item, and manL is a righteous item of the noun "man", and so on. John SAW BILL. èpast (see (john, bill)). à non-word-predicate 3) Each LFL predicate has a fixed number of variables, which may be constant, variables, or other logic. 4) If there is at least one variable element in a predicate, we say this predicate is intersional.
A representative example is given below: John Only Buys Books At Smith's. èonly ((Book (X) & Buy (John, X)): E, AT (Smith, E)). 5.2.1.2. Reference 1. Shi Junyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principles", Tsinghua University Press. 5.2.2. Montague.5.2.2.1. History. 5.2.2.2. Reference 1. Course Note1, Montague Grammar, http://odur.let.rug.nl/~sjaak/semantiek3/courseNotes/3-1.pdf .5.3. Reading Notes .5.3.1. "How to know" Dong Zhendong, Dong Qiang, http://www.keenage.com/zhiwang/c_zhiwang_r.html
Knowing the network (English name "is a common sense knowledge base for the concepts represented by Chinese and English words to disclose the relationship between the concept and the concept and the concept of the concept. Common Sense Knowledge Base è Professional Knowledge Base: The Framework of Knowledge Base How to know the fundamental point of the language is: all things in the world (substances and spirit) are constantly moving and changing within specific time and space. They are usually from a state to another, and is usually reflected by a change in its attribute value. How to know the network operation and description of the basic unit: Wanwu (material spirit), component, attribute (quantity class, unit, ...), attribute value, event, time, space. How to understand the part of the part and function of things in its entirety in its entirety in its entirety. The relationship between the attribute and its host is fixed. Concept of commonality and personality. 16 relationships between concepts: 1) Upper, synonymous, antisense, regimening, related 2) Entity-value, overall-component, finished-material, host-attribute, attribute-value 3) event-role event- Time Event - Site Event - Tool Event - Ship / Experience / Relationship Subject - Benefits / Content / Acquisite
Chinese characters and simple words à
Cognitive Behavior: (1) Local Lenovo (Vocabulary Term) è Concept Status The origin of the local Lenovo clipphor is trying to form a expected and judgment capability so that the computer can implement a "bottom-up" (BOTTOM-UP) The understanding of the "top-down" is combined. (11) Abstract concept (111) Five-tuple (external performance of abstract concept): Dynamic V, Static G, Attributes u, Value Z and Effect R Any concept is needed to express from different sides, this phenomenon is called Conceptual diversity. Abstract concepts need to be expressed from three sides of dynamics, static, attributes, values, and effects. (112) Conceptual Semicity Network (Connotation of Abstract Concept) The expression of concept-related association is the primary goal of the semantic network. Semantic networks are typed hierarchical structures, and several nodes of each layer are represented by numbers. All nodes in the network can be uniquely determined by starting from the highest layer, and this number is uniquely determined by a string number ends to the node. The string is called hierarchical symbol. (1121) The primitive (φ): two major categories (11211) body primitive primitive (6 first-level nodes), process, transfer, comparatively, relational, state, and their constitutional utility chain. (11212) Composite primary concept (8 first-class concept nodes): Human activity physiological activities, general rational activities and social activities (1122) Basic (J): 9 first-level nodes and generalized space, time, space , Number, quantity and scope, quality and class, degree, objective basic attribute, basic attribute (1123) logic (L): two major classes (11231) Language logic concept (11 first-class concept nodes) quite The virtual word of Chinese. Semantic block distinguishing flag, semantic block combination flag, semantic block and inter-semantic relationship specification (11232) Basic logic concept (JL): 2 first-level concept nodes comparison, basic judgment (12) specific concepts (hanging approximation Methods) are divided into three categories, people, and physical properties (divided into separate symbols W, P, x). Basic (7 first-level nodes): hot, light, sound, electromagnetic, microscopic basic, macroscopic basic and life. Hanging against an approximation method: for a concrete concept, hanging on an abstract concept, plus some descriptions of the specific concept (13) concept, the three major semant networks are designed to express the connotation of abstract concept, and will ultimately use it to describe nature. The semantics of the language vocabulary, but the network itself is not directly oriented, but is a concept primitive for constituting the word quall. Any nodes on the network are themselves concept, they are concept primitives. They make composite concepts through a combination of different ways.
(131) Concept -> ((Category Symbol String) (Combined Sign Symbol) (Category Symbol Strings)) Category Symbol String -> (Category Symbol) * Hierarchy String -> (High " ) ((Medium) (under layer)) * | (body layer) (mid-layer) combined structure symbol -> (# | $ | & || --0 ... | * | | / |||| |! | ^ | () | (, LM,)) Category Symbol -> (V | G | U | Z | R | J | φ | L | S | f | q | h | w | p | x) Medium Layer -> (CNK | DNK | K | EMK | - | -0 | -00 ...) Hierarchical Symbol -> (0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | A | B | C | d) (132) Symbol Description: (1321) Category Symbol Set (13211) Five Words Symbols: V, G, U, Z, R (13212) Category Symbols of Abstract Concept: J, φ, L (13213) has Abstract concept of three major class abstract concepts: s (13214) "Syntax" concept symbol: f, q, h (13215) specific concepts and human symbols: W, P (13216), have abstract specific double characteristics Material concept: x Note: These 15 categories symbols are specifically used to express the concept of the concept, and cannot be used for variables of hierarchical symbols. They are the primitive representation of the concept category, where the primitive concept symbol φ can be omitted when specifically expressed. (1322) The definition of the high-level hierarchical symbol is shown in the J-table, φ table, L and JL table. The JW Table Hierarchical symbol is mainly used to express the three types of characteristics of the concept local Lenovo convolution: Contrastibility, concept of concept, Concept of inclusion. Contrast concept: CNK or DNK K = (1, N) coupled concept: K or EMK K = 0, 1, 2, 3 or K = 4, 5, 6, 7 inclusion concept: -, - 0, - The numbers "C, D" in the formula "are comparative signs," e "is a coupled logo (but the symbol is omitted for commonly used dualistic concept primitives). Comparative concepts have used two markers to distinguish the dominant properties, and the digital c represents the positive order, D represents the inward sequence. n represents the total level, k represents the value in the sort. The positive order is defined as a natural order, or the array of values is smaller from small to large, and the inverse sequence is defined as the value of the value from large to small. (1322) Concept Combination Structure Symbol Set: # Effects: $ Object: & Content: | Contains: - 0 ... Hand: * Expand: Positive: / Response: || Logic:, Logic Election:; Logic is not:! Logic Anti-parentheses: () General logic combination: (, LM,) (2) Global Lenovo (sentence, chapter level) è semantic block and sentence class (21) semantic block (semantic components of sentence) : The core part (also called statement element) Description section (211) constitution of semantic block
(212) The main language block ("indispensable"): Feature E, the rolers A, object B and content C. The naming of the e-semantic block is consistent with the name of the 6 links of the action effect chain, ie the X, Process P (Process), Transfer, Effect Y, Relational R (State). Effect chain: The role exists in the internal and mutual mutual effects. The role must have a certain effect. Before achieving the final effect, it will inevitably accompany some process or transfer. After the final effect, there must be a new relationship or status. The new effect will trigger a new role, so loop reciprocation, so endless. The main content of the natural language is to conduct local and overall specific expressions in these six links, and sentence class division is based on this. The active effect chain is both the basis of local Lenovo convolution, and the basis of the global Lenovo convolution. Two Lenovo convolutions are linked by it, so in a certain sense can be said that the effect chain is the theoretical basis of HNC. Generalized object language is a symmetric representation of A, B, and C semantic blocks. (213) Auxiliary speech block ("no"): Conditions, means MS (Means), tool in (instruction), WY (WAY), Refer to Re (Refer), due to PR (Premise) , Fruit RT (214) semantic block represents a semantic block letter definition category letter E, A, B, C; MS, IN, WY, RE, CN, PR, RT function letter x, p, t, y, R, S, D semantic block Number Defines the subclasses of the basic sentence classes after the different categories of the same category after the category letter, usually correspond to the second or even three level symbols of the main base concept node, for example: Mr. Zhang is afraid Miss Li loses his temper. X2B X2 XAC (22) sentence class (semantic category of sentence) is only expressing a clause class of a link to the effect chain, referring to the basic sentence class, expresses the clause class that expresss two or more links to the hybrid clause. The so-called hybrid classes refers to two basic sentence classes in one sentence, such as a functional sentence, process transfusion sentence, status judgment sentence, and more. The sentence consisting of the e-semantic block, named the function, the process, the transfer, the transfiguration, the effector, the relationship sentence, and the status sentence. According to the six sentence classes defined by the action effect chain, the judgment sentence is formed, and the seven basic sentence classes of the HNC are constituted. The knowledge base of sentence class analysis includes the concept level, vocabulary, statement level, and contextual knowledge. The knowledge of these four levels should be centered on the statement level and named sentence class knowledge. (221) Basic clause: Seven, named: scaff, effector, process sentence, transfer sentence, related sentence, status sentence, judgment sentence. The corresponding symbols are: x, y, p, t, r, s, d. (222) Hybrid class: refers to the mixing of two basic sentence classes. (223) Compound clause: Note There are two or even multiple E blocks in one sentence, and they contain information on different aspects of the effect chain. (224) Class Class Form: Sort by the main language block. J Table: Basic Concept 1, Second Level Concept Nice Table
Node Meaning Description J0 Order and Generalized Space J00 (General) J01 Generalized Space (Generalized Location and Generalized Direction) J02 Generalized Distance J1 Time J10 Time J11 Time J12 Time Interval J2 Space J20 Space J21 Space J22 Space Distance J3 J30 Basic number J31 Space J32 number of conversion J4 and range J40 All, topical, individual (inclusive) J41 (single, double, half; less) J42 range (boundary, internal, external; cross ,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,:: J71 couplet, confrontation J72 mainly with the secondary J73 special, general; precious, ordinary J74 nature, appearance J75 relative and absolute J76 conventional and abnormal J77 simple and complex; pure and mixed J78 new and old J8 basic attributes (including subjective evaluation ), In general, the ethics J80 is working with evil J81 true and false; real and virtual J82 good and evil J83 beauty and ugly J84 pairs and error J85 is with non-J86 positive and negative
φ Table: The primitive concept 1, the secondary concept node table (the φ φ φ φ φ) body primitive concept effect effect chain 0 action 008 Physical action 009 Activity 02 Active Reaction 03 Effective Reflection A special form 04 constrained action A special form 1 Process 10 General Process 100 The basic feature 1008 process of the process 1009 process 100A process 129 motion process 10a evolved process 10B life process 11 process 1 process 12 process And the tendency of the cause and the tendency of the source, the exchange, and the flow 13 process, the metabolic 2 Transfer 20 General Transfer 200 Transfer (one of the basic features) 204 Transfer of the transfer (Basic Features 2) 209 Direction transfer 20A Transmission 20B itself Transfer 21 Receive 219 Receiving 21B information of receiving 21B information of 21A information Transfer 23 Information Transfer 24 Exchange 24 Exchange, Alternative and Transform 3 Effect 30 General Effect 300 Effects Basic Features 309 Implementation and Completion 31 Generation and Elimination 32 Harm 33 reveals and hidden 34 expansion and reduction 35 vertical and broken 36 promotion and inhibition 37 interruption and communication 38 selection, save, abandon 39 composite 3A obtaining and pay 3B accumulation and consumption 4 relations 40 general relationship 400 Basic composition of the relationship (aspect) 401 I am 402, 403, 404, 405, 408, 408, 408, 409, 409 Relationships (Distance)
40A relationship 41 binding and separation 42 dependent and exclusion 43 Support and opposition 44 dominance and slave 45 with abandon 46 with loss 47 adaptation and interference 5 state 50 general state 500 state 508 Natural state 509 life status 50A Human State 51 Morphical 52 Dynamic 53 Potential 54 Structure 55 Hierarchical 56 Grade Composite Base Concepts Mainly for Human Activities 2 Physiology and Energy Activity 7 Psychological Activities and Mental State 70 Psychological and Spirit 71 Psychological Reaction 710 General Psychological 711 Attitude 7110 The attitude of the relationship between 7111 emotional color is 3112 attitude 7113 to the career 7114 attitude 7115 attitude 7115 to interpersonal relationship 7116 Attached to general things including public welfare activities 7117 At the attitude of close relationships 712 Wish 7120 General Wish 7121 The desire to be achieved 7122 The desire to achieve 7123 To achieve the desire to achieve 713 Emotion 7130 General Emotion 7131 Happy 7132 Worry 7133 Fear 7134 External Trigger-based emotional performance 7135 Love 7136 evil 7137 hate 7138 Because the Lord's psychological reaction 7139 Remendant 713A regret 713B 7 714 Emotional and mental state 72 mental state 720 General Mental State 7201 Will 7202 Capacity 7211 Congenital Dental Selection 7212 After Day Dental 722 Tsemmetric 7221 Character 7222 Bing Sex 8 Thinking Activity 80 General Thinking Activities 81 Cognition and Understanding 810 General Understanding and Understanding 811 Analysis and Comprehensive 812 Deduction 813 Judgment 82 Exploration and Discovery 821 Exploration 822 Discovery 83 Planning and Design 831 Planning and Plan 832 Design and Planning 84 Evaluation and Decision 841 Evaluation 842 Decision 843 Provisions 844 Agreement 845 Recognized and Recognizing 9 IT Activity A Professional Activity A0 General Professional Activity A00 General Professional Activity A01 General Organization Activity A02 Implementing A1 Politics A10 System A11 Organization A12 Governance and Management A13 Political struggle A14 Diplomatic Activities A15 Conquest and anti-regular A2 Economic A20 General Economic Activity A21 Industrial A22 Commercial A23 Service A24 Financial A25 Countries
Economic A26 Technology A27 Natural Economic A3 Culture A30 General Cultural Activity A31 Literature A32 Art A33 Technique A34 Military A40 General Military Activities A41 Organization A42 War A43 War Effect A44 Military Action A45 Military Technology A5 Law A50 General Law A51 Legislative A53 Legal Difference A53 Law and Ethics A54 General Law Enforcement A55 Procuratorate A56 Judgment A57 Implementation A58 People A59 illegal A5A Relationship A5B Relations People Reaction A6 Technology A60 Philosophy Exploration A61 Research Activity A62 Technology Activity A63 Natural Science A64 Humanities Science A65 Theory Science A66 Experimental Science A67 Science and Technology and Productivity A68 Discipline A7 Education A70 General Education A71 Education A72 School A73 Examination A74 School Education A8 Health A80 Life Sanitary A81 Check A82 Governance A83 Raise A84 Environmental Sanitary A85 Global Health A86 Local Sanitary A87 Local and personal health B pursue activities B0 pursue B00 rational pursuit B01 action pursuit of B02 accompanying battle B03 fight against fate B1 reform B10 general reform B11 overall or fundamental reform B12 Local reform B2 inheritance B20 general inheritance B21 development inheritance B22 Negative inheritance B23 local inheritance B3 competition B30 general competition B31 attack B32 B33 challenge and should fight B4 collaborative B40 general collaborative B41 strategic collaborative B42 tactics collaboration B43 actively collaborative C Social activities D regiment Activity (behavior) D0 behavior D00 General behavior D01 behavior and rational D02 big public behavior D03 selfish behavior D04 to others D1 concept D10 general concept D11 fundamental and global concept D12 point of view D2 guidelines D20 general code of conduct D21 Social Guidelines D22 Communication guidelines D23 Bunch of guidelines D24 Guidelines for others L and JL Table: Logic Concept 1, Secondary Concept Node Table
Language logic concepts substantially correspond to Chinese virtual words L0-3 semantic block distinguish markers Mainly corresponding to the word language in traditional ling, but it is difficult to indicate the meaning of the meaning of the traditional concept and the language logic in different languages Single word L0 main language SLR L00 feature L01 Role L02 Objects L03 Content L1 Acupuncture block single flag L11 method L12 Tool L13 Way L14 L15 Conditions L16 Motion L17 L17 L17 LED L2 Primary Bills Match Monograph L20 Features L21 role L22 object L23 content L3 Acupuncture block with flag L31 method L32 tool L33 pathway L34 L35 Conditions L36 Motion L37 destination L4-5 Semantic block combination flag corresponds to a variety of word sex words, including relational pronouns and partial conjunctions, Words and the traditional concepts of traditional concepts L4 semantic block combination flag L41 Positive L42 reverse partial L43 and L44 or L5 semantic block combination Description L6-a semantic block Description L6 Times Description L7 Temporary L8 Acupuncture Declaples L9 Refers to Logic LA E Elements Logic Description LB Statement Logic Description Basic Logic Concepts Basic Logic Concepts Basic Logic Concepts Comparison of JL02 and One Comparison of JL02 and One Standard Comparison JL1 Judgment JL11 judgment JL111 is, affirm that JL112 No, no JL115, there is a potency of JL116 without JL12 judgment (pure objective) JL12C31 may jL12C32 can jL12C33 inevitable JL13 judgment potential (including subjective) JL13C21 should, should be JL13C22 must, must JW Table: Basic Concept First, Second Level Concept Nice Table
JW0 hot JW1 light JW2 sound JW3 electromagnetic JW4 microscopic basic material JW5 macro base JW51 gland JW518 atmosphere JW52 liquids JW528 water JW53 solid matter JW538 Soil JW6 life JW61 plant JW62 animal JW63 human JW61-plant parts and tissue JW62-animal components And organize JW63-body parts and organizations
S Table: S11, Second Different Typography S12 S11 Specific Ways S12 Strategy S2 Means S21 S22 Method S3 Conditions S31 Time Conditions S32 Space Conditions S33 Social Conditions S34 General Substances S35 Prerequisites S4 Tool S41 Specific Tool S42 Materials S43 raw material S44 energy
.5.3.3. "Chinese grammar integration network" Luchuan Business Press
.5.4. Reference
6. Pragmatic .6.1. How to use different situations for different situations, how different usage affects it. .6.1.1. How to influence the knowledge of the next sentence affects the interpretation of the next sentence, it is especially important for the interpretation of the pronouns. . 6.1.2. The user of the world's knowledge language is a general knowledge that must have the world, including language users, to understand what language users have to understand other users. .6.2. 7. Application. 7.1. Semantic understanding of applications in information processing. 7.1.1. System model
Question 1. What is the concept system (or conceptual framework), how to organize, is it effective? 2. How is the article mapped into a conceptual system? 3. How is user interest to describe the concept system? 4. How is the concept system? 8. Prototype system construction automatic collection and corpus generation system Constructs self-reset English Chinese dictionary (correspondence relationship between the word number and the word) Constructing the word dictionary (correspondence of the word number and the word) Constructive name, place name, historical event dictionary structure Dictionary (one of the words belonging to which discipline) Semantic dictionary of the texture of the texture of the textbook
Jinshan Word Terms WordNet9. Topic. 9.1. Let the machine understand the philosophical thinking of natural language to let the machine understand the person's language, first must answer such a question: What is the language? In other words, what is the mechanism of human language? We must say that the language is a symbolic representation. This can be confirmed from the following phenomena: spoken, written, and body language are languages, and the speech can be sound symbol, or may be image symbols, or even behavioral symbols. In addition, early Egyptian text, Babylon text, and Chinese text are like a hieroglyph, which also reflects the language from one aspect. The language is a symbolic system. Now the language is a set of abstract symbolic systems, even the initial pictographs are no longer needed. So what should this symbol system say? What is it? The author believes that this symbolic system is to represent the objective world of reality and people's understanding of this objective world. This should be understood from the following aspects: First, the objective world is the body to be described by the language; but the world described in the language is not equal to the objective world. It is the perspective of the person, it is the world recognized by people; Second, people's understanding of the objective world, both people's perception of this world, and some people think about this world, and also add to people's emotions. So the problem becomes a classic philosophical problem. But the author does not want to discuss the relationship between material and consciousness, thinking and existence. The author is interested in the objective world à people's perception à people 'thinking such a chain. In other words, since the language is the objective world of the reality and the symbol of the people's understanding of this objective world, what is the objective world in the human eye? How do people know the objective world? How is the language of the language and people's understanding of this objective world? Let us concentrate the history of human history and people's growth history, you can get a sense of inspiration: I wow, my brain is blank. I can see that I can hear, I can touch it, I can smell, I can taste, one sentence, I can feel it. And my vision, the instinum feels that the light is strong (brightness), the color is very shallow, the distance is far close, the orientation is different. And these are instinctive feelings, that is, the objective world is stimulating to me. But I don't know how light, color, distance, orientation. Similarly, I can hear the sound, and the sound is strong, and the feeling of sound generation; I touched, let me have a temperature and smooth feelings, there is painful; I smell the smell; I tasted the taste. In addition, there are two things that are neither. One is a sense of sense. This feeling creates a time concept, this is an aftermath; the other is a sense of interest, or says good misma. This feeling, it seems that Darwin's biological evolution can be evidence. And the sense of interest is derived from the complex feelings of human beings, which is also after. These above constitute human perceived ability.
I still imagine a blank in my mind, except for the above feelings, there is no concept. I see people, see the house (of course, I can't see the house, but I see the sky, the wood, or the cave, I heard people's voice, bird's voice, I wrote Go to the scent, I feel the relatives to my touch, etc. Obviously, I can feel that this is different, and I don't want to mix the cave with birds. Therefore, the difference and segment is one of the characteristics of human thinking. Imagine, I am still a baby, every time I see a visual image of my mother, others have issued a "mother" sound, so I also sent a "mother". I will contact the mother's visual image with "Mom" in the future. That is, when a person came over, I first took her with my mother in my mind. If I agree, I recognize that my mother's visual image is "Mom". I said that it is obviously assumed that I have a memory ability. Then there is another woman A, people let me call her "aunt". So every time A, I called her "aunt." Although I don't know what "aunt" means, or say "aunt". Anyway, I saw she and found it as "aunt." In the future, there are other women B and C come. People also call me "aunt", I just call, I don't feel anything wrong. Until one day, I came again, but no one told me what to call her. At this time, I just thought: I called her "mother" or call her "aunt"? Maybe you will ask why not call "Sister"? Sorry, there is no sister in my mind. Maybe I will call her "mother", maybe I will call her "aunt". When I called her mother, I told me: "No, call her aunt". "Why?" "Because you raise you, you are your mother." So "Mom" and "Aunt" are subdivided and the difference. Later, "Aunt" is collectively referred to as "aunt" because of the comparison of "grandmother" "sister" "aunt" and other concepts, or because there is a common feature This is clustering. So, memory, identification (or symbol representation), pattern matching, subdivision, classification, cluster constitutes the characteristics of human thinking. First, human beings determine each individual, the new individual is always mode-matched with individuals in memory until it is divided into one individual, because of the difference, because of the common point With clusters, because new categories and clusters have been reassigned, there is a new pattern matching. In this way, on the basis of humans perceive the objective world, constantly identify objective world and people's perception, pattern matching, subdivision, classification, and clustering, the objective world in the human eye. The author believes that the contradiction in philosophy is another description of classification and clustering.
The above is the basic part of the perceived ability and thinking ability. Below I will explain how these components constitute a large unit. First, let me say how to make a sense of space and sports concept. Since people's vision can sense light, color, distance and orientation, distance and orientation, resulting in the length of the length, the concept, the orientation and shape of the shape, the orientation, and the shape of the shape. And what? Movement can be said to be ray and color changes, or changes in object space. Time and space, exercise and stationary, the concept of these classic philosophy is such a truth. A special movement along the time axis is life and death. Let's talk about this important concept. The orientation of time and space constitutes the order. Let me talk about a very famous thinking method in thinking: inductance. From individual real instances, all examples that grow out from these individual instances are true (thus achieving clustering) according to the growth conditions (common attributes on the basis of classification). The authors believe that all of human thinking ability, eventually can be treated as subdivision, pattern matching, classification, clustering on the collection. To this end, it seems that we have answered how people perceive the world and how to think about the world. The language is to record these and make the symbol of the system. The following references the two words in "Semantic" further explain the relationship between language and objective world and human cognition: "The language is a single concept system, how many languages have a concept system? Although the current views of many people tend to assume that there is a common concept system in all human language, most observations have shown that various languages are different. "If you don't recognize a concept system in human language It emphasizes that various languages are different, then "all languages have strongly add their own 'bar frames' to experience, or for a metaphor, that is, the language provides a set of 'category', we basis These 'sortegers have aware of the awareness of the world. "Today," "The language of a person uses to largely affect his thinking process and understand the objective world." Obviously, according to our Views, various languages have a conceptual system, the reason why the various languages are classified, because people are determined, subdivided, classified, and clustering in various languages. Said so much, let the machine understand the language, what should I do? First of all, we have to determine how many concepts have been identified, and we must ultimately form a concept system; secondly, we have to form a new concept of generating system, so that our knowledge system is an open system; and this new concept The formation of the new concept is the process of pattern matching, subdivision, classification, and clustering. Third, on the basis of two, we map the language to our concept system. The understanding and processing of the language is turned into model matching, subdivision, classification, and clustering of the concept system.
.9.1.1. Reference 5. Russell, "Western Philosophy History" 6. Feng Youlan, "Simple History of Chinese Philosophy" 7. "Marxism Principles", Higher Education Press 8. Shi Zhongzhi, "Senior Artificial Intelligence" 9. Shi Chunyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principles", Tsinghua University Publishing House 10. Frei. Jiqi, "Semantic", Shanghai Foreign Language Education Publishing House; 11. Yu Jiangsheng, "Philosophy of Language", http://icl.pku.edu.cn/yujs/papers/html/what's Linguistics.htm 12. Yang Xiongi, "Brain Research Changes to the World", http://www.people.com.cn/gb/guandian/183/8778/8779/20020812/798217.html .9.2. Natural Language Understanding and Artificial Intelligence. 9.2.1. AI knowledge representation and reasoning (19991211) predicate restriction and generating system are above the formalized description of the abstraction of the rules system, expressing human thinking can perform strict reasoning. They are suitable for strict reasoning for limited space, and the heuristic search function can reduce the search space. Human thinking is faster than computer computing speed, but can find the best or preferred way to find the same problem while solving the same problem. è1) Humans have several parallel processing units, each unit's performance is not very good, but the overall performance is very good; 2) Various solution methods are different from the degree of correlation in the brain and their respective storage levels. When people mention Dad, the most likely that mothers, children, grandfather, etc., usually don't think of tiger. è is strong in logical relevance, and is also strong in the physical storage of the human brain. Therefore, the human brain has a certain connection mechanism. Semantic network, framework, script reflects the associative ability, classification capacity, inheritance capacity, and object-oriented ability, but the association is limited, and only part of the relevance is described. Human thinking also has memory capabilities, and memory ability has mechanical memory and meaning memory, mechanical memory children are good at, meaning memory adults are good at. The memory time has a length, which involves the strength problem of memory. è memory time length: Memory intensity memory capacity After learning a certain course, after two or three years, I forgot. But when you re-hack, it is easy to get started and it is easy to recover. è memory is intensity instead of capacity. Memory has strong survivability. The human brain has a "forget" "abandoning knowledge". Attachment: Human also has the following ability to understand: attention and observation; on the other hand, human beings also have practical ability, including self-learning ability, expression, communication skills, competitive capacity, self-control, creativity, operational ability, and adaptivity. Human learning knowledge and can extract core knowledge from many knowledge, organize and express. è Lenovo: 1) Relevance of things; 2) Trueability of things; 3) Correlation> = a, store the correlation, otherwise the relevance is stored, otherwise there is no correlation . .9.2.2. Investigating AI (20000307) 1 from the perspective of various disciplines. Each discipline 1) Mathematics à logic, proposition, <= summary, anti-regular law, succession (intuition) => forward reasoning, reverse reasoning, two-way reasoning humans in more time is two-way reasoning.
2) Data Structure à Application 3) English Word: Word Missing Analysis, Synonyms, Synonyms, Antonym French Language Exercise 4) Physics, Chemistry: Experiments à (induction) à Guiji à Empirical, Argument 5) History: History à (Overall Feeling) à History Laws 6) Philosophy, Economics: Agreement for some characteristics in real life è a) boldly guess, careful attention è b) In human thinking, logic does not occupy all, but only small parts. Incident and the agreement of the results of the summary, it is mostly occupied by the guessment results. è Logic <- (Deterministic | Uncertainty) -> The most current research is form logic, not semantic logic 2. Human intelligence is 3 of history. The dependence of AI is mainly based on the utilization of knowledge, and uses existing knowledge to analyze, guess, judgment, prediction, etc. 4. Inspiration, subconscious problem. .9.2.3. Logical thinking is a habitual approach to cultivating problems. Limited logic: From some facts, it can introduce an object with some nature P, which is all of the objects that meet the nature P.
Default logic: Under normal circumstances or typical case, assertion A is established, then default in all cases, A is established. Correct when an antique appears.
Based on example learning reasoning: original example? (Memories, derived) à target example
Problem à Analysis problem (==> ") à solve the problem (==> guess argument)
Incident, the class ratio à guess à argumentation. 9.2.4. At the stage Ai research characteristics (20000317)
Predit, generated: Process default reasoning semantics network, frame: structure (logic) imitation people's thinking neural network: structure (physics) induction
Consciousness: Reflection of human brain on objective things ->? It can use intelligent activities to start
"Only can not be swimmed"
The characteristics of the human brain: 1. Usually thinking is not logical, but jumping, the disadvantages of the informality: 1) Slow calculation; 2) Shallow depth and low. Memory; 3. Imagination, image thinking; 4. Time and space; 5. People's thinking can be cultivated, can be trained (playing chess)
At present, AI has been imitating individual properties of people's thinking.
People's intelligence has its limitations, and the current research has exceeded people's intelligence range.
Prediction statistical analysis ==> Conclusion
Several principles thinking about WHO, WHEN, WHERE, WHAT. Natural language understanding (20000307) 1. NLP is a huge project; 2. NLP learning mechanism (20000407) Knowledge learning is a step-by-step process 3. (20000407) NLP, speech recognition, MT, ML (induction, class ratio, explanation, etc.), DW & DM, CBR4. Whether human beings have heard that they are very natural, they must be familiar. If people speak, otherwise, people generally don't understand, to give explanations. .9.2.6. My natural language is understood (20000317) 1. Taking noun, introducing object models (inheritance, development, coverage, etc.); Taking the focus of semantics, introducing metimony, synonyms, synonyms, antonyms; 3. The semantic network is semantically expressed, introduced in semantic generation rules; 4. Have learning mechanism, open, fault tolerance, and intelligence; 5. Memory model and associative model; 6. Really grammar analysis is the secondary, semantic analysis is mainly, form is content services; 7. Object Method 1) Main object (Feature Description): Attribute behavior; 2) Event object (WHEN, WHERE, WHAT, HOW, wh). 8. (20000505) Semantic understanding - starting from the virtual word (helid, conjunction), starting with "people". .9.2.7. Human language mechanism: Naming and identity mechanism (20000421) 1. Naming and identification mechanism of entities name "people" this entity, but "people"? Since ancient times, no one is difficult to express in words. 2. Named fuzziness, uncertainty and fine division of "computer", I think of a machine when I high school is only; but I know it is big or different; I see "people" concept, I think "Alien", man, woman; cry -> Weeping, sob, big bitterness, crying, crying 3. The content of the language is, for example, "national feet" 4. The form is a content service, grammatical serving Chinese early syntax studies. There is no word language, and the language is strengthened. From a certain extent, grammar is a summary of a semantic basis. .9.2.8. World 1 in the human eye. Composition; Is it useful; 3. Is it beneficial or harmful?
.9.2.8.1. 10. Related resources 11. Reference 4) Bai Shuo, "Computational Lingu", Chinese Academy of Sciences Research Institute 5) Zhan Weidong, "Introduction to Computational Lingu", Peking University Chinese Department 6) Li Tang Qiu "Natural language processing" lecture, Xiamen University Computer Science Department 7) Shi Chunyi, Huang Changning, Wang Jiaqin, "Artificial Intelligence Principle", Tsinghua University Press 8) Chris Manning and Hinrich Schutze, Foundations of Statistical Natural Language Processing, HTTP : //www-nlp.stanford.edu/fsnlp/9) DR TED BRISCOE, Nature Language Processing Course Handout, http://www.cl.cam.ac.uk/teaching/2000/natlangproc/handout.ps 12. Reading notes .12.1. Experience 2) Problems and background, research history, research status, latest research progress, application; 3) solution, ideas, theoretical basis; 4) What are the advantages, new, disadvantages, limitations ;12.2 Article. 12.3. Summary sporadic also read a lot of articles, mostly because of the patient, but did not see it, so I can't talk about it, but I have an impression. If the machine learning, pattern recognition, knowledge excavation is the underlying basic theory, and information classification, information retrieval, information filtration, information extraction is relatively high-level theory, and these four are complementary .