Principles of compression and decompression within the form of multidimensional language representation applied to educational tasks generation
Автор: Lichargin D.V., Safonov K.V., Nikolaeva N.V., Chubareva E.B.
Журнал: Сибирский аэрокосмический журнал @vestnik-sibsau
Рубрика: Математика, механика, информатика
Статья в выпуске: 4 (44), 2012 года.
Бесплатный доступ
In the given work the problem of semantic decompression patterns addition to the meaningful sentences generated as functions in multidimensional vector space of the notions of the natural language is considered. The model of adding semantic decompression patterns based on stylistically-oriented sets of generative grammar rules is offered. The complexity of natural language translation problem is discussed. The considered model provides the algorithm of adding semantic decompression patterns that can be used for improving the performance of educational tasks generation software.
Natural language generation, meaningful sentences generation, turing test
Короткий адрес: https://sciup.org/148176911
IDR: 148176911
Текст научной статьи Principles of compression and decompression within the form of multidimensional language representation applied to educational tasks generation
At the present moment, people are facing a huge amount of information that is not always well absorbed and efficiently used because of the complexity of its structure. Presenting the language as a model of a multidimensional data set can improve the quality of linguistic software. A multidimensional view of data on natural language is important for the construction of electronic translators, abstracting systems, expert systems, generative grammar, etc. In this regard, the analysis of a multidimensional model of language data is relevant at the present stage of development the information technology and mathematical foundations of computer science.
The problem of formal modeling of natural language, particularly, English, is the central task for computational linguistics – a discipline that lies at the intersection of computer science, mathematics, systems analysis, linguistics, philosophy, psychology, etc.
Solving the problems of developing linguistic software successfully implemented numerous theories, concepts and software systems. Numerous works in the field of semantics, discrete mathematics, linguistics and Artificial Intelligence, let people hope for solution many of the problems of formalization of natural language and passing the Turing test in increasingly tough conditions for the test systems in the near future.
For solving the problem of generating the meaningful speech a lot of tools are used today by both Semantics and Artificial Intelligence within the notional apparatus and the various models of mathematical Semantics. In particular, the analysis of natural language was traditionally applied within the following models and tools such as the method of ontology, the method of linguistic classification, the method of multidimensional data, OLAP systems, relational database, frames, generative grammars, in particular, generating Montague grammar, semantic network theory graphs and the resolution method, hybrid systems, and linguistic methods, such as component analysis, the paradigmatic method, the approach of American structuralism, etc [1–8].
There is a topical problem of generating meaningful subsets of the natural language with various approximations. The solution to this problem greatly simplifies tasks such as the construction of expert systems, e-learning systems, automatic transfer systems, programs to support dialogue with users, creating naturallanguage interface. The solution to this problem is mainly determined by the problem of passing Turing test by software systems, providing identity and the inability to distinguish a dialogue with a person and a dialogue with a software system.
The novelty to offer a classification of generative grammar rules or relational patterns subsets, based on the multidimensional model of the natural language based on the proposed vectorized semantic classification of words and notions of natural language.
The main idea is to view the grammatical and lexical spaces sequence linked by the generation process according to this or that generative grammar rules subset or relational patterns subset. The idea can be applied to solving the task of automatic educational tasks generation algorithms creation.
In the work of D. V. Lichargin “The Methods and Tools for the Generation of Semantic Structures is the Natural Language Interface of Software Systems” and “A Multidimensional View of Data on Vocabulary and Grammar of the English Language” the following model of the meaningful natural language generation is offered.
One can specify the states space for such units of the natural language, as words and notions. The space of the grammar of language is described by the space coordinates (Fig. 1, 2):
– Members of the Sentence (Subject, Predicate, Object, …);
– Parts of Speech (Noun, Pronoun, Verb, …);
– Grammar Categories (Plural, Collective, Superlatives, …).
Next, we construct the lexical space of words of the language (data cube) with the following coordinates is presented:
– Word Order (Doer, Action, Receiver, Property of the Receiver, …);
– Topics (Food, Clothes, Body, Building, Money, …);
– Options for Substitution of Words in a Sentence (Possitive, Neutral, Negative; Maximal, Large, Medium, Little, Minimal, …).

Fig. 1. The Lexical and Grammatical Space in the Model of Meaningful Language Generation

Fig. 2. The principle of Lexical Space – Grammatical Space transformation
Methods of Pattern-Based Semantic Decompression patterns Adding and Text Composition / Decompression
Doer |
Action |
Object |
Substance |
I |
Eats |
The … |
With / without … |
We |
Cooks |
Dish |
Beef |
Bob |
Roasts |
Potatoes |
Fish |
I |
Sews |
The … |
From … |
They |
Knits |
Jacket |
Wool |
The girl |
Irons |
Shirt |
Cotton |
Subject |
Predicate |
Object |
Modifier |
My DOER’s |
ACTION.MAKING-ing |
Needs / requires / … |
Good / nice / … + SUBSTANCE |
My cook’s |
roasting |
needs |
(good) beef |
My mother’s |
sewing |
refers to |
(brilliant) silk |
This / the / the given + SUBSTANCE |
Is good / nice / ideal / … for |
For my / his / her / … + DOER |
To + ACTION |
Silk |
Is good for |
My mother |
To sew |
Fish |
Is ideal for |
My brother |
To cook |
ACTION-ing |
Cannot / will not + go on / continue / be done / be all right |
– |
Without + such / this / … > like this + SUBSTANCE |
Cooking |
Will not be done |
– |
Without beef |
Knitting |
Will not be all right |
– |
Without wool |
In this respect, when addressing the problem of real texts processing in a particular language as a part of program – generators and analyzers of meaningful speech, it is necessary to solve the problem of removing the semantic decompression patterns. Semantic decompression patterns are present in the texts written in different language styles: from the academic style with a low degree of semantic decompression patterns up to slang with an extremely high degree of semantic decompression patterns.
Semantic decompression patterns are the formal interpretation of emotional expression and depth of the subject. The computer, in particular, can consider it to be unimportant for the task of generating any units of the natural language: words (such as neologisms) sentences and texts.
For the generation of statements with semantic decompression patterns one can also use the method of generating semantic decompression patterns by analyzing the semantic structures of notions and their transformation. For example, the word “to like” corresponds to the vector of coordinates:
[RELATION–CREATURE–X \ ESSENCE \ POSITIVENESS].
The word “beautiful” corresponds to the vector attributes:
[RELATION–X \ ESSENCE \\ RELATION– CREATURE–X \ IDEA \ ON (NOT) LIVE \ POSITIVENESS].
The word “to look” corresponds to the vector of semantic features:
[RELATION–CREATURE–X \ ESSENCE \\ RELATION–CREATURE–X \ IDEA \ ON (NOT) LIVE \ POSITIVENESS].
As a result, it is possible to take an opportunity to regroup the semes of the natural language within a semantic web for each of the word. For example, the phrase “the apple is beautiful” can be transformed into the phrase “I like the form of the apple”. In this case, the concept of “beautiful” falls into the group of semes with the meaning “to see” and a group of semes with a value of “good”, “love” or, for example, “nice”.
Based on the offered model the following scheme of natural language generation is offered (see the table).
The generative grammar or relational patterns, used for sentences / text composition and decompression and semantic decompression patterns addition, can be subdivided into stylistic subclasses like the ones below.
Below a general scheme of reducing a complex literary phrase to a simplified synonymic equivalent by adding logical, semantic, grammatical, morphological, sound and other noise addition patterns / decompression patterns is presented (Fig. 3).

Fig. 3. The decompression patterns tree going through semantic, grammatical, morphological, phonetic multi-attribute descriptions – multidimensional spaces
This approach supposed to be used for creating different kinds of linguistic ware, for example, summarizing systems, electronic translation systems, expert systems, natural texts data extraction systems and others.
Let’s view some principles of translation of a complex literary text from the point of art – semantic noise computer interpretation. Translation theory, based on a solid foundation on understanding of how languages work, translation theory recognizes that different languages encode meaning in differing forms, yet guides translators to find appropriate ways of preserving meaning, while using the most appropriate forms of each language.
Translation theory includes principles for translating figurative language, dealing with lexical mismatches, rhetorical questions, inclusion of cohesion markers, and many other topics crucial to good translation, which are supposed to be viewed as lexical, grammatical, morphological, phonetic categories multidimensional spaces data presentation as multilevel means of compression tree generation for example.
-
1. “The driver carries the cans …” – Водитель привозит консервы … .
-
1.1. “The cans carried by the driver …” – Консервы, привозимые водителем … .
-
1.2. “The driver, who carries the cans …” – Водитель, который привозит консервы … .
-
1.3. “Carrying the cans by the driver …” – Перевозка консервов водителем … .
-
1.3.1. “Carrying the cans from the side of the driver …” – Перевозка консервов со стороны водителя … .
-
1.3.2. “Carrying the cans by the efforts of the driver …” – Перевозка консервов усилиями водителя … .
-
1.3.2.1. “The process of carrying the cans by the efforts of the driver …” – Процесс перевозки консервов усилиями водителя … .
-
1.3.2.2. “The task of carrying the cans by the efforts of the driver …” – Задача перевозки консервов усилиями водителя … .
-
1.3.2.3. “Carrying the goods, for example, cans by the efforts of the driver …” – Перевозка товаров, например, консервов усилиями водителя … .
-
-
-
-
1.3.2.3.1. “Carrying the consumer goods – cans by the effort of the driver …” – Перевозка потребительских товаров – консервов – усилиями водителя … .
-
1.3.2.3.2. “Carrying the goods – round cans by the effort of the driver …” – Перевозка товаров – круглых консервов – усилиями водителя … .
-
1.3.2.3.3. “Carrying the goods – metal cans by the effort of the driver …” – Перевозка товаров – металлических консервов – усилиями водителя … .
It is also important for human translation process [9; 10] and teaching future translators and interpreters. The translator should understand perfectly the content and intention of the author whom he is translating. The principal way to reach it is reading all the sentences or the text completely so that you can give the idea that you want to say in the target language because the most important characteristic of this technique is translating the message as clearly and natural as possible. The translator should have a perfect knowledge of the language from which he is translating and an equally excellent knowledge of the language structure into which he is translating. At this point the translator should have a wide knowledge in both languages for getting the equivalence in the target language, because the deficiency of the knowledge of both languages decompression and compression principles will result in a translation without logic and sense.
The translation adequacy must be taken into account in fiction literature translation especially carefully, and much more intellectual electronic translators should be used for this in the future. Khatuna Beridze, lecturer of the translation theory and practice at the Batumi State University mentions that there are both linguistic and extralinguistic aspects that hinder to reach adequacy in fiction translation. Semantic information of the text differs essentially from the expressive-emotional information of the text but they have one common trait: both can bear and render extralinguistic information. Extralinguistic information often becomes a stone to stumble over by a translator, as it is a lingvoethnic barrier for a fiction literature translator; misunderstanding or misinterpretation of the extralinguistic information means can be the following:
-
1) Either what was actually communicated in the SL text, what means the pragmatic core of the SL text may be lost or therefore in the TL text ambivalence may arise for the recipient reader;
-
2) Or there may be misrepresented the author's communicative intention, the social context of the scene/situation as well as disposition or relationships of the communication participants.
Good examples could be brought from translation of stories and novels by V. P. Astafiev. The works of world famous Siberian writer V. P. Astafiev are very difficult to translate because his language is very distinctive, he uses a lot of non-standard events (phenomena), and the translator must handle the material carefully, for not to lose the value and meaning of the content. All nonstandard language features in his works can be divided into the following groups: dialect, vernacular, colloquial vocabulary, considering that the vernacular vocabulary is relatively the largest group.
Abnormalities in the speech of the characters serve as a mean of typification or, alternatively, as a mean of linguistic characteristics of a story hero. Abnormalities in the speech of the characters appear at all levels of the language: phonetic – phonetic distortion of the word shape, morphological – mistakes in coordination of number, gender, declension of nouns, syntactic – the use of specifiers (“эдак”, “нате вам”) to connect the sentence. However, most examples of deviations from the norm are at the level of vocabulary that can be considered as a marker of social status of heroes.
So, the semantic compression and decompression should take into account all the tiny particularities of the natural language.
In the author’s speech non-normative events often serve as a mean of creating imagery, expressing the author’s expression. Sometimes the vernaculars in the speech can be considered as the author’s irony.
Let’s see some examples of vernaculars from Astafiev’s text.
«Варнак» – villain (злодей, негодяй).
«Охламон» – dolt (дурень, болван).
«Поросята хоркают» (хрюкать) – to snort (фыркать).
«Закокать курочкой» (кудахтать) – cackle like a hen.
Sometimes interpreters can’t find the appropriate translation of definite phrases, because of distinctive character of Astafiev’s prose.
«Фулиган» – hooligan.
«За что жисть погубил» – What have I made such a mess in my life for.
«А тютюшеньки-тютю» – A-diddums, a-diddums, a-diddums!
«А люлюшеньки-люлю» – A-doodle, a-doodle, a-doodlums!
«А малюшеньки-малю» – A-doodle, a-doodle, a-doodlums!
In some cases interpreters translate words and phrases in such case that readers can understand the sense, but originality of author’s language is lost.
«Жили худо, бедно, натужно и недружно, вразнопляс» –
We had a mean, poor, strained, discordant life, dancing to different tunes.
«Исподличался совсем» – I had turned into a real little scoundrel.
The algorithm of more effective translation will be viewed as the following sequence of actions.
-
1. Writing generative grammar rules for potentially fully generation of a natural language subset in a definite approximation.
-
2. Potential generation of all possible states of the language around the standard classifications of the needed natural language units – standard words, new, occasional words, simple sentences, sentences with artistic phenomena (conditionally) interpreted by computer as semantic noise.
Because of the fact that deep emotional connotation is a very difficult task for linguistic software, the task to create the software, algorithms and approaches for automatic task generation was selected for applying the semantic noise addition model. It can be very useful in e-learning courses systems [10]. Such a task was realized based on language combinability theory and now such educational task generation systems are improved (Fig. 4). One of the next steps is to use second level patterns for language generation (see the table).

Fig. 4. Software for Automatic Educational Tasks Generation
In this way the two-levels patterns can be used for example for generating an educational task based on word combinability principles: “the user + takes + some wire” (Topic: Details of Equipment, Positions: Doer + Action + Receiver, Variants: Passive + Presentation-Passive + Metal-Long-Flexible); based on the phrase a task can be formed: “the user + takes / eats / wears / lives in + some wire”. The head of the task can be like: “select a proper word from the list” or “fill in the gaps with one of the variants, offered below”, etc. But further the decompression of the phrase can be made by adding semantic noise: “the user’s taking the wire was necessary” or “the wire after taking by the user was given to me” (see the table). Corresponding tasks could be made based on the decompressed sentences of the non-reduced form, for example, “the wire after taking / eating / wearing / living by the user was given to me”. The database of the second level semantic patterns should be created for the applications in e-learning linguistic software.
In the article the observation of multidimensional data by the natural language, particularly English, is given. It is possible to apply a multidimensional model of the natural language, semantic vectorized classification of words and notions of the natural language to make an assumption about the stylistic classification structure of the generative grammar rules set or relational wordsentences pattern sets, used for natural language generation. The structure of such rules has been analyzed. They can be used for text and sentences compression and decompression. The methods of using decompression by second level speech patterns for language generation are offered for improving linguistic software for automatic tasks generation for examples for the lessons of a foreign language. Further investigation in the sphere is necessary.