Conference PaperPDF Available

Building a Framework for Arabic Ontology Learning

Authors:

Abstract

This paper presents the ArOntoLearn a Framework for Arabic Ontology learning from textual resources. Supporting Arabic language and using domain knowledge or previous knowledge in the learning process are the main features of our framework, besides it represents the learned ontology in Probabilistic Ontology Model (POM), which can be translated into any knowledge representation formalism, and implements data-driven change discovery, therefore it updates the POM according to the corpus changes only, and allows user to trace the evolution of the ontology with respect to the changes in the underlying corpus. Our framework analyses Arabic textual resources, and matches them to Arabic Lexico-syntactic patterns in order to learn new Concepts and Relations.
Building a Framework for Arabic Ontology Learning
Nada Ghneim, Informatics Department, HIAST, Damascus, Syria, nada.ghneim@hiast.edu.sy
Waseem Safi, Informatics Department, HIAST, Damascus, Syria,
waseemm2005@hotmail.com
Moayad Al Said Ali, AI Department, IT Faculty, Damascus University, Damascus, Syria,
moayad@live.com
Abstract
This paper presents the ArOntoLearn a Framework for Arabic Ontology learning from textual
resources. Supporting Arabic language and using domain knowledge or previous knowledge
in the learning process are the main features of our framework, besides it represents the
learned ontology in Probabilistic Ontology Model (POM), which can be translated into any
knowledge representation formalism, and implements data-driven change discovery, therefore
it updates the POM according to the corpus changes only, and allows user to trace the
evolution of the ontology with respect to the changes in the underlying corpus. Our framework
analyses Arabic textual resources, and matches them to Arabic Lexico-syntactic patterns in
order to learn new Concepts and Relations.
Keywords: Ontologies, Ontology Learning, Knowledge Acquisition, Arabic Natural
Language Processing.
Introduction
Since Ontologies provide a shared understanding of a domain of interest, they became the key
technology of modern knowledge based systems: natural language processing, information
retrieval and the Semantic Web. Since building ontology for a huge amount of data is a
difficult and time consuming task, several ontology learning frameworks have been designed
and implemented in the last decade.
The Mo’k workbench [1], for instance, basically relies on unsupervised machine learning
methods to induce concept hierarchies from text collections.
OntoLT [2] is ontology learning plug-in for the Protege ontology editor. It is targeted more at
end users and heavily relies on linguistic analysis. It basically makes use of the internal
structure of noun phrases to drive ontology knowledge from texts.
The OntoLearn [3] framework by Velardi et al. focuses on the word sense disambiguation
problem, presents a novel algorithm called SSI relying on the structure of the general ontology
for this purpose.
TextToOnto [4] framework is implementing a variety of algorithms for diverse ontology
learning subtasks. In particular, it implements diverse relevance measures for term extraction,
different algorithms for taxonomy constructions as well as techniques for learning relations
between concepts.
Text2onto [5] is a new version from textToOnto. It presented Preliminary Ontology Model
provided with probability, and it relay on data-driven change discovery and NLP techniques.
All mentioned frameworks do not support Arabic language, and they do not use domain
knowledge or previous knowledge such as assistant ontology. In this paper, we will present
our framework which overcomes these deficient points by using Arabic Lexico-syntactic
patterns and semantic annotator, which is based on a novel idea of using domain or previous
knowledge in the traceable incremental ontology learning process. Supporting Arabic
language is not that easy task because current linguistic analysis tools are not efficient enough
to process unvocalized Arabic corpuses that rarely contain appropriate punctuation, so we
tried to build a flexible and freely configured framework whereas any linguistic analysis tool
can be replaced by more sophisticated one whenever it is available.
This paper is structured as follows: in the first section we will describe the System
Architecture which explains usage scenario and the system unit’s interaction. These units will
be detailed in the next three sections: The Probabilistic Ontology Model, Arabic Natural
Language Processing, and Algorithms Controller. Some experimental results are presented in
the Results and Discussion section. Conclusions and future work appear in final section.
System Architecture
The architecture of ArOntoLearn is divided into three units: Probabilistic Ontology Model
POM, Algorithms Controller, and Arabic NLP tools (see figure 1).
POM stores the results of the different ontology learning algorithms, which are initialized by
the Algorithms Controller, the purpose of which is to trigger the linguistic preprocessing of the
data, execute the ontology learning algorithms in the appropriate order and apply the
algorithms change requests to the POM. The fact that none of the algorithms has the
permission of directly manipulating the POM guarantees maximum transparency and allows
for the flexible composition of arbitrarily complex algorithm.
The execution of each algorithm consists of three phases: (i) in the notification phase, the
algorithm learns about recent changes to the corpus, (ii) in the computation phase, these
changes are mapped to changes with respect to the reference repository, which contains the
pointers to all occurrences of the concept, (iii) in the result generation phase, request for POM
changes are generated from the updated content of the reference repository.
Arabic Natural Language Processing NLP Unit is based on the GATE framework [6]. Our
choice of GATE is based on two facts: its flexibility with respect to the set of linguistic
algorithms used, and presenting Java Annotation Patterns Engine (JAPE) which provides
finite state transduction over annotation based on regular expressions that gives flexible
modeling and developing of Lexico-syntactic Patterns.
The Probabilistic Ontology Model
Probabilistic Ontology Model (POM) is presented by text2onto [5] framework as collection of
instantiated modeling primitives which are independent of a concrete ontology representation
language. The benefits of POM are twofold. On the one hand, adding new primitives does not
Algorithm Controller
Corpus
NLP
Evidence
store
Reference
store
POM
POM Visualization
RDF
Ontology
Fig 1. Architecture of ArOntoLearn
imply changing the underlying framework thus making it flexible and extensible. On the other
hand, it can be translated into various ontology representation languages such as OWL and
RDF. The modeling primitives we use in ArOntoLearn are:
i.Concepts.
ii.Instances.
iii.Concept inheritance (subclass-of).
iv.Concept instantiation (instance-of).
POM is not probabilistic in a mathematical sense, but because every instantiated modeling
primitive gets assigned a value indicating how certain algorithm in question is about the
existence of the corresponding instance. The purpose of these ‘probabilities’ is to facilitate the
user interaction by allowing him to filter the POM and thereby select only a number of
relevant instances of modeling primitives. Each POM change has references to the underlying
corpus changes, that is, why the user can trace the evolution of the ontology with respect to the
changes in the underlying corpus.
Arabic Natural Language Processing
Our Natural Language Processing unit is based on GATE framework which is very flexible
and can be freely configured by replacing existing algorithms or adding new ones such as a
deep parser if required, another advantage of using GATE is the ability of using existing
Arabic linguistics analyzers.
The Gate pipeline Application we used starts by Arabic tokenization and sentence splitting.
The resulting annotation set serves as input for an Arabic Morphological Analyzer and POS-
Tagger [7] which assigns appropriate morphological and POS tags to all tokens, such as
(token=, POS-tag=‘common_noun’, root=‘’, pattern=‘’, stem=‘’, stem
pattern=‘). The resulting annotation set will serve as input for an Arabic Syntactical
Analyzer [8] which will assign appropriate syntactic categories to all tokens and construct the
syntactical trees of each sentence. For example, the sentence ‘     ’ will
produce the annotations in the figure 3, where and ‘ are Proper noun phrase (NPP),
and they construct a Noun Phrase (NP).
Fig 3. Syntactical Analyzer Annotations
Then, by using an assistant ontology such as Arabic WordNet (AWN) [11], which represents
previous or domain knowledge in order to keep the result in the same domain and to get more
accuracy, a Semantic Annotator will find and annotate all known concepts in the texts to get
benefits from the knowledge we already had. For example (‘  ’ will be
annotated by the concept ‘Syria’ or super-concept ‘country’).
NP NP
DTJJ DTNN NNP NNP
   
Tokenizer
Morph POS-tagger
Syntactical Analyzer
Semantic Annotator
Lexico-syntactic
Patterns Matcher
Sentence Splitter
Fig 2. Gate Pipeline Application
After the basic linguistic preprocessing is done, a JAPE transducer is run over the annotated
corpus in order to match a set of particular Arabic Lexico-syntactic patterns required by the
ontology learning algorithms. Whereas the left hand side of each JAPE pattern defines a
regular expression over existing annotations, the right hand side describes the new annotations
to be created. We developed JAPE patterns to identify modeling primitives, i.e. concepts,
instances and different types of relations such as has-a and is-a. In the following we
explain some of our Lexico-syntactic patterns written in JAPE:
Pattern 1: is-a relation
(syntaxnode.type =NP): instance
({Token.category=is-a})//is-a={


}
(syntaxnode.type = NP): concept
This pattern aims to match sentences, for example (‘  ’, ‘  ’, 
 ’, ‘  ’, ‘  ’), which all express is-a relation between the
instance and the concept.
Pattern 2: has-a relation
({Token.root==’’}| {Token.root=’’})
(syntaxnode.type = NP):sub_concept
({Token.string=})
({Token.string=’’})
(syntaxnode.type = NP):super_concept
This pattern aims to match sentences, like (‘’, ‘
), which all express has-a relation between sub_concept and super_concept.
We have developed a core of 15 Lexico-syntactical pattern to capture is-a and has-a
relations in different Arabic sentence forms. Once a pattern captures a relation, it adds
annotation tags that refer to the relation elements such as instance, concept and sub_concept.
Algorithms Controller
Algorithms Controller drives the whole learning process. First, it runs the NLP unit and gets
produced annotations, then fires the learning algorithms corresponding to these annotations
types, in appropriate order. In fact, algorithms can be classified according to the modeling
primitives they produce. In order to obtain a more reliable probability for each instantiated
primitive, the algorithms controller allows performing several algorithms of different types,
and then combines their results. Because algorithms can’t directly change the POM, they send
POM change requests to the algorithms controller which applies them to the POM.
Hereafter, we briefly describe the algorithms we used in the learning process. In particular we
describe the approach to calculate the probability of an instantiated modeling primitive:
i. Concept and instances algorithms:
We implemented different algorithms calculating the following measures: Relative Term
Frequency (RTF), Term Frequency Inverted Document Frequency (TFIDF) [9], and
Entropy [10]. For each term, the values of these measures are normalized into the interval
[0..1], and used as corresponding probability in the POM.
ii. Is-a and has-a Relation Algorithms:
In order to assign the confidence of extracted relations we implemented different
algorithms. We used Google search engine to calculate the confidence according to the
count of results returned when searching for the relation, the search query is formed
according to relation type and elements, and we used AWN [11] lexical ontology to search
for the relation, if it’s found then the confidence will be 1 else it’ll be 0.
Results and discussion
To evaluate the quality of the ontology learning results we tested our system on texts selected
from Arabic Wikipedia. We have selected about 125 documents in the domain of countries
and cities. We used Arabic WordNet as assistant ontology. We have established 15 Lexico-
syntactic patterns (5 for instance-of relations, 10 for subclass-of relations). Our test focused on
the instance-of relation. Since applying Stanford Syntactic Parser on Arabic sentences requires
that these sentences are well formed and this are not the case Arabic Wikipedia documents, we
have performed two tests: (i) by using Stanford Syntactic Parser on a limited number of
sentences (34 sentences), (ii) by using only Arabic Morphological Analyzer, without Stanford
Syntactic Parser on the whole corpus (12779 sentences) (see Table 1). We divided the results
according to a relation confidence threshold that we have assigned the value 0.5. The results
were as following:
i. Using Morphological Analyzer:
The number of relations with confidence more or equals 0.5 is 68. After the manual
verification of these relations we found that, there are 26 true instance-of relation and 42
false instance-of relation.
The number of relations with confidence less than 0.5 is 40. After the manual verification
of these relations we found that, there are 28 true instance-of relation and 12 false instance-
of relation.
The precision of the results of this test is 0.5.
ii. Using Stanford Syntactic Parser:
The number of relations with confidence more or equals 0.5 is 10. After the manual
verification of these relations we found that, there are 8 true instance-of relation and 2 false
instance-of relation.
The number of relations with confidence less than 0.5 is 2. After the manual verification of
these relations we found that, there are 2 true instance-of relation and none false instance-
of relation.
The precision of the results of this test is 0.83.
Table 1: Results
We remarked that using Stanford Syntactic Parser has increased the precision of the results,
because we could use the syntactic annotations of the sentences in the Lexico-syntactic
patterns. Unfortunately we could not apply Stanford Syntactic Parser on the whole corpus,
because it requires that the sentences are well formed and with appropriate punctuations.
Conclusion and Future Work
We have developed a framework for incremental ontology learning, using Arabic natural
language processing, machine learning and text mining techniques, in order to extract
ontology from Arabic textual resources. The novel aspects about our framework are: (i) the
flexibility with respect to use other Arabic linguistic analyzers, and add new Lexico-syntactic
patterns to reach more accuracy, (ii) the independence of a concrete ontology representation
Sentences
>=0.5
True
False
<0.5
True
False
Morphological
12779
68
26
42
40
28
12
Syntactical
34
10
8
2
2
2
0
language, (iii) benefits from the previous knowledge by using an assistant ontology, (iv) using
the probability for capturing uncertainty and enhancing user interaction, (v) the integration of
data-driven change discovery strategies increasing the efficiency of the system, as well as the
traceability of the learned ontology with respect to changes in the corpus, making the whole
process more transparent.
In the future, we intend to develop more Arabic Lexico-syntactic patterns to involve more
relations and sentence forms. We can extend the POM with more taxonomic and non
taxonomic relations. In view of increasing the precision of the system, we can replace existing
Arabic morphological and syntactical analyzer with more accurate ones, and add an Arabic
lemmetizer. We can use machine learning and text mining techniques in our algorithms used
for extracting modeling primitives.
References
[1] G. Bisson, c. Nedellec, and L.canamero. Designing clustering methods for ontology
building The Mo’K workbench. In proceedings of the ECAI ontology Learning Workshop,
pages 13-19, 2000.
[2] P. Buitelaar, D. Olejnik, and M. Sintek. OntoLT: A protégé plug-in for ontology extraction
from text, 2003, In Proceeding of the international Semantic Web Conference, (ISWC).
[3] P. Velardi, R. Navigli, A. Cuchiarelli, and F. Neri. Evaluation of ontolearn, a methodology
for automatic population of domain Ontologies, 2005, In P. Buitelaar.
[4] A. Maedche and S. Staab. Ontology learing, 2004, In s. Staab and R. Studer, editors,
handbook on Ontologies, pages 173-189. Springer.
[5] P. Cimiano and J. Volker. Text2onto a framework for ontology learning and data-driven
change discovery, Proc. NLDB 2005, Lecture Notes in Computer Science, vol. 3513,
Springer, Alicante, 2005, pp. 227-238.
[6] H. Cunningham, D. Maynard, K. Bontcheva, and V. Tablan. GATE: A framework and
graphical development environment for robust NLP tools and applications, 2002, In
Proceedings of the 40th Annual Metting of the ACL.
[7] R. Sonbol, N. Ghneim, M. Desouki, Arabic Morphological Analysis: a new approach,
ICTTA 2008, Damascus, Syria.
[8] Arabic Stanford parser, Retrieved October 2,2008, from :
http//www.nlp.stanford.edu/software/parser-faq.shtml.
[9] G. Salton. Developments in automatic text retrieval, 1991, Springer 253:974-979.
[10] R.M. Gray. Entropy and information Theory, 1990, Springer.
[11] Navigli, R., Paola Velardi, Alessandro Cucchiarelli and Francesca Neri. Extending and
Enriching WordNet with OntoLearn, 2004, Proc. of The Second Global Wordnet Conference
2004 (GWC 2004), Brno, Czech Republic.
... The ontology has been implemented manually using protégé-4.3. The system also provided a validation framework for the concepts and the relationships between them [10]. ...
... The framework generates an Arabic ontology from XML documents. The framework builds Arabic ontology from semi-structured data by exploiting the same notations used in [10]. It consists of four phases: Extraction, XML schema parsing, Generation, Validation, and Evolution. ...
... The ontology has been implemented manually using protégé-4.3. The system also provided a validation framework for the concepts and the relationships between them [10]. ...
... The framework generates an Arabic ontology from XML documents. The framework builds Arabic ontology from semi-structured data by exploiting the same notations used in [10]. It consists of four phases: Extraction, XML schema parsing, Generation, Validation, and Evolution. ...
Conference Paper
Full-text available
Arabic is the official language of millions of people in the Middle East and northern African countries. Arabic ontology is the foundation of the formulation of Arabic Semantic Web. Surprisingly, little has been done in the field of computerized language and lexical resources. This paper introduces a framework that generates an Arabic Ontology from a semi-structured data (XML documents associated with graph schema), in which, XML schema is created and utilized in the graph schema development (XSG). Finally, the paper provides a case study, insectivore’s case, where the developed Arabic ontology is applied. The results consist of 143 words, 10 Concepts, 10 elements and 20 relations. The generated ontology is evaluated using data-driven evaluation method and tree based mining. 65% of the source XML documents have been included in the insectivore’s case study. This result can be refined more to reach satisfying results.
... Lee et al. [9] presented approaches for Chinese text processing for news summarization in Chinees language. Ghneim et al. [10] proposed a framework for Arabic Ontology Learning (ArOntoLearn). The Probabilistic Ontology Model (POM) used in the (Text2Onto) framework to represent the extracted ontology [11]. ...
... This paper presents the ArOntoLearn a Framework for Arabic Ontology learning from textual resources [9]. It supports the Arabic language and using domain knowledge or previous knowledge in the learning. ...
... Ghneim et al. [2] proposed a framework for Arabic Ontology learning (ArOntoLearn) based on previous domain knowledge. The framework is an extension of Text2Onto [8] framework. ...
Conference Paper
Full-text available
Ontology, the backbone of Semantic Web, is defined as the formal specification of conceptual hierarchy with relationships between concepts. Ontology Learning (OL) is a process to create an ontology from text automatically or semi-automatically. OL is an important topic in the Semantic Web field in the last two decades but it is still not mature in Arabic not like Latin languages. Currently, there is a limited support for using knowledge from Arabic literature automatically in semantically-enabled systems. Deep Learning (DL), an artificial neural networks learning based application, has proved a good improvement in multiple areas including text mining. By using DL, it is possible to have word embedding as distributed word representations from textual data. The application of DL to aid Arabic ontology development remains largely unexplored. This paper investigates the performance of implementing DL with Arabic ontology learning tasks using major models such as Continuous Bag of Words (CBOW) and Skip-gram. Initial performance results are promising as an effective application of Arabic ontology learning.
... Ghneim et al. [2] proposed a framework for Arabic Ontology learning (ArOntoLearn) based on previous domain knowledge. The framework is an extension of Text2Onto [8] framework. ...
... The ArOntoLearn [13] presents a framework for building Arabic Ontology from textual resources. The proposed framework applies a set of NLP tools as GATE pipeline application. ...
Conference Paper
Full-text available
Quran Recitation relies on identifying and applying different Tajweed rules [قواعد التجويد] such as Muddud [مدود] and Tanween [تنوين] in the Quran text. This research is aimed at providing a tool that automatically finds and annotates letters that embody Tajweed rules in Quran text. This field remains an open research area due to the lack of open source NLP tools that support the Arabic language. Applying Natural Language Processing (NLP) techniques on Quran text to extract Tajweed letters is considered an important Information Extraction (IE) step. This research explores the field of applying IE techniques on Quran text. Rule based IE techniques are well known to achieve optimal results. This research explores NLP techniques on Quranic text using GATE, an open source flexible NLP environment. GATE is employed for this research to build the application that processes un-annotated Quranic text corpus. The developed application is evaluated using the well known IE evaluation metrics precision and recall. By comparing the system’s automatically annotated text with a gold standard (i.e. Quran text). The system proved to be efficient by achieving 100% precision and recall of the implemented Tajweed rules.
Conference Paper
This paper introduces a novel resource for Arabic Qur'anic textual annotations: AQD, Arabic Qur'anic Database, providing an annotation-level search that draws on a number of available resources in a single query. In addition, it allows implementing a set of queries as rewrite rules, which is performed in a recursive way. The experiments show that our AQD is able to discover knowledge from very simple to very complex queries.
Article
Full-text available
Ontology is the gateway for the interaction between human and the machines because it facilitates the gaining of the information for the end users. The semantic web terms and vocabulary are considered as form of ontology. So the improvement of the ontology helps in the semantic web progration. Arabic ontology is a critical natural language processing field; it helps to enrich the Arabic language resources. Although, the searchers started to focus their efforts creating ontologies in Arabic language, the Arabic terms and fields have not yet been covered. This paper introduces an innovative framework, annotated as NAAO (Novel Automatic Arabic Ontology), which automates the ontology generation process from XML documents. The novelty of NAAO resides in generating the Arabic ontology, in the form of XML graph schema (XSG), from semi-structured data (XML documents associated with graph schema). The automation process entails four main phases (extrication, XML schema parsing, ontology generation and refinement and evaluation). Extrication phase, where all the data required for the ontology generation is extricated from the information source. XML schema parsing phase, where the extracted data is going to be investigated and the hierarchical analysis to identify all the parts of the ontology is going to be created. Ontology generation phase deal with the production of the Arabic ontology (classes, elements, and relations). Finally, refinement and evaluation phase that deals with improvements of the Arabic ontology. Moreover, the paper comprises versatile examples for an automatic generation to Arabic ontologies regarding the animal kingdom. Such as the invertebrate's ontology, in which the XSG is automatically extracted using 103 XML documents, 320 concepts (classes), 603 elements and 783 relations. Finally, NAAO enjoys certain advantages than others since it shows an automatically generate Arabic ontology from semi-structured data (XML file). It provides mapping connections to ensure the right relationships with the classes and the elements to its corresponding nodes in the XML documents and supplies huge ontology contents (concepts elements and relations). Moreover, NAAO offers an automatic evaluation of the generated Arabic ontology using cosine similarity measurements.
Conference Paper
Full-text available
. This paper describes Mo'K, a configurable workbench that supports the development of conceptual clustering methods for ontology building. Mo'K is intended to assist ontology developers in the exploratory process of defining the most suitable learning methods for a given task. To do so, it provides facilities for evaluation, comparison, characterization and elaboration of conceptual clustering methods. Also, the model underlying Mo'K permits a fine- grained definition of similarity measures and class construction operators, easing the tasks of method instantiation and configuration. This paper presents some experimental results that illustrate the suitability of the model to help characterize and assess the performance of different methods that learn semantic classes from parsed corpora. 1. INTRODUCTION In this paper we propose a workbench that supports the development of conceptual clustering methods for the (semi-) automatic construction of ontologies of a conceptual hierarchy typ...
Conference Paper
Full-text available
Morphological analysis is an important step in the process of Arabic language processing which became a must in today's IT development. We present here a new approach for Arabic root extraction. Our algorithm outperforms most other morphological analysis algorithms; we achieved more reliable results with better performance and minimal storage. Our Morphological analyzer provides different levels of reliability and performance to support the needs of different applications. It also provides a separate stage which can be plugged easily in any other Morphological analyzer to improve its performance, and with no negative effect on its reliability.
Article
Full-text available
OntoLearn is a system for word sense disambiguation, used to automatically enrich WordNet with domain concepts and to disambiguate WordNet glosses.
Article
language technology tools into ontology development environments is to be expected. Language technology tools will be essential in scaling up the Semantic Web by providing automatic support for ontology monitoring and adaptation. Language technology in combination with approaches in ontology engineering and machine learning provides linguistic analysis and text mining facilities for ontology mapping (between cultures and applications) and ontology learning (for adaptation over time and between applications).
Article
In this paper we present Text2Onto, a framework for on- tology learning from textual resources. Three main features distinguish Text2Onto from our earlier framework TextToOnto as well as other state-of-the-art ontology learning frameworks. First, by representing the learned knowledge at a meta-level in the form of instantiated model- ing primitives within a so called Probabilistic Ontology Model (POM), we remain independent of a concrete target language while being able to translate the instantiated primitives into any (reasonably expressive) knowledge representation formalism. Second, user interaction is a core as- pect of Text2Onto and the fact that the system calculates a confldence for each learned object allows to design sophisticated visualizations of the POM. Third, by incorporating strategies for data-driven change dis- covery, we avoid processing the whole corpus from scratch each time it changes, only selectively updating the POM according to the corpus changes instead. Besides increasing e-ciency in this way, it also allows a user to trace the evolution of the ontology with respect to the changes in the underlying corpus.
Article
Recent developments in the storage, retrieval, and manipulation of large text files are described. The text analysis problem is examined, and modern approaches leading to the identification and retrieval of selected text items in response to search requests are discussed.
Book
Contents Prologue xi 1 Information Sources 1 1.1 Introduction .............................. 1 1.2 Probability Spaces and Random Variables ............. 1 1.3 Random Processes and Dynamical Systems ............ 5 1.4 Distributions ............................. 6 1.5 Standard Alphabets ......................... 10 1.6 Expectation .............................. 11 1.7 Asymptotic Mean Stationarity ................... 14 1.8 Ergodic Properties .......................... 15 2 Entropy and Information 17 2.1 Introduction .............................. 17 2.2 Entropy and Entropy Rate ..................... 17 2.3 Basic Properties of Entropy ..................... 20 2.4 Entropy Rate ............................. 31 2.5 Conditional Entropy and Information . . ............. 35 2.6 Entropy Rate Revisited ....................... 41 2.7 Relative Entropy Densities ...................... 44 3 The Entropy Ergodic Theorem 47 3.1 Introduction ..........
OntoLT: A protégé plug-in for ontology extraction from text
  • P Buitelaar
  • D Olejnik
  • M Sintek
P. Buitelaar, D. Olejnik, and M. Sintek. OntoLT: A protégé plug-in for ontology extraction from text, 2003, In Proceeding of the international Semantic Web Conference, (ISWC).