ArticlePDF Available

Abstract and Figures

The organized capturing and sharing of knowledge is very important, and a lot of tools, such as wikis, social communities and knowledge-management or e-learning portals, exist for supporting this purpose. The community content- and knowledge-capturing, management and sharing portal of the European project “Realising an Applied Gaming Eco-system” (RAGE)†† combines such tools. The goal of the RAGE project is to boost the collaborative knowledge asset management for software development in European applied gaming (AG) research and development (R&D). To support this process, the so-called RAGE ecosystem implements a portal to support the related asset, content and knowledge exchange between diverse actors in AG communities. Therefore, the community portal in RAGE is designed as a so-called ecosystem and is intended to provide its users different tools for the capturing, management, and sharing of knowledge. In this study, we rely on the term and model definition of spiraling knowledge exchange between explicit and tacit knowledge given by Nonaka and Takeuchi.1 To achieve the goal of extracting, i.e., externalizing and explicitly representing and sharing this knowledge to its users, we propose to generate a taxonomy for faceted search automatically by extracting named entities form the knowledge sources and to classify documents using Support Vector Machines (SVM). In this paper we present our architectural approach for the NLP-based IR concepts and discuss how cloud services based on data distribution and cloud computing can improve the outcome of our system.
Content may be subject to copyright.
Procedia Computer Science 68 ( 2015 ) 206 216
1877-0509 © 2015 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license
-review under responsibility of Institute of Communication and Computer Systems.
doi: 10.1016/j.procs.2015.09.236
Available online at
HOLACONF - Cloud Forward: From Distributed to Complete Computing
Towards Cloud-Based Knowledge Capturing Based on Natural
Language Processing
Christian Nawroth,
Matthäus Schmedding,
Holger Brocks,
Michael Kaufmann,
Michael Fuchs,
Matthias Hemmje
University of Hagen, Faculty for Mathematics and Computer Science, Hagen, Germany
Research Institute for Communiction and Cooperation, Dortmund, Germany
Lucerne University of Applied Sciences and Arts, School of Engineering and Architecture, Horw, Switzerland
Wilhelm Büchner University of Applied Sciences, Darmstadt, Germany
The organized capturing and sharing of knowledge is very important, and a lot of tools, such as wikis, social communities and
knowledge-management or e-learning portals, exist for supporting this purpose. The community content- and knowledge-capturing,
agement and sharing portal of the European project Realising an Applied Gaming Eco-system (RAGE)
combines such tools.
The goal of the RAGE project is to boost the collaborative knowledge asset management for software development in European
applied gaming (AG) research and development (R&D). To support this process, the so-called RAGE ecosystem implements a
ortal to support the related asset, content and knowledge exchange between diverse actors in AG communities. Therefore, the
mmunity portal in RAGE is designed as a so-called ecosystem and is intended to provide its users different tools for the capturing,
agement, and sharing of knowledge. In this study, we rely on the term and model definition of spiraling knowledge exchange
etween explicit and tacit knowledge given by Nonaka and Takeuchi.
To achieve the goal of extracting, i.e., externalizing and
explicitly representing and sharing this knowledge to its users, we propose to generate a taxonomy for faceted search automatically
by extracting named entities form the knowledge sources and to classify documents using Support Vector Machines (SVM). In this
paper we present our architectural approach for the NLP-based IR concepts and discuss how cloud services based on data
stribution and cloud computing can improve the outcome of our system.
© 2015 The Authors. Published by Elsevier B.V.
Peer-review under responsibility of Institute of Communication and Computer Systems.
Corresponding author.
E-mail address:
© 2015 The Authors. Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license
Peer-review under responsibility of Institute of Communication and Computer Systems.
Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
Keywords: Storage Cloud; Scientific Cloud; Natural Language Processing; Named Entity Recognition; Support
Vector Machines; Knowledge Management
1. Introduction and Motivation
The EU-based industry for applied gaming is an emerging business field. However, it is still fragmented and needs
to build critical mass for global competition. The actors in the game development in
dustry, namely the developers,
resellers, users, researchers, etc., are not working together optimally because they do not yet have an integrated
orm for knowledge interchange. The European Commission (EC)-funded RAGE project will help overcome this
fragmentation and aims to exchange knowledge through an online portal and social space that will connect research,
gaming industries, intermediaries, education providers, policy makers and end users.
The goal of this online portal, which is called the RAGE ecosystem, is to allow its participants
x to get hold of advanced, usable gaming assets (technology push);
x to get access to associated bus
iness cases (commercial opportunity);
x to create bonds with peers, s
uppliers and customers (alliance formation);
x to advocate their expertise and demands (publicity);
x to develop and publish their ow
n assets (trade) and
x to contribute to creating a joint agenda and road
map (harmonization and focus).
Seen as a whole, RAGE is a technology- and know-how-driven research and innovation project. Its main purpose
is to be able to equip industry players (e.g., game develo
pers) with a set of technology resources (so-called assets, i.e.,
software and related knowledge resources) and strategies (i.e., knowledge management support features) to strengthen
their capacities to penetrate a market (non-leisure) that is new for most of them, and to consolidate a competitive
position in it.
Knowledge-interchange portals mean a lot of human effort to accu
mulate useable information. Automated
knowledge extraction can improve the efficiency of the portal because it generates useful content automatically.
Therefore, RAGE will support knowledge extraction from differen
t knowledge sources such as a social network, the
software repository, the media archive and the learning management system. In that context, we have identified the
following general research questions for which possible answers are described in sections 3 and 4, respectively:
(1) How can AG knowledge be captured automatically, and how can this knowled
ge be shared optimally?
(2) How can cloud technology support knowledge c
apturing in the RAGE ecosystem?
Usually, knowledge management is performed by th
e usage of assistance tools, e.g., wikis, blogs, e-learning portals,
etc. The RAGE ecosystems initial design already consists of such tools, and we want to go a step further. A lot of
implicit and tacit knowledge, as introduced by Nonaka and Takeuchi,
is produced by the communities interacting
with and within such a portal. For instance, with contents that are commonly used together, e-learning module users
have to pass to reduce colleagues’ knowledge gaps, etc. Considering this description, RAGE is a system that contains
multiple kinds of content and functionality. The storage of heterogeneous contents and knowledge management in
RAGE require an especially high input of resources due to the complexity of the related tasks. For this, a cloud-based
approach relieves the user from hosting and setting up the complex RAGE (NLP-IR) environment and transfers this
to a specialized cloud provider. This is our main motivation for investigating the need and the advantages for a
ud-based NLP-IR architecture in RAGE.
The main components of the RAGE ecosystem considered as knowledge sources are:
x A social network system-integration component will enable the RAGE ecos
ystem users to collaborate with each
other in social network systems and across those and the RAGE ecosystem, which is developed and described by
Salman et al. in two papers.
Thus, the users can directly exchange their knowledge, e.g., in chats or other forms
of interaction, but they can also use the RAGE ecosystem for content and knowledge management and sharing.
x The software repository of RAGE will contain so-called assets that represent, e.g., advanced technology
components for the support of game development.
x The media archive contains multimedia content, such as video instructions, and the digital library includes
ments for example, scientific research papers, n digital form.
x The learning management system enables ecosystem users to author, access, and consume online courses to build
p knowledge for the usage of specific assets or reduce knowledge gaps among project participants.
208 Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
In our cloud-based architecture, these four knowledge sources are the blueprint for the structure of the RAGE
storage cloud, which is an abstract concept of data distribution, which we introduce in section 4.4. Before we describe
the cloud architecture, we will take a short look at our hybrid approach, which uses methodologies from several fields
of computer science, such as recommender systems (R
S), natural language processing (NLP) with its sub-discipline
named-entity recognition (NER) and information retrieval (IR) for the capturing and sharing of such knowledge.
Due to the high number of different components of t
he RAGE ecosystem and the corresponding content types, it is
very important to manage the flow of knowledge in the RAGE ecosystem. There are different ways by which
knowledge can be captured, represented, managed, further developed and re-used, i.e., shared and disseminated to
other users. Thus, it is not sufficient to provide tools that already exist in the initial design of the RAGE ecosystem. It
is important to know which kind of content types are commonly used together, which content items rely on other
content items, whom to ask if ecosystem users have questions regarding specific assets, and much more. Therefore,
we have decided to develop a hybrid approach that aims at extracting tacit user knowledge and implicit content
wledge from data, and share this emergent knowledge with the users of the RAGE ecosystem. One pipeline in this
hybrid is the NLP-IR system. In this paper we will introduce our overall concept for this system. In our cloud-based
architecture, the NLP-IR system is part of the RAGE computing cloud, which focuses on high-performance computing
C) tasks to support all functionality to provide the NLP-IR services defined by the use cases. Therefore, in this
paper we describe RAGE, our overall NLP-IR approach and combine both with a data distribution and cloud
omputing based architecture.
Before we describe the cloud architecture, we take a short look at our hybrid approach, which uses methodologies
from several fields of computer science, such as recommender systems (RS), natural language processing (NLP) with
its sub-discipline named-entity recognition (NER) and information retrieval (IR) for the capturing and sharing of such
2. Related Work
The base technology of the RAGE ecosystem will be built upon the so-called Educational Portal (EP) tool suite that
was developed by the software company GLOBIT.
The EP tool suite was developed as a solution for the growing
need of scientific communities to manage their document and media collections as well as their educational and other
kinds of knowledge resources. Furthermore, one of its additional purposes is to support continuous professional
ucation and training of practitioners, experts and scientists who are members of professional communities of practice
or scientific communities. The EP tool suite has been developed in order to provide an internet-platform which allows
or interaction among community members and access to their digital assets. Su
ch assets include, e.g., published
papers, lecture videos, abstracts and entire online courses. A reference application of the EP tool suite is, e.g., the
United European Gastroenterology (UEG) association’s educational platform 2. In order to find information in the EP,
a UEG community member can either search for specific words in all texts or set a filter for specific categories provided
by the communities. UEG experts have created domain-specific, hierarchical categories to w
hich a given document
can belong. In order to make use of this feature, the need to assign specific categories to given documents arises.
Assigning categories to digital documents can be done either manually or automatically. The manual approach
yields the best results but has two intuitiv
e drawbacks: domain experts time is a sparse resource and every
categorization created is subjective to the individual expert’s opinion. The automated assignment of a specific text to
a given category is commonly called text categorization (TC, also known as text classification or topic spotting).
initial class paper is largely based on Mohri et al.
The UEGs EP-based application solution is based on the TYPO3
content management system (CMS). The underlying architecture paradigm is the service-oriented architecture
paradigm. Before the work of Mohri et al.,
the EP employed an automated TC approach using the Apache Solr 3
enterprise search engine to calculate scores for individual terms. The higher the score, the more representative of the
ment the term is. The terms used were the names of the available cate
gories. If the score exceeded a predefined
threshold, the document was automatically assigned to the category represented by the term.
Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
Besides its application by the UEGs EP-based application solution, the EP tool suite was already used in the
European project Alliance Permanent Access to t
he Records of Science in Europe Network (APARSEN).
The EP
tool suite offers a wide variety of tools. This includes a web-based, user-friendly course-authoring tool, and content &
knowledge management tools for the management of documents, multimedia objects, taxonomies and more. Figure 1
displays the components and services in the Educational Portal tool suite underlying the RAGE ecosystem. The EPs
three-tier architecture is divided into core functions, extensions and the data storage.
In the NLP-IR, part our work is motivated by several concepts that use NLP to support knowledge capturing in
ain-specific environments: No
bata, Cotter et al.
present an integrated IR system for biology that is built upon
conditional random field (CRF)-based named-entity recognition. Ananiadou, Thompson et al.
developed an IR system
based on text-mining techniques that provides access to documents for the education community. The NLP-IR
component has to address the heterogeneous structure of the RAGE ecosystem and will rely on a couple of base
es: "Text classification based on support vector machines h
as been known for one and a half decades. Our
classification approach for the wiki and other text-based components relies on the base ideas of SVM text classification
as p
ublished by Joachims
and Sebastiani.
Filipova and Hall
demonstrated how to classify videos based on textual
meta-data as well as on video content, a concept which could be used in the media archive component. Swoboda
developed an SVM-based classifier for the medical domain, which works with the bag of words approach. Our work
ill extend his approach, not using the bag of words approach but using features extracted from the RAGE ecosystem’s
ntent, knowledge resources and software assets. Ritter, Clark et al.
present how to use NER in social media
(Twitter), which we will use to extend the knowledge-capturing features of the social network system integration
support and for integrating knowledge-capturing features in the LMS of the RAGE ecosystem. Dascalu, Dessus et al.
developed ReaderBench, an NLP-based system tailored to the needs of learning and educational environments.
ench will also be a part of the RAGE ecosystems core system. Dit, Revelle et al.
provide a comprehensive
overview of concepts for feature extraction in source code to be used, for example, in software repositories.
Our ideas about cloud are based on several concepts. First of all, we reference GATECloud, developed by Tabland,
oberts et al.
GATECloud is a cloud-based NLP solution of the GATE framework, which we mentioned before.
Klenner, Bergmann et al.
present UIMA-HPC, an unstructured information management applications (UIMA)-based
NLP system that uses grid-based HPC. This project shows which amount of CPU capacities may be used for large-
Fig. 1. Architecture of the EP
210 Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
scale NLP purposes. We use it as a starting point to transfer this grid-based UIMA approach to a cloud-based service.
Evangelinos, Hill et al.
give an example of how HPC applications can be transferred to cloud-based services. Our
approach on cloud-based SVM classification is primarily based on Pouladzadeh, Shirmohamaddi et al.,
who present
a cloud-based SVM classifier for object classification. They also show that in their approach the cloud-based SVM
classifier outperforms a traditional implementation
with LIBSVM. Apache SolrCloud is a cloud-based implementation
of the well-known open-source search server Apache Solr.
SolrCloud could be the base of our IR subsystem
described in the architecture section.
3. NLP-based Information Retrieval Support (NLP-IR)
3.1. NLP-IR Use Cases (as a Cloud Service)
The NLP-IR component of RAGE will provide a search functionalit
y for RAGE. Due to the heterogeneous structure
of the content, knowledge resources and software assets in the RAGE ecosystem, the NLP-IR component will cover
two dimensions. On one side, the NLP-IR component will have to deal with the multiple types of content in the
different RAGE modules, for example text documents, multimedia content, source code, social media content etc. On
e other side, an adaptation of the functional requirements of the applied gaming sector is necessary, represented by
36 RAGE asset categories and existing taxonomies. So from the NLP-IR point of view, we talk about two dimensions
NLP-IR support in this project. Based on these assumptions, we defined two major use cases for NLP-based IR
support, which are keyword search and faceted search. In the case of keyword search, the user specifies one or more
keywords and is able to filter and refine his search results by a multi-dimensional categorization scheme, which covers
both dimensions mentioned above. In a faceted search, the user does not specify any keywords but uses solely the
orization to access assets in the RAGE ecosystem.
From a user perspective, both use cases have the potential to be classical” use cases for a cloud
While all the data as well as the functionality may be covered by two cloud systems (RAGE Storage
Cloud and RAGE
Computing Cloud), on the user side just a simple interface (browser, app) is necessary to provide
ubiquitous and smart
user access.
3.2. NLP-IR - Working Hypothesis and Research Questions
To help explain our NLP-IR approach, we give a short description of three working hypotheses that are the basis
of NLP-IR. Our first hypothesis is that a bag of words approach is not the appropriate co
ncept for IR support in the
RAGE ecosystem due to its heterogeneous content structure. As there are a
ssets that will not work with full-text IR
(e.g., videos or software libraries) we try to extract the relevant features using NLP
either from text documents as well
as from meta-data. Our idea of using NER for this kind of knowledge extraction and document classification is
arily based on Guo, Xu et al.
These authors found that 71% of user search queries contain named entities. Based
on this observation, our second hypothesis is that in addition to queries, named entities are also a good feature to
tract from the content to support search and access. Our third hypothesis is that in domain-specific content, the
eight of named entities as an IR feature is even higher than in multi-purpose IR. Nadeau and Sekine
show several
examples of successful domain-specific named-entity recognition. Because we work in the highly specialized domain
applied gaming, we believe that a domain-specific approach similar to that presen
ted in the paper mentioned before
will work in this case too. Based on these hypotheses, the NLP component should rely on state-of-the-art named-entity
recognition to identify named entities and deal
with heterogeneous content. Due to the two dimensions of NLP-IR
support mentioned above, there will be an adaptation of the NER component to the different types of content and the
ain-specific asset categories necessary. As these three (in t
his context yet unproven) working hypotheses are the
main anchor of the NLP-IR sub-project, the development of an ap
propriate evaluation strategy will be a central aspect
of our future work. The three working hypotheses lead to the following main research question of the overall project:
(3) Is an NLP- and classification-based IR approach in a domain-specific software repository indeed
more efficient
than traditional bag-of-words-based IR-approaches?
Hence, from a technical perspective we will investigate which kinds of named entity categories besides the standard
ories (Name, Place and Organization) are suitable for the applied g
aming sector. First (yet-unevaluated) examples
Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
for this are Story-Character,” Player-Class and Emotion. Named-entity recognition, besides the standard
categories, requires the use of NER components that can be trained on special named-entity classes of the applied
gaming sectors. For this we use NER based on conditional random fields (CRF). One of the state-of-the-art named-
entity recognizers fulfilling this requirement is the NER component of Stanford CoreNLP, which we will use in this
To realize interoperability and to support our modular approach, we are going to integrate this NER
component in a standard NLP framework such as GATE
or Apache UIMA
using the appropriate wrappers.
3.3. NLP-IR Architecture
The main idea is to extract named entities from the data and use these to automatically categorize documents in
RAGE so that they can be used as taxonomy in faceted search. Figure 3 displays a pipeline representing our multi-
layer architecture for the NLP-IR workflow. The primary layer is the RAGE ecosystem with its sub-components, as
described above. The adapter
layer is intended to connect the relevant RAGE ecosystem subsystems containing
different types of content with the
NER component. Adoption in this context essentially means to extract (textual)
meta-data from the non-text assets
such as multimedia content or software components and to normalize unstructured
content, e.g., from social network
systems and the learning management system, to get consistent input for the NER
component. In the following
NLP layer, a tailored state-of-the-art NLP component will be used to extract the named
entities (NE-chunks) from
the textual input. The output of this layer is a vector whose features are NE-chunks to be
processed in the following
classifier layer. During the project we will investigate whether a binary or a weighted
output vector is suitable for
the following classifier layer. The classifier layer, which contains the SVM-based
classifier, is used for the document
classification within a domain-specific taxonomy. Finally, the pipeline is concluded
by the information-retrieval layer,
which supports keyword search as well as faceted search based on the taxonomy built
by the classifier. The output of
the IR layer is merged with the result sets of the recommender system (RS), addressing
our hybrid approach, and will be presented to the user on the GUI. Our main goal with this layered architecture is to
allow a modular and independent
development of the several components after inter-layer interface specification.
3.4. RAGE Asset Categories
To support access to the heterogeneous contents of the RAGE ecosystem, we introduce a hierarchical classification
scheme, which will be used to provide a structure for the user to browse through. Therefore, on one side our
categorization scheme is initially based on 36 categories of software assets that support applied game development
and will be delivered through the RAGE ecosystem.
3.5. SVM Classifier
To categorize the documents in RAGE using the extracted categories in the NER component, an SVM classifier is
applied. Due to the potentially complex structure of the assets, we prefer an SVM-based approach to
Fig. 2. NLP-IR Workflow
212 Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
approaches of text categorization such as probabilistic or linear techniques as described by Sebastiani.
To categorize
the RAGE content based on the classification scheme outlined above, an SVM classifier will be trained
on the domain
specific NE-chunks produced by the NER component. With this approach we hope to increase the
of the classification compared to the full-text approach with regard on the RAGE use cases. To implement the
classifier we will rely on the state of the art SVM library LibSVM.
3.6. Evaluation
Although the evaluation strategy will be presented in detail i
n a future paper, here we share the main idea of how
we are going to evaluate our designs. During the implementation we are going to evaluate cloud user experience using
ctured interviews to discover the benefits and drawbacks of our cloud-based approach. For this we will design
tandard-use cases and a test collection to be tested with users of
the RAGE project in a case study. The user-based
evaluation is intended to compare the cloud and the non-cloud approach. Besides utilizing those stakeholder-
interviews to evaluate the cloud approach, we are going to use a semi-supervised standard evaluation methodology to
ate the quality of the NLP-based classification and IR system based on test collections of the RAGE project.
hese standard evaluation methodologies (e.g., recall, precision, F-measure) are described by, among others, Croft,
Metzler et al.
and Manning, Raghavan et al.
and are used to answer research question 3, whether an NLP (NER)-
and classification-based IR approach outperforms the bag-of-words IR approach.
4. Towards Cloud Support
In the previous sections we described the basic ideas of the RAGE ecosystem and the NLP-IR support for knowledge
capturing in the heterogeneous content within it. Now we want to discuss how cloud technology can enhance our
cloud approach to improve the outcome of the overall system. In the first step we take a look at the several sub-systems
the NLP-IR architecture and discuss how they can be implemented using
cloud technology. In the second step we
present a common architecture that
integrates both cloud and non-cloud subsystems to implement a complete system.
Our discussion is primarily based
on the NIST cloud definition.
In this paper we focus on cloud support in the NLP-
IR system, but we also address
aspects of data distribution in the RAGE storage cloud.
4.1. Use Cases for Cloud Support
In the following section we investigate which possible use cases for cloud support exist within the NLP-IR system.
For that, we consider some of the Essential Characteristics” proposed by NIST in their widely used definition of
cloud computing. Taking a look at them, it comes to mind that the Rapid Elasticitycharacteristic in particular may
be usefully applied to our overall NLP-IR and will lead to a reasonable use case. Both NER and SVM classification
are tasks that may require a high volume of CPU time depending on the volume and complexity of the input data.
the same time, both components are only in use when new (heterogeneous) content is added to the RAGE ecosystem,
which has to be classified in order to support knowledge capturing and access support, as described
above. This
means that there is a high volatility in this component: While most of the time there is no workload on the
there are other (relatively short) times where the CPU load is relatively high when new content has to be
analyzed and
classified. By using rapid elasticity, an efficient and goal-oriented application of the resources can be
Combined with the measured service characteristic in a third-party model, rapid elasticity may reduce costs
computing power in our use case while increasing usability by providing a suitable amount of CPU time at the
time. As the RAGE ecosystem and NLP-IR in the actual architecture are designed as non-cloud web services
that are
used by thin clients, the broad network access characteristic of cloud computing also may be applied in
a cloud-
based approach. For the end user there will be no distinction between a cloud and a non-cloud RAGE/NLP-IR
system. In the use case described above, the two other characteristics resource pooling” and on demand
service” in our architecture also primarily refer to provision and use of CPU resources for NER and SVM. If
the user
needs them, he doesn’t have to care about the underlying resources (on demand), while the provider, in house or third
party, may share the resources using resource pooling with ot
her (scientific) cloud projects. Based on these
Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
considerations, there are good reasons to develop a cloud-based version of NLP-IR in RAGE. Therefore, the strongest
characteristic indeed is the rapid elasticity, which was mentioned first.
4.2. Service Model
The next step is to decide which service paradigm proposed by NIST should be used in the architecture. For
both the RAGE storage cloud with data distribution and the RAGE computing cloud with NLP-IR-INPUT, NLP-
IR-HPC and NLP-IR-OUTPUT system we decided to utilize the platform as a service (PaaS) paradigm throughout
the complete pipeline. This paradigm allows us to use a suitable abstraction of the underlying hard- and software
resources while allowing us to develop our own solution based on the technology of the runtime environment. As
candidates for the underlying NLP/NER framework are both based on Java technology, we will have to choose
scalable cloud-based Java runtime environment.
4.3. Deployment Model
The NIST definition gives four different deployment models of clouds: public, private, community and hybrid. In
our architecture we have chosen the community cloud, as the applied gaming developer community is a “classic
example” of a community described by NIST: by a specific community of consumers from organizations that have
shared concerns (e.g., mission, security requirements, policy, and compliance considerations).”
The cloud may be
managed by a stakeholder of the applied gaming developer community or an external cloud provider, depending on
the resources needed for the implementation. The question of which kind of provider model is chosen influences if
characteristics measured service” and resource pooling” may be used to add value to the project, as both primarily
refer to providers who operate a broad cloud infrastructure. As indicated, the community cloud approach allows the
users to deal with their core business (AG development) without having to deal with IT management,
as all necessary
services are provided by the RAGE storage cloud and the NLP-IR cloud. If the community cloud is
operated by a
trusted third party, it is even possible to operate multiple distinct instances of the RAGE cloud services
for competing
communities in one cloud environment. This allows the deployment of distinct RAGE instances as a
whole in different
sub-communities of the applied gaming c ommunity.
Besides the pure community cloud approach, a hybrid approach seems to be also constructive. In this deployment
e mission-critical resources and specializations are deployed in
a community cloud. In the use case, this applies to
the NLP and IR resources as specialized, for example, as well as to the storage of secret and business-critical company
ormation. While these critical assets remain internal to the community, public cloud services may be utilized to
tegrate open-source services for software developme
nt, presentation and storage of non-critical assets.
4.4. Cloud-based NLP-IR Architecture
Our architecture is divided into four cloud-based subsystems: the STORAGE CLOUD, NLP-IR-INPUT, HPC and
NLP-IR-OUTPUT subsystems. In our actual approach, these four subsystems are distinct and distributed due to their
heterogeneous structures. The four systems may be divided into two subgroups: data distribution and cloud computing.
The data distribution subgroup contains the RAGE contents, whereas the cloud computing subgroup is used to provide
the services of the NLP-IR architecture; the NLP-IR system has to deal with a high I/O load. The focus of the
system is primarily to provide CPU time. The NLP-IR-OUTPUT system will be optimized for database
and index
access. In a later development stage after evaluating cloud platforms, we may come to the conclusion that
two or three
subsystems may be combined on one common cloud platform.
214 Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
The cloud storage offers all the data sources of the RAGE system in a distributed environment. A data distribution
approach follows the classification of the RAGE knowledge sources shown in section 1. The advantages of the cloud
storage co
mpared to non-cloud storage are the scalability and adaptability of
the storage regarding the heterogeneous
content. The storage infrastructure is hidden from the user and tailored to the contents, such as multimedia with high
olume and I/O load or textual documents with smaller file sizes. The end user and following cloud services just use
tandard service-based interfaces. In the RAGE storage cloud we will rely on state-of-the-art cloud storage techniques.
The task of the NL-IR-INPUT subsystem is similar to the task of the adaptor layer of our
non-cloud approach. Its
main objective is to provide an interface to the cloud storage and the RAGE subsystems
within it. Compared to our
original approach, a second component has been added to the NLP-IR-INPUT subsystem: the infrastructure adaptor.
his component is used to connect the cloud-based NLP-IR system to cloud- as well as non-cloud-based RAGE
ystem repositories using appropriate abstraction techniques. Besides the infrastructure adaptor, you find the
ntent adaptor, which fulfills the normalization of the heterogeneous contents for NER as
described above.
The HPC subsystem is the core system of our cloud-based NLP-IR architecture, as it provides the main functionality
of the system and which has the most benefit of the cloud deployment due to the rapid elasticity characteristic. It
contains both the NER and the SVM components, which require high CPU load while in use. The integration of NER
and SVM in this layer will follow state-of-the-art NLP/SVM cloud-based implementations, as mentioned in related
The OUTPUT subsystem contains the IR components (index, tax
onomy, user interface). Here we are planning to
use a state-of-the-art cloud-based IR technology, as provided by Apache SOLR, for example. Due to our special needs,
this server application will have to be tailored to our domain-specific issues.
5. Conclusion
We have introduced a design for a cloud-based architecture for knowledge extractio
n in a game industry knowledge-
sharing platform. The intended architecture will extract named
entities from the data and use these named entities for
document categorization so that they can be applied in a taxonomy for faceted search. Regarding this architecture, we
have identified three research questions. Also, we have hinted at a possible evaluation of our design to answer these
research questions. The research is still in an early stage, and the evaluation scenarios will be elaborated further in the
ear future.
Fig. 3. NLP-IR Cloud Architectural Overview
Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
As we have shown, the NLP-IR in the RAGE ecosystem has the potential to be
implemented using cloud
technology to optimize use cases, particularly because in use cases that contain high volumes of heterogeneous data
and require easy and fast scalability a cloud support is recommended. From an administrator’s perspective, a cloud-
implementation offers more flexibility and usability due to the reduction in installation and operation efforts.
This allows users
to concentrate on their primary objective of developing applied games (which is the main goal of the
RAGE project).
As a first consequence, the component design of our non-cloud based approach is tailored so that we
will use
standard components which are cloud ready, as mentioned above, to ultimately ensure easy
implementation of a
non-cloud solution as well as a cloud-based solution.
Acknowledgements and Disclaimer
This publication has been produced in the context of the RAGE project. The project has received funding from
the European Union’s Horizon 2020 research and innovation programme under
grant agreement No 644187. However,
this paper reflects only the author’s view and the European
Commission is not responsible for any use that may be
made of the information it contains.
1. Nonaka I, Takeuchi H. The knowledge-creating company: How Japanese companies create the dynamics of innovation.
New York: Oxford University Press; 1995.
2. Salman M, Hemmje M, Heutelbeck D. Towards Social Network Support for an Applied Gaming Ecosystem. In: Acc
for: ECGBL, European Conference on Games Based Learning; 2015
3. Salman M, Hemmje M, Heutelbeck D, Fuchs M, Brocks H. Towards Social Media Platform Integration with an Applied
aming Ecosystem. In: Submitted to: SOTICS, The Fifth International Conference on Social Media Technologies,
Communication, and Informatics; 2015.
4. Educational Portal; 2015. Available at: [accessed
5. Sebastiani F. Machine Learning in Automated Text Categorization. A
CM Comput. Surv. 2002;34(1):147.
6. Mohri M, Rostamizadeh A, Talwalkar A. F
oundations of Machine Learning: The MIT Press
7. Schrimpf S. APARSEN - Alliance Permanent Access to the Records of Science in Europe Network. Dia
log mit Bibliotheken
8. Nobata C, Cotter P, Okazaki N, Rea B, Sasaki Y, T
suruoka Y, et al. Kleio: A Knowledge-enriched Information Retrieval
System for Biology. In: Proceedings of the 31st Annual International ACM SIGIR Conference on Research and
Development in Information Retrieval. New York, NY, USA: ACM; 2008, p. 787788.
9. Ananiadou S, Thompson P, Thomas J, Mu T, Oliver S, Rickinson M, et al. Supporting the education evidence portal via text
ining. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences
10. Joachims T. Text categorization with Support Vector Machines: Learn
ing with many relevant features. In: Nédellec C,
Rouveirol C, editors. Machine Learning: ECML-98: Sp
ringer Berlin Heidelberg; 1998, p. 137142.
11. Filippova K, Hall KB. Improved Video Categorization from Text Metadata and User Comments. In: Proceedings of the 34th
International ACM SIGIR Conference on Research and Development in Information Retrieval. New York, NY, USA: ACM;
2011, p. 835842.
12. Tobias Swoboda. Towards effectivity augmentation of automated sc
ientific document categorization by continuous
feedback, Master Thesis. Hagen; 2014.
13. Ritter A, Clark S, Mausam, Etzioni O. Named Entity Recognition in Tweets: An Experimental Study. In: Pro
ceedings of the
Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA Association for Computational
Linguistics; 2011, p. 15241534.
14. Dascalu M, Dessus P, Trausan-Matu Ş, Bianco M, Nardy A. ReaderBench, an Environment for Analyzi
ng Text Complexity
and Reading Strategies. In: Lane H, Yacef K, Mostow J, Pavlik P, editors. Artificial Intelligence in Education: S
Berlin Heidelberg; 2013, p. 379388.
15. Dit B, Revelle M, Gethers M, Poshyvanyk D. Feature lo
cation in source code: a taxonomy and survey. Journal of Software:
Evolution and Process 2013;25(1
16. Tablan V, Roberts I, Cunningham H, Bontcheva K. GATECloud. net: a platform for large-scale, open-source text processing
on t
he cloud. Philosophical Transactions of the Royal Society of London A: Mathematical, Physical and Engineering
216 Christian Nawroth et al. / Procedia Computer Science 68 ( 2015 ) 206 – 216
Sciences 2013;371(1983):20120071.
17. Klenner A, Bergmann S, Zimmermann M, Romberg M. Large scale chemical patent mining with UIMA and UNICORE.
urnal of Cheminformatics 2012;4(1):12.
18. Evangelinos C, Hill C. Cloud computing f
or parallel scientific HPC applications: Feasibility of running coupled atmosphere-
ocean climate models on Amazon’s EC2. ra
tio 2008;2(2.40):234.
19. Pouladzadeh P, Shirmohammadi S, Bakirov A, Bulut A, Yassine A. Cloud-based SVM for food categorization. M
Tools and Applications 2015;74(14):524360.
20. Apache Foundation. SolrCloud; 2014. Available at: https://cwiki.a [accessed
21. Guo J, Xu G, Cheng X, Li H, editors. Name
d entity recognition in query: ACM; 2009.
22. Nadeau D, Sekine S. A survey of named entity recognition and classification. Lingvisticae Investigationes 2007;30(1):326.
23. Manning CD, Surdeanu M, Bauer J, Finkel J, Bethard SJ, McClo
sky D. The Stanford CoreNLP Natural Language
Processing Toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System
monstrations; 2014, p. 5560.
24. Cunningham H, Maynard D, Bontcheva K, T
ablan V. GATE: A Framework and Graphical Development Environment for
Robust NLP Tools and Applications. In: Proceedings of the 40th Anniversary Meeting of the Association for Computational
Linguistics (ACL’02); 2002.
25. Ferrucci D, Lally A. UIMA: an architectural approach to unst
ructured information processing in the corporate research
environment. Natural Language Engineering 2004;10(3-4):32748.
26. Chang C, Lin C. LIBSVM: A library
for support vector machines. ACM Transactions on Intelligent Systems and Technology
27. Mell P, Grance T. The NIST definition of cloud computing. National Institute of Standards and Technology 2009;53(6):50.
28. Croft WB, Metzler D, Strohman T. Search engines: Information retrieval in practice. Boston: Addison-Wesley; 2010.
29. Manning CD, Raghavan P, Schütze H. In
troduction to information retrieval. New York: Cambridge University Press; 2008.
... Besides the research into TC itself, these techniques can also benefit researchers using VREs, as new publications in the digital library can be found by using this visual interface to identify NERs and found TC categories for all existing and new documents. In [16] we discussed our ideas on combining named entity recognition (NER) and support vector machine (SVM)-based TC in a domain-specific information retrieval (IR) system as part of an overall environment called AGNERS in the EU research project RAGE. In this system design, three use cases are being implemented: (1) related entity finding [1], (2) faceted search and (3) traditional keyword search. ...
... The system model shown in [16] has been refined and extended. It now contains a NLP/IR-research-visualization component. ...
... We think that saliency is the Fig. 3. The extended model based on the AGNERS IR in [16] suitable filter and ranking criterion for the intended use case of investigating the correlations between named entities and classification results. ...
Conference Paper
Full-text available
In big data analytics, visualization and access are central for the creation of knowledge and value from data. Interactive visualizations of analysis of structured data are commonplace. In this paper, information visualization and interaction for text analysis are addressed. The paper motivates this issue from a data usage standpoint, gives a survey of approaches in the area of interactive visualization of text analytics, and presents our proposal of a specific solution design for visual interaction with results from a combination of named entity recognition (NER) and text categorization (TC). This matrix-based model illustrates abstract views on complex relationships between abstract entities and is exemplary for any combination of feature extraction and TC. The aim of our proposal is to support feature extraction and TC researchers in distributed virtual research environments by providing intuitive visual interfaces.
... This whill allow its key participants to effectively and efficiently collaborate and access the associated innovative business cases and commercial opportunities [5]. To achieve this goal, RAGE will refer to Social Network Analysis (SNA) by mean of applying technologies for Natural Language Analysis (NLA) for discourse analysis, as well as Semantic Representation and Annotation (SRA) of its results [6]. This will help to extend the envisioned Ecosystem with features of a social mediation engine going beyond content syndication, i.e. it will serve as a social space that mediates collaboration partners, while content remains the main attractor [5]. ...
... This will help to extend the envisioned Ecosystem with features of a social mediation engine going beyond content syndication, i.e. it will serve as a social space that mediates collaboration partners, while content remains the main attractor [5]. But, RAGE is facing a greater challenge which is Information Overload (IO) due to the constant flow of web resources [7], the heterogeneous nature [6] and diversity of its contents [8], and the increasing number of tools and channels used to manage them [9]. This means for RAGE that new innovation requirements will generate new imports into Ecosystem which will saturate the target system with large piles of relevant materials, hence leading the user into overload. ...
... IO can decrease innovation because it has a negative impact on work productivity (interruptions [8], distractions [10]) and individual behaviour (like anxiety [11], infobesity [12]). Automated knowledge extraction can help tackle IO because it can retrieve useful knowledge (in form of named entities) from naturally written texts automatically, which can be used for efficiently searching and reasoning about existing documents in various domains on the Web [6]. Existing techniques dealing with IO have often focused on developing search tools [13,14] that harvest the Web and/or construct complicated queries to find useful documents. ...
Full-text available
The RAGE research project will provide access to a wide range of software assets for Applied Gaming (AG) enabling AG software developers to better collaborate, share their knowledge, and be able to react to new requirements and trends more e�ciently. But, RAGE is facing a greater challenge which is Information Overload (IO) because of the permanent flow of new documents published daily on the Web including their different formats. To solve IO, existing contributors have used complicated mechanisms that index the Web and harvest large piles of documents to detect relevant materials. Others have used Semantic Web Technologies to enrich web resources with additional facts and meaning, but have let the system (and not the end-user) decide about the relevance of the search result failing to satisfy the users' requirements. Thus, we propose a novel software architecture that combines NLP (by mean of Named Entity Recognition) and Domain Ontology to support searching and browsing (faceted-search), as well as reasoning about named entities on the Semantic Web. AG software developers will get only relevant information according to their speci�c need without much time spent on information harvesting.
... This evidence can then be used to recommend treatments and provide rational arguments as to why specific treatments are suggested [EEH19]. The identification of evidences in medical text resources is generally hindered by the permanent growth of medical information [Byy12] and emerging named entities (eNEs, i.e., new terms appearing suddenly and unexpectedly [PG11] in medical datasets like PubMed [Naw+15]. ...
... Recent NER techniques are making use of user feedback through crowdsourcing which can also optimize the result of IR systems [Fin+10]. These methods are using, besides ML, other state of the art NLP techniques, such as Part-of-Speech (POS) tagging, regular expressions, or dependency parsing [Naw+15]. In his recent PhD thesis, Christian Nawroth [Chr21] relied on a similar approach. ...
Full-text available
This dissertation addresses the challenges of accessing textual resources and the resulting Information Overload faced by software developers on the Web. A new approach based on methods of Named Entity Recognition and rule-based Document Classification is proposed to facilitate the extraction of domain-related named entities, enabling faceted search and browsing in an innovative content and knowledge management ecosystem portal. The concept of the approach with model designs, overall architecture, and the initial prototype is described. Finally, several evaluations are shown to demonstrate the feasibility of the approach and usability of the prototype.
... Web crawling can be used for systematically analyzing web pages, starting with yellow pages for qualification offers, e.g., Higher Education Compass 4 . Pages are automatically classified for relevance for the SQW, using the NLP technique of named entity recognition (NER) [16]. ...
... Sound experience is mandatory and agreed-on engineering guidelines are scarce. ML libraries like scikit-learn 10 16 . ML cheat sheets provide an overview of major ML concepts and how they interrelate. ...
Full-text available
Machine intelligence, a.k.a. artificial intelligence (AI) is one of the most prominent and relevant technologies today. It is in everyday use in the form of AI applications and has a strong impact on society. This article presents selected results of the 2020 Dagstuhl workshop on applied machine intelligence. Selected AI applications in various domains, namely culture, education, and industrial manufacturing are presented. Current trends, best practices, and recommendations regarding AI methodology and technology are explained. The focus is on ontologies (knowledge-based AI) and machine learning.
... Until now, factual diagnosis argumentation knowledge has been mostly based on textual information [19] [16] [20][1] [17] [18]. However, particularly in medical applications, visual argumentative evidence can contribute many additional features to diagnosis [ [31]. ...
Conference Paper
Full-text available
In this paper, we present a hybrid and interdisciplinary approach for the calculation of argument strength, based on the coupling of two existing frameworks-the emerging Argument Recognition (eAR) framework of: (1) the emerging Named Entity Recognition Information Retrieval System (eNER-IRS) for textual medical articles, and (2) the Generic Multimedia Analysis Framework (GMAF) for semantic extraction of visual Multimedia features. We focus on combining textual and visual features to increase the capability of both frameworks and facilitate applications with a higher level of confidence in argumentation.
... Further databases identified were grouped into category "Other" (14% -10 studies), for example, Wikis [18] and Tweets [25]. ...
Conference Paper
Context: A large number of information is generated and manipulated in Software Engineering (SE) projects. The technology surrounding this domain is constantly evolving. To keep up with such evolution , developers share their knowledge and seek help from other developers by means of interactive and collaborative environments. Understanding and extracting knowledge from these environments can enable developers to identify useful information for the project. Objective: This work aims to identify the main textual analysis approaches to extract useful information in the SE. Method: To achieve the proposed objective, we conducted a Systematic Mapping (SM). Results: We analyzed 69 relevant primary studies addressing approaches to extract useful information in the SE. Conclusion: Among the main conclusions of this study, we can infer that discussion forums attracted a significantly attention in SE context and it becomes one of the main textual databases investigated to extract useful information.
... In contrast, machine-learning-based approaches exhibit superior performance over dictionary-based approaches, particularly machine learning algorithms, support vector machines (SVMs) (Tsochantaridis et al. 2005;Nawroth et al. 2015), hidden Markov models (HMMs) (Sarkar and Shaw 2017), CRFs (Arnab et al. 2015;Zeng et al. 2017) and maximum entropy Markov models (MEMMs) (Finkel et al. 2004). These approaches consider NER as a sequence labeling problem, where the goal of each algorithm is to search the best label sequence for a given input sentence. ...
Full-text available
Many detailed geoscience reports lie unused, offering both challenges and opportunities for information extraction. In geoscience research, geological named entity recognition (GNER) is an important task in the field of geoscience information extraction. Regarding numerical geoscience data, research on information extraction remains limited. Most conventional NER approaches are heavily dependent on feature engineering, and such sentence-level-based methods suffer from the tagging inconsistency problem. Based on the above observations, this paper proposes a neural network approach, namely, attention-based bidirectional long short-term memory with a conditional random field layer (Att-BiLSTM-CRF), for name entity recognition to extract information entities describing geoscience information from geoscience reports. This approach leverages global information learned from an attention mechanism to enforce tagging consistency across multiple instances of the same token in a document. Experiments on the constructed dataset show that our method achieves comparable performance to that of other state-of-the-art systems. Additionally, our method achieved an average F1 score of 91.47% in the NER extraction task.
... The EP tool suite was developed as a solution for the growing need of scientific communities to manage their document and media collections as well as their educational and other kinds of knowledge resources" (cf. [9]). The EP is a flexible software solution, that is built around the content management functionality that is provide by the Typo3 framework. ...
Conference Paper
Analytics and visualisation of Big Data appearance is still a challenging task. In this paper, the management and visualisation of big data streams will be discussed at hand of data that comes into existence during the care of people with dementia in their own homes. Therein, basic requirements are explored towards the development of a data management, analytics and visualisation platform stemming from application scenarios in which various data streams are created, processed, analysed, visualised and stored for ad hoc or later reuse. The platform will be realised on open ICT standards, implemented within the EC co funded SenseCare project.
Main problems occurring in Product-Service Systems (PSSs), are due to an inadequate requirements analysis and lack of a strong PSS conceptual design. Problems vary from exceeding budgets, to missing functionalities, unsuccessful market launch, or even project abortion. Furthermore, the special characteristics of a PSS have to be considered already at an early stage of the development process. Requirements Engineering (RE) and design methodology as well as supporting Information and Communication Technologies (ICT) need to establish a common perception of the targeted PSS. At the same time, the inner complexity of PSS leaves requirements analysis, design activities and development tasks fragmented among many disciplines and sometimes conflicting, unstable, unknowable or not fully defined. In this context, a concurrent, transdisciplinary and collaborative design of PSS is required to create feasible and successful solutions. The objective of this chapter is to present a structured approach to face the specific challenges of PSS development in detail, to elaborate a general framework that features a systematic approach for PSS development, and to consider the effects of changes in specific product and service design on a systematic PSS development process.
Conference Paper
Full-text available
The MAT project aims to develop and evaluate an initial set of algorithms that can detect agitation, restlessness and aggression in dementia patients. MAT uses vision based analytics to track a subjects facial expressions in real-time. The first version of MAT has implemented two use cases for the detection of restlessness and aggression. The project also has the potential to be used in more advanced machine learning and data analytics applications typically for research purposes on elder care. Data sets for each subject’s monitoring period are generated in CSV file format which could be used to populate a database or as inputs to machine learning classification algorithms/platforms.
Conference Paper
Full-text available
We describe the design and use of the Stanford CoreNLP toolkit, an extensible pipeline that provides core natural language analysis. This toolkit is quite widely used, both in the research NLP community and also among commercial and government users of open source NLP technology. We suggest that this follows from a simple, approachable design, straight-forward interfaces, the inclusion of robust and good quality analysis components, and not requiring use of a large amount of associated baggage.
Full-text available
As people across the globe are becoming more interested in watching their weight, eating more healthily, and avoiding obesity, a system that can measure calories and nutrition in everyday meals can be very useful. Recently, due to ubiquity of mobile devices such as smart phones, the health monitoring applications are accessible by the patients practically all the time. We have created a semi-automatic food calorie and nutrition measurement system via mobile that can help patients and dietitians to measure and manage daily food intake. While segmentation and recognition are the two main steps of a food calorie measurement system, in this paper we have focused on the recognition part and mainly the training phase of the classification algorithm. This paper presents a cloud-based Support Vector Machine (SVM) method for classifying objects in cluster. We propose a method for food recognition application that is referred to as the Cloud SVM training mechanism in a cloud computing environment with Map Reduce technique for distributed machine learning. The results show that by using cloud computing system in classification phase and updating the database periodically, the accuracy of the recognition step has increased in single food portion, non-mixed and mixed plate of food compared to LIBSVM
Full-text available
In our project UIMA-HPC, we describe the combined usage of Unstructured Information Managment Architecture (UIMA) and Uniform Interface to Computing Recources (UNICORE) for large-scale chemical patent mining. Our approach incorporates existing software such as chemoCR for image processing (image-to-structure) and OCR for text reconstruction. All components are wrapped inside the UIMA framework pipeline. Using the UIMA framework ensures compatibility between different components of the pipeline and makes it possible to connect arbitrary annotation modules into this system.