Conference PaperPDF Available

Inflectional Review of Deep Learning on Natural Language Processing


Abstract and Figures

In the age of knowledge, Natural Language Processing (NLP) express its demand by a huge range of utilization. Previously NLP was dealing with statically data. Contemporary time NLP is doing considerably with the corpus, lexicon database, pattern reorganization. Considering Deep Learning (DL) method recognize artificial Neural Network (NN) to nonlinear process, NLP tools become increasingly accurate and efficient that begin a debacle. Multi-Layer Neural Network obtaining the importance of the NLP for its capability including standard speed and resolute output. Hierarchical designs of data operate recurring processing layers to learn and with this arrangement of DL methods manage several practices. In this paper, this resumed striving to reach a review of the tools and the necessary methodology to present a clear understanding of the association of NLP and DL for truly understand in the training.
Content may be subject to copyright.
International Conference on Smart Computing and Electronic Enterprise. (ICSCEE2018) ©2018 IEEE
Inflectional Review of Deep Learning on Natural
Language Processing
SK Ahammad Fahad
Faculty of Computer and Information Technology
Al-Madinah International University
Shah Alam, Malaysia
Abdulsamad Ebrahim Yahya
Faculty of Computing and Information Technology
Northern Border University
Rafha, KSA
Abstract— In the age of knowledge, Natural Language
Processing (NLP) express its demand by a huge range of
utilization. Previously NLP was dealing with statically data.
Contemporary time NLP is doing considerably with the corpus,
lexicon database, pattern reorganization. Considering Deep
Learning (DL) method recognize artificial Neural Network (NN)
to nonlinear process, NLP tools become increasingly accurate
and efficient that begin a debacle. Multi-Layer Neural Network
obtaining the importance of the NLP for its capability including
standard speed and resolute output. Hierarchical designs of data
operate recurring processing layers to learn and with this
arrangement of DL methods manage several practices. In this
paper, this resumed striving to reach a review of the tools and the
necessary methodology to present a clear understanding of the
association of NLP and DL for truly understand in the training.
Efficiency and execution both are improved in NLP by Part of
speech tagging (POST), Morphological Analysis, Named Entity
Recognition (NER), Semantic Role Labeling (SRL), Syntactic
Parsing, and Coreference resolution. Artificial Neural Networks
(ANN), Time Delay Neural Networks (TDNN), Recurrent Neural
Network (RNN), Convolution Neural Networks (CNN), and
Long-Short-Term-Memory (LSTM) dealings among Dense
Vector (DV), Windows Approach (WA), and Multitask learning
(MTL) as a characteristic of Deep Learning. After statically
methods, when DL communicate the influence of NLP, the
individual form of the NLP process and DL rule collaboration
was started a fundamental connection.
Keywords—Deep Learning; Natural language processing; Deep
nural Network; Multitask Learning
I. I
In the age of information, Natural Language Processing
(NLP) create its demand by a comprehensive area of
application. To present the significant knowledge to non-
programmer from computer system Natural Language
Processing was deliberate as a working field from 1950. Non-
Subject material experts obtain an answer of simple queries by
improvement of NLP. Previously NLP was dealing with
statically data. Recent year NLP is doing well with the corpus,
lexicon database, Neural Network. Since Deep Learning (DL)
method allow artificial Neural Network (NN) to nonlinear
process, Natural Language processing tools enhance more
accurate and valuable that make a revolution. Multi-Layer
Neural Network expanding the influence of the natural
language processing for its capability with decent acceleration
and reliable producing.
In the field of NLP, DL completely takeover Text
Classification and Categorization, Named Entity Recognition
(NER), Part-of-Speech Tagging (POST), Semantic Parsing
and Question Answering, Paraphrase Detection, Language
Generation and Multi-document Summarization, Machine
Translation, Speech Recognition, Character Recognition, Spell
Checking etc. Hierarchical representations of data operate
complicated processing layers to learn and have by this
imagined Deep Learning methods dominate many realms.
Those tools are powered by DL and applied on Natural
Language to complete the NLP process to achieve the goal.
Text Summarization, Caption Generation, Question
Answering, Text Classification, Machine Translation,
Language Modeling, Speech Recognition all NLP tools are
working with DL go obtain desire accurate result. Several
methods are proposed and the system was implemented to
NLP by DL and they are doing well.
NLP methods are modified when DL was associated. NLP
process like POS, NER, Morphology, Syntactic Parsing,
Coreference resolution is discussed in subsequent section.
Different type of Neural Network like; ANN, TDNN, RNN,
CNN are discussed with the relation and in NLP process was
discussed in this paper. Different technique and tools of DL
like; LSTM, MTL, DV, CBOW, VL, WA, SRL, Non-Linear
Function are discussed with the possible relation with NLP. In
this paper, it was trying to make a review of the tools and the
basic methodology. After statically methods, when DL take
over the control to NLP, a new form of the NLP process and
DL process collaboration was presented with the basic
During a document record recognized for accomplishing
besides, this is not continued to direct accomplishment. It
requires processing background for authentic execution. There
has some obligatory progression in contemporary Natural
Language Processing (NLP). Deep Learning (DL) is a superior
form of Neural Network (NN) and it deals with preprocessed
data. Before NN applied in NPL subsequent processing the
document. DL also operated among concocted document file.
That’s why unusual primary actions of the process of textual
documents are extremely valuable. There have six compulsory
International Conference on Smart Computing and Electronic Enterprise. (ICSCEE2018) ©2018 IEEE
rounds should ensue to perceive more sustainable and accurate
result by implementing DL. Those are; Splitting and
Tokenization, Part-of-Speech Tagging, Morphological
Analysis, Named Entity Recognition, Syntactic Parsing, and
Conference Resolution [7,8].
Splitting and Tokenization segment clean document from
outcast tags and absolute splitting document in to token. The
input file has any tags for designing the text. NLP consider
simply real fresh text for processing. Tags require cleaning
before performing a better and correct result in NLP.
Tokenization is that the approach of separating a stream of
text into words, phrases, symbols, or different principal
elements specified as tokens. How to split and character of the
token is should describe as per the demand for output.
Succeeding cleaning the writing from unwanted content and
splitting text to token, NLP proceeding by tokens [3].
Part of speech tagging (POST) or POS tagging is a
mechanism of NLP, it performs the extremely significant role
on text phrase, syntax, translation, and semantic analysis.
Rule-Based tagging perform POS tagging by match everything
on a lexical dictionary and match rest of words independently
with each part of speech precept. In Stochastic POS tagging is
accomplished by applying prospect including a tagged corpus.
Both Rule-Based and Stochastic POS tagging are worked
blended in Transformation-Based POS tagging [2].
Morphology allots with the connection between the
structure and purpose of words. Morphological Analysis is the
insignificant semantic systems including a lexical or a logical
definition of words. Morphological analyzers and lemmatizes
typically demand training data. The latter is practiced on
character-aligned combinations of stems and lemmas, where
stems are extorted and neophytes stem into lemmas [9].
Named Entity Recognition (NER) is the policy that
recognizes the name and number and eliminates from tokens
which are allowing for additional processing. Hand-Made
NER sharpens on obtaining name and numbers applying the
rules which are human-made. Those rules utilizing
grammatical syntactic and Orthographic innovations in
combination with dictionaries. Machine Learning-based NER
method applies a classification analytical pattern to determine
and converting description query into a distribution problem.
As tike POS tagging there has a Hybrid scheme developed
which one is the combination of human-made and machine-
made rules [6].
Syntactic Parsing produces comprehensive syntactic
analysis and Sentiment analysis including both ingredient and
dependence description with a compositional pattern atop trees
utilizing deep learning. WordNet® is an immense on-line
database of English. Nouns, verbs, adjectives and adverbs
section organized within collections of subjective peculiarity
synonyms (synsets), individual expressing a particular
representing. Synsets section interlinked with hints of
conceptual-semantic and lexical associations [6].
Coreference resolution is the responsibility of defining
semantic definitions that belong to the corresponding real-
world existence in natural language. Coreference consists of
two semantic appearances—antecedent and anaphor.
Fig. 1. Basic steps for Natural Language Processing for Deep Learning
The anaphor is the appearance whose interpretation (i.e.,
comparing by both pavement and theoretical real-world
article) depends on that of the differential phase. The
predecessor is the semantic appearance on which an anaphor
depends. Coreference resolution complete specifying
disclosure and both pronominal from token and dismissed
them [4,5].
A. Dense Vector
Artificial Neural Networks (ANN) Input textual data to the
data vector x in-dimensions and produce the product in out-
dimensions. DL distributes with raw words (tokens) and not
engineered features, the first layer has to map words into real-
valued vectors for processing by consequent layers of the NN.
Alternatively, forming a unique dimension for individual and
every feature, implant every feature into a D-Dimensional
space and describe as a Dense Vector (DV) in the space [1]. If
there exists a correlation distribution which can learn, that can
be captured by DV. It is the principal influence of DV
representation. DV training will cause comparable
characteristics have to experience vectors-information
between them [7].
International Conference on Smart Computing and Electronic Enterprise. (ICSCEE2018) ©2018 IEEE
B. Feature Extraction
For deriving a feature from tokens; Distance and
Dimensionality Feature, Continuous Bag of words (CBOW),
and Variable Length (VL) signify implemented. Distance can
measure by deducting the identical tokens word to feature
vector. Distance invariably positive, it can deduct whatever
from anyone. This measurement data can be used to train NN
and make DL more specific. Distance, the weight of token,
synset (synonyms) everything is a feature of the embedded
token and all of the dimension are property and those are
associated with processing speed, and accuracy. To obtain a
feature dimension is an important feature of a token. CBOW is
a feature based on dense vector feature extraction. NN get
trained from inputted token individually with a unique feature.
The weight of word executes this feature extends the
important. A number of unique words are proffered to vector
representation (One hot vector illustration) and it’s a
significant part of train DL [1]. Averaging embedded feature
and weight averaging of an embedded feature is essentially
allied to expand weight. DV has the primary tokens within the
sequence. But DL inadequate to supervise to variable length.
Consider an established measurement Windows Approach
(WA) can be a simplistic explication of this limitation. WA
strongly collaborate with POS but it not proper to operate with
Semantic Role Labeling (SRL) [7].
C. Deep Neural Network
Determination of basis of deep learning Neural Network
depends on the impression was operating to supervise to
accomplish the purpose. Non-Linear Function, Output
Transformation Function, Loss Function are numerously
accessible on NLP. NN is approximator and it can
approximate any non-zero liner function. Infinity number of
positive and negative range and a bound output range is the
property of a Non-Linear Function. In Outer Transformation
Function the peripheral function of NN use as a transmutation
function. To represent a Multi-class distribution, Outer
Transformation is pretty much recommended and used
extensively. How greatly the network encouraged from the
genuine output is was intimate by Loss Function. It depends
on the specification, character, and extent of tolerance.
Ranking loss, categorical entropy loss, log loss are the current
function of Loss Function [11]. Time Delay Neural Networks
(TDNN) can be the best alternative when fashioning for long-
distance dependencies. TDNN accumulated to local feature in
the deeper layer and tiny local (more global) function
inconsequently. TDNN perform a linear into tokens and it
accomplishes on POS, NER and further complicated steps of
NLP, like SLR [10]. Recurrent Neural Network (RNN) is
techno scientifically on acknowledged handwriting. It presents
dynamic ephemeral arbitrary by performing the primary
element of the network. RNN connection into blends
accompanies an addressed series. RNN further deal with
variable length and present involvement by timestamps [10].
Other hand Convolution Neural Networks (CNN) obtain token
to engrave feature by implementing convolution method and
build an artificial network. CNN produces output from
convolution to D Dimensional vector and transfers them to
choose the most appropriate feature of presented features
called pooling. Deciding the best feature by pooling it
developed the ANN for text classification. CNN good for
clustering but starving on learning sequential knowledge.
When RNN train neural nets by backpropagation algorithm,
vanishing gradient problem is occurring. To supervise this
problem, Long-Short-Term-Memory (LSTM) is implemented.
LSTM control log-logistic function and select the parameters
D. Multi-Tasking Learning
Multitask learning (MTL) in Deep Neural Network is the
method where NN perform several learning processes at the
identical time to gain a complementary advantage. In NN,
associated task particular feature is essential for another
feature. In NLP, POS prediction feature activity is also
accomplished for SRL and NER. Adjustment or upgrade to
POS task also generalization for SRL with NER. In the deep
layers of architected NN automatically learn. NN get trained
on related task by according deep layers. Maximum the time,
latest layer in the network is liability specified and according
to the layer enhance the performance. In MTL tasks, cascading
feature is the numerous dynamic way to accomplish the
aspired output. Use task feature to another task is obvious to
use an SLR and POS classifier and use the result to another
feature to train a parser [11]. When several tasks labeled for in
one dataset, shallow manner can be applied to the task jointly
and one all task labels at the same time by a unique model. In
shallow joint training can improve joint training on POS
tagging and noun-phrase chunking task. Relation extraction,
parsing, and NER can be jointly training in the statically
parsing model to improve the achievement.
Natural Language Processing (NLP) formulate its demand
with Deep Learning (DL) method. Artificial Neural Network
(NN) and non-linear process, Natural Language processing
tools enhance increased accurate and efficient and DL rule
collaboration was introduced with the primary relationship. In
the field of NLP, DL completely takeover Text Classification
and Categorization, Named Entity Recognition (NER), Part-
of-Speech Tagging (POST), Semantic Parsing and Question
Answering, Paraphrase Detection, Language Generation and
Multi-document Summarization, Machine Translation, Speech
Recognition, Character Recognition, Spell Checking etc. NLP
methods are modified when DL was associated. NLP process
like POS, NER, Morphology, Syntactic Parsing, Coreference
resolution are associated with Neural Network like; ANN,
TDNN, RNN, CNN are discussed with the relation and in
NLP process and tools of DL like; LSTM, MTL, DV, CBOW,
VL, WA, SRL, Non-Linear Function maintain relationship
with NLP process to collaborating basic relation.
[1] J. Schmidhuber, "Deep learning in neural networks: An overview",
Neural Networks, vol. 61, pp. 85-117, 2015.
International Conference on Smart Computing and Electronic Enterprise. (ICSCEE2018) ©2018 IEEE
[2] M. Hjouj, A. Alarabeyyat and I. Olab, "Rule Based Approach for Arabic
Part of Speech Tagging and Name Entity Recognition", International
Journal of Advanced Computer Science and Applications, vol. 7, no. 6,
[3] S. Fahad, "Design and Develop Semantic Textual Document Clustering
Model", Journal of Computer Science and Information Technology, vol.
5, no. 2, 2017.
[4] J. Zheng, W. Chapman, R. Crowley and G. Savova, "Coreference
resolution: A review of general methodologies and applications in the
clinical domain", Journal of Biomedical Informatics, vol. 44, no. 6, pp.
1113-1122, 2011.
[5] A. Kaczmarek and M. Marcińczuk, "A preliminary study in zero
anaphora coreference resolution for Polish", Cognitive Studies | Études
cognitives, no. 17, 2017.
[6] W. Yafooz, S. Abidin, N. Omar and R. Halim, "Dynamic semantic
textual document clustering using frequent terms and named entity",
2013 IEEE 3rd International Conference on System Engineering and
Technology, 2013.
[7] H. Li, "Deep learning for natural language processing: advantages and
challenges", National Science Review, vol. 5, no. 1, pp. 24-26, 2017.
[8] S. Fahad and W. Yafooz, "Review on Semantic Document Clustering",
International Journal of Contemporary Computer Research, vol. 1, no. 1,
pp. 14-30, 2017.
[9] O. Bonami and B. Sagot, "Computational methods for descriptive and
theoretical morphology: a brief introduction", Morphology, vol. 27, no.
4, pp. 423-429, 2017.
[10] G. Goth, "Deep or shallow, NLP is breaking out", Communications of
the ACM, vol. 59, no. 3, pp. 13-16, 2016.
[11] R. Gunderman, "Deep Questioning and Deep Learning", Academic
Radiology, vol. 19, no. 4, pp. 489-490, 2012.
... ANNs are inspired by the structure of the human brain to learn from the data, and they consist of multiple neurons that interact between each other via weighted connections (Abiodun et al., 2018). ANNs models are data-hungry, as the scale of training data increases the more accurate forecasting (Fahad and Yahya, 2018). There are multiple types of ANN structure such as Feed Forward Neural Network (FFNN) (Svozil et al., 1997), Radial Basis Functional Neural Network (Dash et al., 2016), and Multilayer Perceptron (MLP) (Agirre-Basurko et al., 2006). ...
... DL is a type of neural networks that can learn representation of the raw data, and solve complex problems by using multiple processing hidden layers in the network (Dara and Tumma, 2018). Recently, DL models have achieved great success in diverse applications such as computer vision (Chai et al., 2021), time-series forecasting (Torres et al., 2021), NLP (Fahad and Yahya, 2018), computational linguistics (Latif et al., 2020) and self-driving cars (Grigorescu et al., 2020). ...
The building sector accounts for 36 % of the total global energy usage and 40% of associated Carbon Dioxide emissions. Therefore, the forecasting of building energy consumption plays a key role for different building energy management applications (e.g., demand-side management and promoting energy efficiency measures), and implementing intelligent control strategies. Thanks to the advancement of Internet of Things in the last few years, this has led to an increase in the amount of buildings energy related-data. The accessibility of this data has inspired the interest of researchers to utilize different data-driven approaches to forecast building energy consumption. In this study, we first present state of-the-art Machine Learning, Deep Learning and Statistical Analysis models that have been used in the area of forecasting building energy consumption. In addition, we also introduce a comprehensive review of the existing research publications that have been published since 2015. The reviewed literature has been categorized according to the following scopes: (I) building type and location; (II) data components; (III) temporal granularity; (IV) data pre-processing methods; (V) features selection and extraction techniques; (VI) type of approaches; (VII) models used; and (VIII) key performance indicators. Finally, gaps and current challenges with respect to data-driven building energy consumption forecasting have been highlighted, and promising future research directions are also recommended.
... Recently, deep learning (DL) has developed rapidly, and it has been widely used in computer vision and natural language processing [24,25]. DL has attracted much attention in terms of HU due to its strong feature representation and learning abilities. ...
Full-text available
The purpose of hyperspectral unmixing (HU) is to obtain the spectral features of materials (endmembers) and their proportion (abundance) in a hyperspectral image (HSI). Due to the existence of spectral variabilities (SVs), it is difficult to obtain accurate spectral features. At the same time, the performance of unmixing is not only affected by SVs but also depends on the effective spectral and spatial information. To solve these problems, this study proposed an efficient attention-based convolutional neural network (EACNN) and an efficient convolution block attention module (ECBAM). The EACNN is a two-stream network, which is learned from nearly pure endmembers through an additional network, and the aggregated spectral and spatial information can be obtained effectively with the help of the ECBAM, which can reduce the influence of SVs and improve the performance. The unmixing network helps the whole network to pay attention to meaningful feature information by using efficient channel attention (ECA) and guides the unmixing process by sharing parameters. Experimental results on three HSI datasets showed that the method proposed in this study outperformed other unmixing methods.
... Since there are numerous techniques of Machine Learning, therefore, there are numerous techniques to train a model and solve problems. Logistic regression [6], naive Bayes classifier [7], Decision trees [8] are very normally used gadget mastering based strategies that might be used to decide whether a few is proper or wrong, right or bad. They have additionally been used in the beyond to decide numerous diseases, like cancer [9]. ...
Today, the proportion of bits of knowledge making is incredibly tremendous. Dependent upon the adjustments of estimations, immense information involves social Data, machine data, and trade-based Data. Social estimations gathered from Facebook, Twitter, etc. Machine information is RFID chip examining, GPRS, etc. Trade based bits of knowledge consolidate retail site's information. Around the assortments of different sorts of estimations first segment is printed content real factors. Content information is sorted out information. Deriving of high five star sorted out records from the unstructured printed content is artistic substance examination. Changing over unstructured real factors into critical records is a book assessment process.CV parsing is one of the substance examination strategies. It is keep parsing or extraction of CV.CV parser combines the candidate's resume with selection gems flow and thusly systems moving toward CV's. This paper proposes a CV parser adjustment of the usage of artistic substance examination. The proposed CV parser interpretation isolates substances required in the enlistment methodology inside the associations
... Essentially, the difference between these architectures is the type of neurons that form them, and the manner in which information flows through them [54]. The NN consists of multiple neurons or perceptrons in each layer, which enables the NN to capture the correlation distribution between inputs [55]. MLP operates very similar to the NN. ...
Full-text available
This paper presents our research approach and findings towards maximizing the accuracy of our classifier of feature claims for cybersecurity literature analytics, and introduces the resulting model ClaimsBERT. Its architecture, after extensive evaluations of different approaches, introduces a feature map concatenated with a Bidirectional Encoder Representation from Transformers (BERT) model. We discuss deployment of this new concept and the research insights that resulted in the selection of Convolution Neural Networks for its feature mapping aspects. We also present our results showing ClaimsBERT to outperform all other evaluated approaches. This new claims classifier represents an essential processing stage within our vetting framework aiming to improve the cybersecurity of industrial control systems (ICS). Furthermore, in order to maximize the accuracy of our new ClaimsBERT classifier, we propose an approach for optimal architecture selection and determination of optimized hyperparameters, in particular the best learning rate, number of convolutions, filter sizes, activation function, the number of dense layers, as well as the number of neurons and the drop-out rate for each layer. Fine-tuning these hyperparameters within our model led to an increase in classification accuracy from 76% obtained with BertForSequenceClassification’s original model to a 97% accuracy obtained with ClaimsBERT.
... This is required when the order of the inputs can convey valuable information (i.e., data instances of the input (x) are not independent). This situation comes up in major fields, including but not limited to natural language processing (NLP) [29], time series forecasting [30], sound processing [31], and electromyography processing [32]. The RNN and its peers (i.e., LSTM and GRU) are distinguished with their ability to hold valuable information from previous time steps, which, in RNN terminology, is called a hidden state. ...
More grid-connected solar thermal and photovoltaic power plants are coming online every year, which necessitates precise irradiance forecasters for plant management. A major share of these plants is installed at remote sites with no long-term meteorological records to develop precise site-specific global horizontal irradiance (GHI) forecasters, and the process of prior collection of those records is both expensive and time-consuming. This study proposes a unique transfer learning approach for training one-hour-ahead forecasters of hourly average GHI at new locations with limited data records by refining trained recurrent neural network-based models at other locations with abundant data. The methodology is demonstrated by considering Cairo (Egypt) as the source site and the trained models are tuned using limited datasets from five other locations in Tunisia, Morocco, and Jordan. The approach was found valid for all targeted locations, with mean absolute errors (MAEs) and root mean square errors (RMSEs) lower than 46.4 and 73.4 W/m2, respectively, and coefficients of determination (R2) higher than 93.7. The proposed models also outperformed two baseline persistence models by reducing MAEs and RMSEs by at least 34 and 44%, respectively, while holding more uniformly distributed and less clearness index-dependent residuals. The size of the source site’s dataset and the type of input predictors were, respectively, found as the least and most influential parameters on the models’ performances. This approach makes GHI forecasting accessible to practitioners and enhances the control of power plants since their start date.
Full-text available
This article provides an overview of the evolution of metamaterials (MTM) and all the aspects related to metamaterial development for antenna applications. It will be a useful collection of information for antenna researchers working in metamaterials applications. It gives an insight into the various metamaterial structures utilized along with miniature antenna designs. Different types of design parameters studied by the previous researchers are showcased to understand better perception of the metamaterial usage.
Recruitment, a worldwide industry with millions of resumes being uploaded on N-number of hiring websites for countless jobs on daily routine, with the great evolution in online-based hiring process. Job requirement is considered one of the major activities for humans which is a very strenuous job to find a fruitful talent. Every job seeker will have their novel to organize the data blocks with unique style and format for representation, thereby the resumes became an extraordinary illustration of data which is unstructured. Choosing a fantastic resume from a collection of unstructured resumes is very simple with a resume parser, which adapts the process of extracting structured data from a collection of unstructured resumes. Machine understandable output is obtained using the set of instructions that investigates and abstracts resume/CV data. This helps to store and analyse data automatically. This paper proposes a smart recruitment system (SRS) which includes resume classification using deep learning along with natural language processing (NLP) techniques and automatic questionaries’ which measures technical proficiency of an applicant, and by using syntactical similarity and semantic similarity measurements to identifying the suitable candidate according to the requirement of the company skill set. To abstract the significant data for hiring process using natural language process, we propose to use named entity recognition. The proposed procedure irrespective of format whether it is a pdf or word document of the resume, will pick up the data associated with the candidate-like work experience and education. Parsing and ranking the resume makes the hiring process easy and efficient.KeywordsParsingNLPNERSPACYSyntacticalSemantic
As the global population ages and longevity increases, health systems around the world are facing significant difficulties in building the workforce required to provide healthcare services due to increasing service demands and increasing innovation costs. The healthcare industry produces large volumes of data, and challenges in cost and patient outcomes are increasing. In order to provide health services in the fastest, most accurate, and highest quality way while also meeting the needs of both patients and healthcare organizations, healthcare professionals must reach the most accurate and up-to-date information and utilize this information by using computerized support systems. For this reason, it is inevitable for healthcare systems to become a structure supported by artificial intelligence that both speeds up and facilitates the way of doing business and provides some basic information and services to patients without being too dependent on the system. Artificial intelligence (AI) aims to mimic human intelligence to perform tasks and can recursively improve itself based on the information it gathers. By increasing the availability of healthcare data and enabling rapid progress in analytical techniques, it brings paradigm changes to healthcare services. Artificial intelligence has been used or tested for a variety of health and research purposes, including management of chronic conditions, workload reduction of doctors and nurses, drug discovery, provision and prevention of health care, diagnosis, treatment of diseases, and patient monitoring. Artificial intelligence has the ability to transform medicine through the role of physicians and nurses and to transform medicine through new science research and delivery models that revolutionize medical practices and enhance person and public health outcomes. Today, artificial intelligence practices are carried out in many private and public health institutions in many countries. In this section, we will discuss in detail the application of artificial intelligence in the modern healthcare system with advantages and possible disadvantages and challenges.KeywordsArtificial intelligenceHealthcare systemHealthcare industryMachine learningOptical character recognition (OCR)Intelligent virtual agent (IVA)Clinical decision support system (CDSS)Deep learningElectronic health records (EHRs)Robotic surgery
Full-text available
The utilization of textual documents is spontaneously increasing over the internet, email, web pages, reports, journals, articles and they stored in the electronic database format. It is challenging to find and access these documents without proper classification mechanisms. To overcome such difficulties we proposed a semantic document clustering model and develop this model. The document pre-processing steps, semantic information from WordNet help us to be bioavailable the semantic relation from raw text. By reminding the limitation of traditional clustering algorithms on the natural language, we consider semantic clustering by COBWEB conceptual clustering. Clustering quality and high accuracy were one of the most important aims of our research, and we chose F-Measure evaluation for ensuring the purity of clustering. However, there still exist many challenges, like the word, high spatial property, extracting core linguistics from texts, and assignment adequate description for the generated clusters. By the help of Word Net database, we eliminate those issues. In this research paper, there have a proposed framework and describe our development evaluation with evaluation.
Full-text available
A preliminary study in zero anaphora coreference resolution for PolishZero anaphora is an element of the coreference resolution task that has not yet been directly addressed in Polish and, in most studies, it has been left as the most challenging aspect for further investigation. This article presents an initial study of this problem. The preparation of a machine learning approach, alongside engineering features based on linguistic study of the KPWr corpus, is discussed. This study utilizes existing tools for Polish coreference resolution as sources of partial coreferential clusters containing pronoun, noun and named entity mentions. They are also used as baseline zero coreference resolution systems for comparison with our system. The evaluation process is focused not only on clustering correctness, without taking into account types of mentions, using standard CoNLL-2012 measures, but also on the informativeness of the resulting relations. According to the annotation approach used for coreference to the KPWr corpus, only named entities are treated as mentions that are informative enough to constitute a link to real world objects. Consequently, we provide an evaluation of informativeness based on found links between zero anaphoras and named entities. For the same reason, we restrict coreference resolution in this study to mention clusters built around named entities. Wstępne studium rozwiązywania problemu koreferencji anafory zerowej w języku polskimKoreferencja zerowa, w języku polskim, jest jednym z zagadnień rozpoznawania koreferencji. Dotychczas nie była ona bezpośrednim przedmiotem badań, gdyż ze względu na jej złożoność była pomijana i odsuwana na dalsze etapy badań. Artykuł prezentuje wstępne studium problemu, jakim jest rozpoznawanie koreferencji zerowej. Przedstawiamy podejście wykorzystujące techniki uczenia maszynowego oraz proces tworzenia cech w oparciu o analizę lingwistyczną korpusu KPWr. W przedstawionej pracy wykorzystujemy istniejące narzędzia do rozpoznawania koreferencji dla pozostałych rodzajów wzmianek (tj. nazwy własne, frazy rzeczownikowe oraz zaimki) jako źródło częściowych zbiorów wzmianek odnoszących się do tego samego obiektu, a także jako punkt odniesienia dla uzyskanych przez nas wyników. Ocena skupia się nie tylko na poprawności uzyskanych zbiorów wzmianek, bez względu na ich typ, co odzwierciedlają wyniki podane dla standardowych metryk CoNLL-2012, ale także na wartości informacji, która zostaje uzyskana w wyniku rozpoznania koreferencji. W nawiązaniu do założeń anotacji korpusu KPWr, jedynie nazwy własne traktowane są jako wzmianki, które zawierają w sobie wystarczająco szczegółową informację, aby można było powiązać je z obiektami rzeczywistymi. W konsekwencji dostarczamy także ocenę opartą na wartości informacji dla podmiotów domyślnych połączonych relacją koreferencji z nazwami własnymi. Z tą samą motywacją rozpatrujemy jedynie zbiory wzmianek koreferencyjnych zbudowane wokół nazw własnych.
Full-text available
Now the age of information technology, the textual document is spontaneously increasing over online or offline. In those articles contain Product information to a company profile. A lot of sources generate valuable information into text in the medical report, economic analysis, scientific journals, news, blog etc. Maintain and access those documents are very difficult without proper classification. Those problems can be overcome by proper document classification. Only a few documents are classified. All need classification and those are unsupervised. In this context clustering is the only solution. Traditional clustering technique and textual clustering have some difference. Relations between words are very imported to do clustering. Semantic clustering is proven as more appropriate clustering technique for texts. In this review paper, there has valuable information about clustering to semantic document clustering technique. In this paper, there has some information provided about advantage and disadvantage for various clustering methods.
Full-text available
The aim of this study is to build a tool for Part of Speech (POS) tagging and Name Entity Recognition for Arabic Language, the approach used to build this tool is a rule base technique. The POS Tagger contains two phases:The first phase is to pass word into a lexicon phase, the second level is the morphological phase, and the tagset are (Noun, Verb and Determine). The Named-Entity detector will apply rules on the text and give the correct Labels for each word, the labels are Person(PERS), Location (LOC) and Organization (ORG).
Neural net advances improve computers' language ability in many fields.
In recent years, deep neural networks (including recurrent ones) have won numerous contests in pattern recognition and machine learning. This historical survey compactly summarises relevant work, much of it from the previous millennium. Shallow and deep learners are distinguished by the depth of their credit assignment paths, which are chains of possibly learnable, causal links between actions and effects. I review deep supervised learning (also recapitulating the history of backpropagation), unsupervised learning, reinforcement learning & evolutionary computation, and indirect search for short programs encoding deep and large networks.
Conference Paper
Data is mostly stored in digital format rather than hard copy because the former is safer, more secure, smaller in size, and faster to retrieve than the latter. With the increasing number of electronic documents to be organized for users to obtain knowledge and integrate information, document clustering has been applied by grouping textual documents based on their similarities. Many attempts have been made to perform textual document clustering with highly accurate results (i.e., close to nature classes) and high processing performance. However, such proposed techniques work in batch (or static) mode in which performance tend to be sacrificed with the use of all the terms in the document, at times resulting in overlapping or scalability issues. Few studies that focus on dynamic clustering also reported on performance issues. This paper contributes in the investigation of textual document clustering approaches and highlights the importance of using dynamic clustering in mining frequent terms with included named entity. This method is used to achieve high efficiency and high-quality data clustering. The method is also beneficial to be used in textual document clustering algorithms for many text domain applications.