Milos Jovanovik

Milos Jovanovik
Ss. Cyril and Methodius University in Skopje · Faculty of Computer Science and Engineering

PhD
Currently focused on Knowledge Graph research.

About

66
Publications
29,306
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
238
Citations
Citations since 2017
31 Research Items
159 Citations
2017201820192020202120222023051015202530
2017201820192020202120222023051015202530
2017201820192020202120222023051015202530
2017201820192020202120222023051015202530
Introduction
PhD in Computer Science and Engineering, working in the fields of Knowledge Graphs, Linked Data and Data Science.
Additional affiliations
May 2017 - January 2022
Ss. Cyril and Methodius University in Skopje
Position
  • Assistant Professor
February 2013 - January 2022
OpenLink Software
Position
  • Knowledge Graph Engineer
September 2011 - May 2017
Ss. Cyril and Methodius University in Skopje
Position
  • Teaching and Research Assistant
Education
March 2012 - November 2016
Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University in Skopje
Field of study
  • Computer Science and Engineering
October 2008 - March 2010
Faculty of Electrical Engineering and Information Technologies, Ss. Cyril and Methodius University in Skopje
Field of study
  • Computer Networks and E-Technologies
October 2004 - June 2008
Faculty of Electrical Engineering and Information Technologies, Ss. Cyril and Methodius University in Skopje
Field of study
  • Computer Science and Engineering

Publications

Publications (66)
Article
Full-text available
Food - drug interactions are well studied, however much less is known about cuisine - drug interactions. Non-native cuisines are becoming increasingly more popular as they are available in (almost) all regions in the world. Here we address the problem of how known negative food - drug interactions are spread in different cuisines. We show that diff...
Article
Full-text available
Background: Drug product data is available on the Web in a distributed fashion. The reasons lie within the regulatory domains, which exist on a national level. As a consequence, the drug data available on the Web are independently curated by national institutions from each country, leaving the data in varying languages, with a varying structure, gr...
Thesis
Full-text available
The vast amount of data available over the distributed infrastructure of the Web has initiated the development of techniques for their representation, storage and usage. One of these techniques is the Linked Data paradigm, which aims to provide unified practices for publishing and contextually interlinking data on the Web, by using the World Wide W...
Article
Full-text available
GeoSPARQL is an important standard for the geospatial linked data community, given that it defines a vocabulary for representing geospatial data in RDF, defines an extension to SPARQL for processing geospatial data, and provides support for both qualitative and quantitative spatial reasoning. However, what the community is missing is a comprehensiv...
Article
Natural language processing (NLP) is an area of artificial intelligence that applies information technologies to process the human language, understand it to a certain degree, and use it in various applications. This area has rapidly developed in the last few years and now employs modern variants of deep neural networks to extract relevant patterns...
Article
Full-text available
In this work, we present a state-of-the-art solution for automatic playlist continuation through a knowledge graph-based recommender system. By integrating representational learning with graph neural networks and fusing multiple data streams, the system effectively models user behavior, leading to accurate and personalized recommendations. We provi...
Preprint
Full-text available
Lexicon-based sentiment analysis (SA) in finance leverages specialized, manually annotated lexicons created by human experts to extract sentiment from financial texts. Although lexicon-based methods are simple to implement and fast to operate on textual data, they require considerable manual annotation efforts to create, maintain, and update the le...
Preprint
Full-text available
Natural language processing (NLP) is an area of artificial intelligence that applies information technologies to process the human language, understand it to a certain degree, and use it in various applications. This area has rapidly developed in the last few years and now employs modern variants of deep neural networks to extract relevant patterns...
Article
Full-text available
Even though named entity recognition (NER) has seen tremendous development in recent years, some domain-specific use-cases still require tagging of unique entities, which is not well handled by pre-trained models. Solutions based on enhancing pre-trained models or creating new ones are efficient, but creating reliable labeled training for them to l...
Conference Paper
In today’s knowledge society, large number of information systems use many different individual schemes to represent data. Ontologies are a promising approach for formal knowledge representation and their number is growing rapidly. The semantic linking of these ontologies is a necessary prerequisite for establishing interoperability between the lar...
Article
The challenge of recognizing named entities in a given text has been a very dynamic field in recent years. This task is generally focused on tagging common entities, such as Person, Organization, Date, etc. However, many domain-specific use-cases exist which require tagging custom entities that are not part of the pre-trained models. This can be so...
Conference Paper
Full-text available
We set out to determine the feasibility of implementing Discrete Global Grid System (DGGS) representations of geometry support in a GeoSPARQL-enabled triplestore, and test the GeoSPARQL compliance for it. The implementation is a variant of Apache Jena's existing GeoSPARQL support. Compliance is tested using an adapted implementation of the GeoSPARQ...
Chapter
Drug repurposing, which is concerned with the study of the effectiveness of existing drugs on new diseases, has been growing in importance in the last few years. One of the core methodologies for drug repurposing is text-mining, where novel biological entity relationships are extracted from existing biomedical literature and publications, whose num...
Chapter
Full-text available
As more private data is entering the web, defining authorization about its access is crucial for privacy protection. This paper proposes a policy language that leverages SPARQL expressiveness and popularity for flexible access control management and enforces the protection using temporal graphs. The temporal graphs are created during the authentica...
Conference Paper
Full-text available
This paper presents the architecture for the development of web applications for exploring semantic knowledge graphs through parameterized interactive visualizations. The web interface and the interactive parameterized visualizations, in the form of a computational book, provide a way in which knowledge graphs can be explored. An important part of...
Article
Full-text available
Checking the compliance of geospatial triplestores with the GeoSPARQL standard represents a crucial step for many users when selecting the appropriate storage solution. This publication presents the software which comprises the GeoSPARQL compliance benchmark – a benchmark which checks RDF triplestores for compliance with the requirements of the Geo...
Preprint
Full-text available
The challenge of recognizing named entities in a given text has been a very dynamic field in recent years. This is due to the advances in neural network architectures, increase of computing power and the availability of diverse labeled datasets, which deliver pre-trained, highly accurate models. These tasks are generally focused on tagging common e...
Preprint
Full-text available
We propose a series of tests that check for the compliance of RDF triplestores with the GeoSPARQL standard. The purpose of the benchmark is to test how many of the requirements outlined in the standard a tested system supports and to push triplestores forward in achieving a full GeoSPARQL compliance. This topic is of concern because the support of...
Conference Paper
Full-text available
With the constant increase in population and the growing impact of climate change, energy efficiency on a household and a city-wide level represents a significant key in the process of transformation of smart cities. Recently, machine learning approaches have been proven to be beneficial in addressing several global problems, especially in areas wh...
Chapter
Analyzing long text articles in the pharmaceutical domain, for the purpose of knowledge extraction and recognizing entities of interest, is a tedious task. In our previous research efforts, we were able to develop a platform which successfully extracts entities and facts from pharmaceutical texts and populates a knowledge graph with the extracted k...
Conference Paper
Full-text available
Geospatial RDF datasets have a tendency to use latitude and longitude properties to denote the geographic location of the entities described within them. On the other hand, geographic information systems prefer the use of WKT and GML geometries when working with geospatial data. In this paper, we present a process of RDF data transformation which p...
Conference Paper
Full-text available
Medical datasets that contain data relating to drugs and chemical substances, in general tend to contain multiple variations of a generic name which denotes the same drug or a drug product. This ambiguity lies in the fact that a single drug, referenced by a unique code, has an active substance which can be known under different chemical names in di...
Conference Paper
Full-text available
Autonomous vehicles (AVs) are a disruptive technology of the 21-st century that are beginning the next revolution of the transportation of people and goods. Their presence has a particular impact on the future directions of development of E-commerce. The number of online orders is in a steep incline, and so is the necessity to deliver goods to the...
Conference Paper
Full-text available
Security is an inevitable part of every information system. It is a cross-cutting concern that affects every part of the system. There is a constant trade-off between a secured system and convenient security management, and this management gets more demanding when the permissions are context dependent. The delegation of authorization is one way to...
Chapter
The aim of the Mighty Storage Challenge (MOCHA) at ESWC 2018 was to test the performance of solutions for SPARQL processing in aspects that are relevant for modern applications. These include ingesting data, answering queries on large datasets and serving as backend for applications driven by Linked Data. The challenge tested the systems against da...
Chapter
Following the success of Virtuoso at last year’s Mighty Storage Challenge - MOCHA 2017, we decided to participate once again and test the latest Virtuoso version against the new tasks which comprise the MOCHA 2018 challenge. The aim of the challenge is to test the performance of solutions for SPARQL processing in aspects relevant for modern applica...
Conference Paper
Full-text available
Following the success of Virtuoso at last year's Mighty Storage Challenge - MOCHA 2017, we decided to participate once again and test the latest Virtuoso version against the new tasks which comprise the MOCHA 2018 challenge. The aim of the challenge is to test the performance of solutions for SPARQL processing in aspects relevant for modern applica...
Conference Paper
Full-text available
AntiVirus products and tools are essential in every business deployment connected to the Internet. Nowadays, with the increase in the number and diversity of malware on the Web, there are also more AntiVirus Tools (AVT) becoming available to protect users and/or companies from malware. However, the quarterly growth at around 12% for known unique ma...
Conference Paper
Full-text available
Transportation is one of the main concerns of the big cities because vehicles are one of the major sources of air pollution. Despite this fact, many cities follow a trend of growth of the number of vehicles. In order to meet the demand, the authorities are obliged to increase the roads capacity or build new road infrastructure. However, this soluti...
Chapter
The Mighty Storage Challenge (MOCHA) aims to test the performance of solutions for SPARQL processing, in several aspects relevant for modern Linked Data applications. Virtuoso, by OpenLink Software, is a modern enterprise-grade solution for data access, integration, and relational database management, which provides a scalable RDF Quad Store. In th...
Chapter
The aim of the Mighty Storage Challenge (MOCHA) at ESWC 2017 was to test the performance of solutions for SPARQL processing in aspects that are relevant for modern applications. These include ingesting data, answering queries on large datasets and serving as backend for applications driven by Linked Data. The challenge tested the systems against da...
Chapter
A large number of emerging services expose their data using various Application Programming Interfaces (APIs). Consuming and fusing data form various providers is a challenging task, since separate client implementation is usually required for each API. The Semantic Web provides a set of standards and mechanisms for unifying data representation on...
Conference Paper
Full-text available
The Mighty Storage Challenge (MOCHA) aims to test the performance of solutions for SPARQL processing, in several aspects relevant for modern Linked Data applications. Virtuoso, by OpenLink Software, is a modern enterprise-grade solution for data access, integration, and relational database management, which provides a scalable RDF Quad Store. In th...
Conference Paper
Full-text available
Synthetic datasets used in benchmarking need to mimic all characteristics of real-world datasets, in order to provide realistic benchmarking results. Synthetic RDF datasets usually show a significant discrepancy in the level of structuredness compared to real-world RDF datasets. This structural difference is important as it directly affects storage...
Chapter
Full-text available
A flexible platform supporting the linked data life-cycle has been developed and applied in various use cases in the context of the large scale linked open data project Fusepool P3. Besides the description of the aims and achievements, experiences from publishing and reusing linked data in public sector and business are summarized. It is highlighte...
Conference Paper
Full-text available
The Linked Data approach in data publishing allows the users and their data-driven applications to have broader use cases which encompass various data sources, either publicly available on the Web, or in private repositories. The use of W3C standards in publishing such data enables uniform access across platforms. Transport information today has hi...
Conference Paper
Full-text available
This project is intended to ease the writing process of dynamic SPARQL queries for applications. Its goal is to make an autocomplete form that can be reused in different applications and will be up to date with the latest ontologies, thus making the process of using Linked Open Data closer to application developers in general. This is done by havin...
Conference Paper
Full-text available
Publishing raw data as Linked Open Data gives an opportunity of data reusability and data understandability for the computer machines. Today, the air pollution problem is one of the biggest in the whole world. Republic of Macedonia, especially its capital Skopje, has big problems with the PM2.5 and PM10 particles in the air approved by several meas...
Data
Full-text available
One of the most active fields of research in the past decade has been data representation, storage and retrieval. With the vast amount of data available on the Web, this field has initiated the development of data management techniques for distributed datasets over the existing infrastructure of the Web. The Linked Data paradigm is one of them, and...
Chapter
Full-text available
The concept of Open Data, which represents the idea that public data should be published in a machine-readable format, starts to take a significant role in modern society. Public data from various fields are being transformed in open data formats and published on systems which allow easier consumption from software agents and applications, as well...
Chapter
Full-text available
One of the most active fields of research in the past decade has been data representation, storage and retrieval. With the vast amount of data available on the Web, this field has initiated the development of data management techniques for distributed datasets over the existing infrastructure of the Web. The Linked Data paradigm is one of them, and...
Chapter
Full-text available
In this paper, a document collaboration platform for enterprise environments is presented. It incorporates the collaboration, security, auditing and reuse features into the document editor, as a tool that has well known interface to the end users. The platform enables template definition with annotation of the collaboration and protection units, ca...
Conference Paper
Full-text available
The Linked Data best practices provide ways for easier data representation, while at the same time raise the quality of the information that comes with it. The idea behind these best practices is to interlink datasets from various sources which are distributed over different locations and publish the data in an open, machine-readable format so that...
Conference Paper
Full-text available
Accessing data on the Web in order to obtain useful information has been a challenge in the past decade. The technologies of the Semantic Web have enabled the creation of the Linked Data Cloud, as a concrete materialization of the idea to transform the Web from a web of documents into a web of data. The Linked Data concept has introduced new ways o...
Conference Paper
Full-text available
The Linked Data concept uses a collection of Semantic Web technologies in order to interconnect, publish, and share pieces of data on the Web, in a machine-readable format. It enables querying and combining data from different datasets together in order to retrieve specific information and enable use-case scenarios which are unavailable over isolat...
Conference Paper
Full-text available
The need to represent data on the Web in a way that will make it easier to manage, has led to new solutions for data representation, visualization, storage and querying. The concepts of Open Data, Linked Data and the Semantic Web offer a significant improvement in information and data dissemination. These concepts aim towards making data on the Web...
Chapter
Full-text available
With the growing popularity of the service oriented architecture concept, many enterprises have large amounts of granular web services which they use as part of their internal business processes. However, these services can also be used for ad-hoc actions, which are not predefined and can be more complex and composite. Here, the classic approach of...
Conference Paper
Full-text available
With the development of IT technologies, the amount of information that the user receives and the sources of information which he or she uses, increased dramatically. There is a need of an intelligent desktop application, which will integrate different information sources from both the desktop and the web. To address this problem, we present a solu...
Conference Paper
Full-text available
The concept of Open Data represents the idea that certain data should be freely available to the public, for use, reuse, republishing and redistributing, with little or no restrictions. The goal is to make data of public nature available in an open manner, in a raw and machine-readable format, so that they can be used for building useful applicatio...
Conference Paper
Full-text available
Information dissemination has always been in the focus of the computer science research community. New ways of information and data representation, storage, querying and visualization are being constantly developed and upgraded. Linked Open Data represents a concept which offers a comprehensive solution for information and data dissemination. It ac...
Conference Paper
Full-text available
Recent activities of governments around the world regarding the publication of open government data on the Web, re-introduced the Open Data concept. The concept of Open Data represents the idea that certain data should be freely available to the public, i.e. the citizens, for use, reuse, republishing and redistributing, with little or no restrictio...
Conference Paper
Full-text available
These days, the number of data sources an ordinary computer user works with every day is very large and continues to grow. With the increasing number of cloud services with specialized functionalities, the users are faced with the necessity to routinely perform manual actions to interchange data among different cloud and web services, in order to p...
Conference Paper
Full-text available
Today, many organizations and institutions have vast collections of datasets and databases filled with information that, in general, can turn out to be very useful for individuals and for the society [1]. The Police is one such institution which is obliged by law to keep an archive of all the information it deals with on a daily basis. All that inf...
Conference Paper
Full-text available
The existence of the Semantic Web is very dependent on the semantic content which exists on the current web. Thus, the process of creating more semantic content actually means building the Semantic Web itself. This process can be done using various information extraction and annotation mechanisms. However, they usually focus on domain specific appr...
Conference Paper
Full-text available
Today, there is a growing trend for publishing public data in an open format, on the web, making it available for everyone to use and reuse. This idea has been widely supported by governments and companies throughout the world, which have made their own public data available in such way. Some of them, like the World Bank, even challenge developers...
Conference Paper
Full-text available
Following the global trends and innovations in the communication and information technology, the way of communication between universities and faculties with the students, and among students themselves, became significantly different than a few years ago. The use of social networks and other forms of interactive communication for this type of commu...
Conference Paper
Full-text available
In the age of information overflow, we face with the challenge of categorizing the digital information we come across on a daily basis, in order to apply different operations and priorities to different types of information and to manage to use it in a more efficient manner. This issue introduces the challenge of automatic text classification. The...
Chapter
Full-text available
A few years ago, the System on Chip idea grew largely and ‘flooded’ the market of embedded systems. Many System on Chip designers started to write their own HDL components and made them available on the Internet. The idea of searching for a couple of pre-written cores and building your own System on Chip only by connecting them seemed time saving....
Conference Paper
Full-text available
The progress in the field of web services has made it possible to publish, locate, and invoke specific applications across the Web. This is the reason why more and more companies and organizations now implement their core business, or outsource other application services, over the Internet. Therefore, the efficient selection and integration of inte...
Conference Paper
Full-text available
We agree that a semi -structured data model offers the right balance of rich structure and flexible schema allowing users to store and manage data as they prefer, making the access as easy as possible. For that reason we build this system, which exposes the richness and flexibility of the data model, offering natural access to the data and hides th...
Conference Paper
Full-text available
The fast technology development causes exponential increase of the number of information available to the people. But people do not to extract useful conclusions from all the information served, simply because they are either overloaded or uninterested about certain topics. By their nature, people are slow, unreliable, they forget and make errors w...
Conference Paper
Full-text available
The world labor market, besides the technical (hard) skills is increasing the focus on the so-called “soft” skills. Soft skills are a set of skills which allow an individual to fit in and successfully function in its environment. Because the engineering market is so dynamic and full of contacts with different cultures, soft skills need to be accomp...
Conference Paper
Full-text available
The World Wide Web is a web of documents. These documents and the data they hold are connected in a “mechanical” manner, with hyperlink references. The W3C with a number of it’s research and industry partners are working on an effort – to extend the principles of the Web from documents to data. This endeavor is called The Semantic Web, and it’s vis...