Conference PaperPDF Available

Abstract and Figures

Extraction and analysis of environmental information is very important, since it strongly affects everyday life. Nowadays there are already many free services providing environmental information in several formats including multimedia (e.g. map images). Although such presentation formats might be very informative for humans, they complicate the automatic extraction and processing of the underlying data. A characteristic example is the air quality and pollen forecasts, which are usually encoded in image maps, while the initial (numerical) pollutant concentrations remain unavailable. This work proposes a framework for the semi-automatic extraction of such information based on a template configuration tool, on Optical Character Recognition (OCR) techniques and on methodologies for data reconstruction from images. The system is tested with a different air quality and pollen forecast heatmaps demonstrating promising results.
Content may be subject to copyright.
Environmental Data Extraction from Multimedia Resources
Anastasia Moumtzidou 1, Victor Epitropou2, Stefanos Vrochidis1, Sascha Voth3,
Anastasios Bassoukos2, Kostas Karatzas2, Jürgen Moßgraber3, Ioannis Kompatsiaris1,
Ari Karppinen4 and Jaakko Kukkonen4
1Information Technologies Institute
6th Klm Charilaou-Thermi Road
Thessaloniki, Greece
{moumtzid, stefanos, ikom}
3Fraunhofer Institute of Optronics,
System Technologies and Image Exploitation
Karlsruhe, Germany
Informatics Systems and Applications Group
Aristotle University of Thessaloniki
Thessaloniki, Greece,,,
4Finnish Meteorological Institute
Helsinki, Finland
{ari.karppinen, jaakko.kukkonen}
Extraction and analysis of environmental information is very
important, since it strongly affects everyday life. Nowadays there
are already many free services providing environmental
information in several formats including multimedia (e.g. map
images). Although such presentation formats might be very
informative for humans, they complicate the automatic extraction
and processing of the underlying data. A characteristic example is
the air quality and pollen forecasts, which are usually encoded in
image maps, while the initial (numerical) pollutant concentrations
remain unavailable. This work proposes a framework for the
semi-automatic extraction of such information based on a
template configuration tool, on Optical Character Recognition
(OCR) techniques and on methodologies for data reconstruction
from images. The system is tested with a different air quality and
pollen forecast heatmaps demonstrating promising results.
Categories and Subject Descriptors
H.3.1 [Information Storage and Retrieval]: Content Analysis
and Indexing – Indexing methods.
Environmental, multimedia, images, heatmaps, OCR, data
reconstruction, template, configuration, pollen, air quality.
Environmental conditions are of particular interest for people,
since they affect everyday life. Thus, meteorological conditions,
air quality and pollen (i.e. weather, chemical weather and
biological weather) are strongly related to health issues (e.g.
allergies, asthma, cardiovascular diseases) and of course they play
an important role in everyday outdoor activities such as sports and
commuting. With a view to offering personalized decision support
services for people based on environmental information regarding
their everyday activities [1], there is a need to extract and combine
complementary and competing environmental information from
several resources. One of the main steps towards this goal is the
environmental information extraction from multimedia resources.
Environmental observations are automatically performed by
specialized instruments, “hosted” in stations established by
environmental organizations, and the data collected are usually
made available to the public through web portals. In addition to
the observations, forecasts are used to foretell the levels of
pollution in areas of interest, and these are usually published on-
line in the form of images, while only a few of the data providers
make available some means of access to their actual (numerical)
forecast data. It should be mentioned that the presentation format
adopted is human oriented and in most of the cases doesn’t allow
for an automatic (or at least semi-automatic) extraction of
information. A characteristic example is the air quality and pollen
forecasts, which are usually encoded in image maps (heatmaps) of
heterogeneous formats, while the initial (numerical) pollutant
concentrations remain unavailable. In this context, we propose a
semi-automatic framework for extracting environmental data from
air quality and pollen concentrations, which are presented as
heatmap images. The framework consists of three main
components: an annotation tool for user intervention, an Optical
Character Recognition (OCR) and text processing module, as well
as the AirMerge heatmaps image processing system [2], [3].
The contribution of this paper is the integration of existing tools
for environmental quality forecast data extraction (i.e. AirMerge)
with text processing and OCR techniques tailored for heatmap
analysis, under a configurable semi-automatic framework for
processing air quality and pollen forecast heatmaps, which offers
a graphical user interface for template-based customization.
This paper is structured as follows: section 2 presents the relevant
work and section 3 describes the problem. Section 4 introduces
the proposed framework, the annotation tool, the text processing
component and the image processing module. The evaluation is
presented in section 5 and finally, section 6 concludes the paper.
The task of map analysis strongly depends on the map type and
the information we need to extract. Depending on the application,
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. To copy
otherwise, or republish, to post on servers or to redistribute to lists,
requires prior specific permission and/or a fee.
MAED’12, November 2, 2012, Nara, Japan.
Copyright 2012 978-1-4503-1588-3/12/11...$15.00.
a straightforward requirement would be to extract meaningful
segments (e.g. rivers, forests, etc.), while in the case of heatmaps
it is to transform color into numerical data. In general, the
information contained in maps can be distinguished by their scale,
colorization, quality, etc.. In the case of air quality and pollen
forecast maps two types of information are mainly covered:
1. Geographical information: points and lines describing
country frontiers or other well-known points of interests or
structures (e.g. sea, land) in a given coordinate system.
2. Feature information: measured or forecasted parameters of
any kind (e.g. average temperature) which is coded via a
color scale representing the (measured or forecasted) values.
Single values are referenced geographically by a color point
at the corresponding geographical point.
Chemical weather maps often use raster map images to represent
forecasted data or spatially interpolated measured data. There are
several approaches to extract and digitalize this image information
The authors in [4] describe the process of the vectorization of
digital image data. Hereby the geographical information, in form
of lines, is extracted and converted to digital storable vector data.
In [5] the authors use the specific knowledge of the known
colorization in USGS maps, to automatically segment these maps
based on their semantic contents (e.g. roads, rivers). Finally, [6]
improves the segmentation quality of text and graphics in color
map images, to enhance the results of the following analysis
processes (e.g. OCR), by selecting black or dark pixels from color
maps, cleaning them up from possible errors or known unwanted
structures (e.g. dashed lines), to get cleaner text structures.
Although research work has been conducted towards the
automatic extraction of information in maps, to the best of our
knowledge only the work performed for the AirMerge system
addresses the extraction of information from chemical weather
maps with image processing. In such works [7], [2], [3], a method
to reconstruct environmental data out of chemical weather images
is described. First, the relevant map section is scraped from the
chemical weather image. Then, disturbances are removed and a
color classification is used to classify every single data point
(pixel), to recover the measured data. With the aid of the known
geographical boundaries, given by the coordinate axis and the
map projection type, the geographical position of the measured
data point can be retrieved. In case of missing data points, a
special interpolation algorithm is used to close these gaps.
The proposed work proposes a framework, which integrates
AirMerge and extends its application to pollen forecasts and in
addition facilitates the procedure of information extraction from
heatmaps using OCR and visual annotation techniques.
In order to clearly state the problem we have conducted an
empirical study on more than 60 environmental websites (dealing
with weather, air quality and pollen) and based on previous works
[8] we reached to a conclusion that a considerable share of
environmental content, almost 60%, is encoded in images.
Specifically, pollen and air quality forecast information is always
illustrated in heatmap images. Overall, it can be said [9] that the
air quality and pollen information is usually presented in the form
of images representing pollutant or pollen concentrations over a
geographically bounded region, typically in terms of maximum or
average concentration values for the time scale of reference,
which is usually the hour or day [7]. These providers present their
air quality forecasts almost exclusively in the form of
preprocessed images with a color index scale indicating the
concentration of pollutants. In addition, they arbitrarily choose the
image resolution and the color scale employed for visualizing
pollution loadings, the covered region, as well as the geographical
map projection. The mode of presentation varies from simple web
images to AJAX, Java or Adobe Flash viewers [10].
The heatmaps that contain environmental information are static
bitmap images, which represent the coverage data (e.g.
concentrations) in terms of a color-coded scale over a
geographical map. An example of such heatmap is depicted in
Figure 1 obtained from the GEMS project1 website.
Figure 1. Typical example of image of air quality forecasts.
After observing the image, we conclude that besides the
geographical information and concentrations, additional
information is also provided, which is the type of environmental
“feature” (e.g. ozone, birch pollen), the date/ time information of
the meaningful information and a color scale. Therefore, the main
parts of information that need to be extracted from the image are:
Heatmap: map depicting a geographical region with
colors representing the environmental aspect value.
Color scale: range indicating the correspondence
between feature value and color
Coordinate axis (x, y): indicate the geographical
longitude and latitude of every map point for a specific
geographic projection
Title: contains information such as the type of aspect
measured, the time and date of forecast
Additional information: watermarks, wind fields
superimposed to concentration maps and any
information that can be categorized as “noise” in terms
of influencing the information content and
representation value of the specific heatmap.
The proposed architecture draws upon the requirements that were
set in the previous section. The idea is to employ image analysis
and processing techniques to map the color variations on the
images on specific categories that can be ranges of values. Optical
character recognition techniques need to be used for recognizing
text encoded in image format such as image titles, dates,
environmental information and coordinates.
Due to the fact that there is a large variation of images and many
different representations, there is a need for optimizing and
configuring the algorithms involved. Specifically, the intervention
of an administrative user is required in order to annotate and
manually segment different parts of a new image type (like data,
legend, etc.), which need to be processed by the content extraction
algorithms. The system workflow and the involved modules are
depicted in Figure 2.
In order to facilitate this configuration through a graphical user
interface we have implemented the “Annotation Tool” (AnT),
which is tailored for dealing with heat maps. The output of this
tool is a configuration file that holds the static information of the
image. The second module is the “Text Processing”, which uses
the information of the configuration file to extract data from the
corresponding image. More specifically, it retrieves and analyzes
the information captured in text format using text processing
techniques including OCR. The third module is the “Image
Processing”, which uses information both from the output of the
“Text processing” module and the configuration file to process the
heatmap found inside the image.
Figure 2. Image content distillation architecture
The input of the framework is a heatmap image and the output is
an XML file, in which each geographical coordinate of the initial
heatmap image is associated with a value (e.g. air quality index).
The results of the empirical study indicated that heatmap images
share common characteristics, which in most cases do not match
exactly, e.g. spatial arrangement. Therefore, it is necessary to
manually identify the interesting parts in the images in order to
automatically extract the important information. This can be
achieved by a semi-automatic approach, which involves an
administrative user, who would annotate parts of the image.
The information provided by this user includes mainly the
position and the dimension of several pre-specified elements of
the image, such as the position and size of the heatmap, the color
scale, the x and y axis, and the title, which are saved in a template
file. To do this, the user has to define region of interests (ROI)
and/or point of interests (POI) inside these images.
Within the context of this work, a tool has been developed to
make the annotation process easier and more user-friendly. The
tool called Annotation Tool (“AnT”) can load images and
predefined templates, and let the user interactively annotate them.
The tool receives as input an image and produces an XML
configuration file as output.
The basic annotation structure of the images is preloaded based on
the image type, in the form of an XML template file, which
includes all necessary elements to describe the images (e.g. for
heat maps at least the Region Of Interest (ROI) of the map, the
ROI of the coordinates, etc.). The Annotation tool (Figure 3)
provides two data views on the annotation data: a tree view and a
graphical view. The tree view can be used to traverse through all
the elements and edit the values of the elements by hand. In case
of ROIs or Points of Interest (POI) these elements are drawn in
the graphical view onto the loaded image as overlay, to verify the
parameters (e.g. size and position of an ROI). By using the mouse
the displayed ROIs and POIs can be modified. In case of ROIs the
four corner points of the drawn rectangle can be used to resize the
actual element. By clicking in the centre of the rectangle, the
object can be moved. POIs can be moved to another position. Any
change of position and size is synchronized between both views.
Figure 3. The user interface of Annotation Tool
This module is driven by the configuration file and focuses on
retrieving the textual information captured in the image using
OCR and text processing.
The first step includes the application of OCR on several parts of
the input image. We used the Abbyy Fine Reader2 OCR software,
which is applied separately on the following parts of the initial
image: title, color scale, map x and y axis. In the second step
empirical text processing techniques are applied on the results of
OCR in order to make corrections by combining different sources
of information. Part of the extracted information (e.g. coordinates)
is used as input for the Image Processing Module. Then, we
describe the two steps by applying them on the typical heatmap
image of Figure 1 and present the results.
4.2.1 OCR on Title, Color Scale, Map Axis
Based on the empirical study, considerable part of the meaningful
information can be extracted from the text surrounding the image.
More specifically, color scale and map axis are essential elements
that provide information about the values and the geographical
area covered. On the other hand, the title contains information
about the environmental aspect measured and the corresponding
date/ time. The location of the aforementioned image parts is
captured in the configuration template.
We apply OCR in the heatmap of Figure 1.The Tables 1, 2 and 3
contain the input and output of OCR for the title, the color scale
and y axis (the results for x axis are omitted, since they are
generated in a similar way to the ones of y axis). The values in
bold indicate the errors produced by OCR. It should be noted that
for the cases of color scale and x, y axis, we also retrieved the
exact position of the text, in order to relate the text position with
geographical coordinates.
Table 1. Title - Image (top) and OCR output (bottom)
Tuesday 29 November 2011 OOUTC GEMS-RAQ Forecast
t4027 VT: Wednesday 30 November 2011 03UTC
Model: EURAD-IM Height level: Surface Parameter: Ozone [
\iq m31
Table 2. Color scale – Image (left) and OCR (with position)
output (right)
Position: left, top, right, bottom – Value
Position: 0, 17,51, 39 – Value: 360
Position: 0, 128, 50, 150 – Value: 240
Position: 0, 236, 51, 258 – Value: 200
Position: 3, 345, 51, 366 – Value: 180
Position: 3, 456, 51, 477 – Value: 180
Position: 2, 563, 51, 585 – Value: 140
Position: 3, 672, 51, 693 - Value: 120
Table 3. Coordinates of y axis – Image (left) and OCR (with
position) output (right)
Position: left, top, right, bottom – Value
Position: 20, 194, 78, 213 – Value: 65°N
Position: 20, 386, 78, 405 - Value: 60°N
Position: 21, 578, 77, 596 - Value: 55°N
Position: 21, 770, 78, 789 - Value: 50°N
Position: 20, 959, 47, 978 - Value: 45
Position: 51, 960, 77, 977 - Value: °N
Position: 21, 1155, 47, 1172 - Value: 40
Position: 51, 1155, 77, 1172 - Value: °N
4.2.2 Text processing on OCR Results
Then, we apply text processing to extract, correct and understand
the semantic information encoded in the aforementioned places.
Each of these segments was treated in different way since the type
of the semantic information included is different. Title
The title (if it exists) usually contains the name of the aspect, the
measurement units and the date/ time. Moreover, in case that the
data included in the map are forecast data, they contain two dates.
Regarding the measurement units, these are usually standard
depending on the measured environmental aspect and therefore
we will not attempt to extract them. The date/time is considered as
the most complex element given that it is presented in several
different formats. In the current implementation, we focus on
formats similar to the GEMs site. In order to correct possible
errors in the textual format of the month, day and aspect we apply
the Levenshtein distance and compare with three English ground
truth sets. Then we correct the initial OCR result by considering
the word from the ground truth dataset that has the minimum
distance from it. For simplicity, in the current version we
considered having only one date/time in the title.
In the specific example, no corrections were required. Thus, the
information we obtained from the title is the following:
Date/ time: 2011-11-29 00:00:00, Aspect: Ozone Color Scale
The color scale shows the mapping between colors variations in
the map and environmental aspect values. The extraction of
information from the color scale is a two step procedure.
The first step corrects OCR results, while the second correlates
values to colors. In order to correct the OCR results, the most
common difference among the scale values is calculated and then
the error values are adapted accordingly. The correlation of values
to colors is achieved by using the top-bottom or left-right
coordinates, depending on the color scale orientation, of the color
scale values and map them to the closest color.
In the specific example, the most common interval among the
values in the scale is 20 and error values are corrected based on
that. Then, values are mapped onto coordinates and thus colors.
For example, 140-160 is mapped onto the color found in
(719,224) coordinates of the initial image. X and Y Axis
Regarding x and y axis, similar processing techniques are applied
since they both represent the geographical coordinates of the map.
Specifically, at least two points of the map, as well as their
position with respect to the map needs to be resolved, in order to
identify successfully all the point coordinates. The procedure
followed includes again two steps: a) correction of the errors
produced by OCR and b) use of the element the coordinates.
For the specific example, after correcting OCR results, we
associated the geographical coordinates (-10, 65) and (-5, 60) to
the image map pixels (98, 125) and (162,189) respectively.
In this section we present the image processing module that
extracts data from different models and coordinate systems. The
tool integrated into the system is the AirMerge engine, which
performs various tasks concerning analysis, reverse engineering
and reuse of heatmaps like Chemical Weather forecasts.
The AirMerge engine combines elements of screen scraping,
image processing and geographical coordinate transformations, in
order to produce uniform, indexed data using a unified format and
geographical projection. The engine is already in use as part of a
more complex production environment [9], [3], it is available as a
REST service via an API, and also in the proposed tool-chain as
the final processing step, as indicated in Figure 2.
4.3.1 Screen Scraping
This step handles the cropping of the original image to a region of
interest (the heatmap) and parsing of it into a 2D data array
directly mapped to the original images’ pixels. Also, it associates
each color to minimum/maximum value ranges of the air pollutant
concentration levels, which is often implied by the color scale
associated with the original images. It should be noted that the
information about where to crop, where each color on the legend
is, etc. are provided by the configuration template of the AnT in
the proposed system. In this phase, the mapping of the images’
raster to a specific geographical grid is performed, since the
images themselves represent a geographical region. The
configuration system allows choosing between the most
commonly encountered geographical projections (equirectangular,
conical, etc.) and choosing keypoints in the image to allow for
precise pixel-coordinate mapping. These functions are semi-
automated in the autonomous AirMerge system.
4.3.2 Reconstruction of Missing Values and Data
This step deals with unwanted elements such as legends, text,
geomarkings and watermarks, as well as regions that are not part
of the forecast area. The image’s pixels are classified into three
main categories: valid data (with colors that satisfy the color
scale’s classification), invalid data (with colors not present in the
color scale), and regions containing colors that are explicitly
marked for exclusion, and which are considered void for all
further presentation and processing. However, regions containing
unmarked invalid data are considered as regions with correctable
errors or “data gaps” which can be filled-in. This distinction is due
to their different appearance patterns: void regions are usually
extended and continuous (e.g. sea regions not covered by the
forecast, but present on the map), while invalid data regions are
usually smaller but more noticeable (e.g. lines, text, watermarks
“buried” in valid data regions) and with more noise-like patterns,
and thus it is more compelling to remove them by using gap-
filling techniques. These techniques include traditional grid
interpolation as well as pattern-based interpolation techniques
using neural networks, which are described in detail in [8].
The evaluation of the framework is carried out into two steps with
different focus. The first step deals with evaluating the OCR and
providing a visual assessment of the output, while the second
evaluates the final system result after running a series of tests on
several images and comparing with ground truth. We omit the
presentation of the final XML output of the system (i.e. mapping
of geographic coordinates to forecast values), since its visual
presentation is not that informative, and instead we present the
reconstructed image, which derives from this representation and is
more appropriate for visual inspection.
5.1 OCR Performance and Visual Results
The tests in this step focus on the recognition of the x and y axis
and evaluate the assignment of pixels to geographical coordinates.
Given the fact that we were not aware of the initial heatmap
values we can only assess the results of AirMerge by visual
comparison of the original image and the one produced by
AirMerge. However, a more detailed assessment and evaluation of
the AirMerge system can be found in [7], [2] and [3].
The images tested during the first step of the evaluation were
extracted from the following sites of the Finnish Meteorological
Institute (FMI): Pollen FMI site
and SILAM model FMI site
Pollen FMI website
Figures 4 and 5 depict the original image and the reconstructed
image produced from the proposed system after visualizing the
XML output. The reconstructed figure is almost identical and in
addition any noise (e.g. black lines) was removed.
In table 4 we report the error introduced by the OCR named as
“absolute error”, which is calculated as the difference when
subtracting the OCR estimation (e.g. 4.98775 in the first line)
from the initial degrees range (e.g. 5 in the first line). In both
cases the absolute error is very low (around 0.3%) and acceptable.
Figure 4: Original Image Figure 5: Reconstructed Image
Table 4. OCR error in Pollen website
degrees Estimation Absolute Error
Longitude step 5
4.98775 0.01225
Latitude step 5
4.98404 0.01596
SILAM model FMI website
Figure 6: Original Image
igure 7: Reconstructed Image
Table 5. OCR error in SILAM website
degrees Estimation Absolute Error
Longitude step 5
4.97516 0.02484
Latitude step 5
4.96523 0.03477
In case of the SILAM site, based on visual assessment the
reconstructed image (Figure 7) is almost identical to the initial one
(Figure 6). The absolute geo-coordinate error is very low (around
0.6%) and thus the error introduced by OCR is not significant.
5.2 System Evaluation
In this step we focus on evaluating the system output with
heatmaps from different providers. The evaluation is realized by
comparing the results of AirMerge system based on manual
configuration, which is considered as ground truth with the results
of the proposed system involving AnT, OCR and AirMerge.
The tests are performed on a set of 60 images, extracted from the
following sites and locations for different times/dates:
GEMS site,
Pollen FMI site,
SILAM site,
Atmospheric and Oceanic Physics Group site,
In Table 6 we report the following results for every site: a) the
number of images, b) the number of different colors in the color
scale, c) the absolute latitude and longitude errors, which indicate
the error introduced by the proposed framework for 5o degrees in
each axis, d) the average percentage of pixels with correct values
(i.e. compared with the values provided by manually configured
AirMerge) and e) the average error introduced in each pixel due to
OCR and thus misalignment of the coordinates. The error  is
calculated as:   
, where is the total number of
pixels, is the value of pixel i using AirMerge with manual
configuration and  is the value estimated by the system.
Based on Table 6, it is evident that both the latitude and longitude
errors are quite low for all sites and the percentage of pixels with
correct values is satisfactory. The error introduced in each pixel
value is in general quite low (around 6%), while only in the case
of Pollen FMI site, the error is higher (around 12%). This is
probably due to the fact that the values between sequential pixels
were highly varying compared to the other sites.
In this paper, we propose a framework for environmental
information extraction from air quality and pollen forecast
heatmaps, combining image processing, template configuration,
as well as textual recognition components. This framework could
serve as a basis for supporting environmental systems that provide
either air quality information from several providers for direct
comparison or orchestration purposes or decision support [1] on
everyday issues (e.g. travel planning). The proposed work
overcomes the limitation of not having access to the raw data,
since it only considers information being publicly available on the
Internet. Future work includes extensive evaluation with more
images in different projections (e.g. conical), recognition of
additional elements with OCR, as well as employment of Linked
Open Data to enrich the semantics of the extracted information.
This work was supported by the FP7 project PESCaDO.
[1] Wanner, L., Rospocher, M., Vrochidis, S., Bosch, H.,
Bouayad-Agha, N., Bugel, U., Casamayor, G., Ertl, T.,
Hilbring, D., Karppinen, A., Kompatsiaris, I., Koskentalo, T.,
Mille, S., Moßgraber, J., Moumtzidou, A., Myllynen, M.,
Pianta, E., Saggion, H., Serafini, L., Tarvainen, V., and
Tonelli, S. 2012. Personalized Environmental Service
Configuration and Delivery Orchestration: The PESCaDO
Demonstrator. In Proceedings of the 9th Extended Semantic
Web Conference (ESWC 2012), Heraclion, Crete, Greece.
[2] Epitropou V., Karatzas K. and Bassoukos A. 2010. A method
for the inverse reconstruction of environmental data
applicable at the Chemical Weather portal. In Geospatial
Crossroads @GI_Forum’10, In Proceedings of the
GeoInformatics Forum Salzburg, 58-68, Wichmann Verlag,
Berlin, ISBN 978-3-87907-496-9.
[3] Epitropou V. Karatzas K., Kukkonen J. and Vira J. 2012.
Evaluation of the accuracy of an inverse image-based
reconstruction method for chemical weather data,
International Journal of Artificial Intelligence, in press.
[4] Musavi, M.T., Shirvaikar, M.V., Ramanathan, E. and
Nekovei, A.R. 1988. Map processing methods: an automated
alternative. In Proceedings of the Twentieth Southeastern
Symposium on System Theory, 300-303.
[5] Henderson, T. C. and Linton, T. 2009. Raster Map Image
Analysis. In Proceedings of the 2009 10th International
Conference on Document Analysis and Recognition (ICDAR
'09). Washington DC, USA, 376-380.
[6] Cao, R. and Tan, C. 2002. Text/graphics separation in maps.
In Fourth IAPR Workshop on Graphics Recognition, 2390,
167–177, Springer, Berlin.
[7] Epitropou, V., Karatzas, K.D., Bassoukos, A., Kukkonen, J.
and Balk, T. 2011. A new environmental image processing
method for chemical weather forecasts in Europe. In
Proceedings of the 5th International Symposium on
Information Technologies in Environmental Engineering,
Poznan, (Golinska, Paulina; Fertsch, Marek; Marx-Gómez,
Jorge, eds.), ISBN: 978-3-642-19535-8, Springer Series:
Environmental Science and Engineering, 781-791.
[8] Karatzas K. 2009. Informing the public about atmospheric
quality: air pollution and pollen, Allergo Journal, 18, Issue
3/09, 212-217.
[9] Balk T., Kukkonen J., Karatzas K., Bassoukos A. and
Epitropou V. 2011. A European open access chemical
weather forecasting portal, Atmospheric Environment, 45,
6917-6922, doi:10.1016/j.atmosenv.2010.09.058
[10] Kukkonen, J., Klein, T., Karatzas, K., Torseth, K., Fahre Vik,
A., San Jose, R., Balk, T. and Sofiev, M. 2009. COST
ES0602: Towards a European network on chemical weather
forecasting and information systems, Advances in Science
and Research Journal, 1, 1–7.
Table 6. Results comparing the proposed system output with the manually configured AirMerge.
Pollen FMI
Atmospheric and Oceanic Physics
Group site
Number of images 15 15 15 15
Number of colors 9 11 12 12
Latitude Error (in 5o) 2.85·10-4 8.54·10-4 2.42·10-4 6.38·10-4
Longitude Error (in 5o) 0.00174 3.88·10-4 0.00164 1.39·10-4
Mean percentage of pixels with correct value 97.424 % 90.957 % 77.23 % 77.3 %
Average error per pixel 0.1265 0.0365 0.0619 0.0504
... While screen poking is used to generate synthetic mouse and keyboard events for inputting data as if done manually, screen scraping is used for the opposite direction of extracting information from an application's UI. If no structured information about the screen contents are available, this can be done via OCR and image processing [50]. Yet if such information is available -e.g., the Document Object Model (DOM) of a website -an extractor or wrapper can be used to find and obtain the relevant information [51]. ...
Conference Paper
Full-text available
Over the past years, the visualization of large and complex data sets brought up various Visual Analytics (VA) tools in order to solve domain-specific tasks. These VA tools are typically implemented as individual software components in data-flow-oriented models, meaning that data is transferred from one component to the next. While most VA frameworks rely on a monolithic architecture with features for the integration of specialized analysis methods, we consider a loose coupling of independent applications, where autonomous VA tools are used in predefined analysis sequences. To this end, we provide a characterization of the data exchange process among individual VA tools in the form of a taxonomy. This taxonomy can be used as a checklist to identify characteristics and improve the data flow of one's own multi-tool VA setup. For this purpose, we conducted a systematic investigation of the individual aspects of data exchange that are commonly found across different usage scenarios. We apply our taxonomy to three existing multi-tool frameworks, the open-source library ReVize, the toolchain editor AnyProc, and the visualization and monitoring framework Plant@Hand3D.
... Li et al. [6] proposed the Haze Optical Thickness retrieval model based on the assumption that surface reflectance varies slowly in a relative short period that could monitor the haze distribution and intensity for Beijing Olympic Games and help Beijing municipal government to carry out more measures to improve air quality conditions. Moumtzidou et al. [18] proposed a configurable semiautomatic framework for processing air quality and pollen forecast heatmaps. They integrated several existing environmental quality forecast data extraction tools with text processing and OCR (Optical Character Recognition) techniques tailored for heatmap analysis. ...
Full-text available
Air pollution is a universal problem confronted by many developing countries. Because there are very few air quality monitoring stations in cities, it is difficult for people to know the exact air quality level anytime and anywhere. Fortunately, large amount of surveillance cameras have been deployed in the cities and can capture image densely and conveniently in the cities. In this case, this provides the possibility to utilize surveillance cameras as sensors to obtain data and predict the air quality level. To this end, we present a novel air quality level inference approach based on outdoor images. Firstly, we explore several features extracted from images as the robust representation for air quality prediction. Then, to effectively fuse these heterogeneous and complementary features, we adopt multikernel learning to learn an adaptive classifier for air quality level inference. In addition, to facilitate the research, we construct an Outdoor Air Quality Image Set (OAQIS) dataset, which contains high quality registered and calibrated images with rich labels, that is, concentration of particles mass (PM), weather, temperature, humidity, and wind. Extensive experiments on the OAQIS dataset demonstrate the effectiveness of the proposed approach.
... There were some approaches to get environmental information from heatmap semi-automatically. They based on annotation tool and OCR techniques (Moumtzidou et al. 2012(Moumtzidou et al. , 2013. ...
The paper is focused on adaptation of the information extraction method for environmental heatmaps to U-matrices of Self Organising Maps—SOM neural networks. Our method bases on OCR, image processing and image recognition techniques. The approach was designed to be as much as possible general but information acquired from a heatmap and the form of a heatmap defer depending on the heatmap type. In the paper we introduce some dedicated processing steps while trying to minimize the number of changes in the previously proposed method. The results for U-matrices of SOM neural network are evaluated in the experimental study and compared with efficiency of the method for environmental maps.
... Selecting appropriate points is done manually by the AirMerge's operator, though it is possible to partially automate the process [18,27]. If no usable geographical reference grid is supplied with the heatmap, it is still possible to use the known geographical coordinates of two easily identifiable landmarks. ...
Full-text available
The AirMerge platform was designed and constructed for increasing the availability and improving the interoperability of heatmap-based environmental data on the Internet. This platform allows data from multiple heterogeneous chemical weather data sources to be continuously collected and archived in a unified repository; all the data in this repository have a common data format and access scheme. In this paper, we address the technical structure and applicability of the AirMerge platform. The platform facilitates personalized information services, and can be used as an environmental information node for other web-based information systems. The results demonstrate the feasibility of this approach and its potential for being applied also in other areas, in which image-based environmental information retrieval will be needed.
... If we consider for instance, applications in ecology, where analysis of sounds within specific landscapes is often desired or required. Here, ecologists seek to investigate the structure of soundscape by using separate sound elements such as sounds generated by organism, sounds generated by physical features of the environment (e.g., running water, wind, rain, etc.), and sounds generated from human activities (e.g., crowd, laughter, vehicle, etc.) [1], [2]. Other applications may include those in the domain of environment recognition and forgery detection applications, e.g., a proper set of environmental sounds can be used to understand higher-level semantics such as malls, offices, restaurants, or quiet streets [3], [4]. ...
Conference Paper
Environmental sounds (ES) have different characteristics, such as unstructured nature and typically noise-like and flat spectrums, which make recognition task difficult compared to speech or music sounds. Here, we perform an exhaustive feature and classifier analysis for the recognition of considerably similar ES categories and propose a best representative feature to yield higher recognition accuracy. In the experiments, thirteen (13) ES categories, namely emergency alarm, car horn, gun, explosion, automobile, helicopter, water, wind, rain, applause, crowd, and laughter are detected and tested based on eleven (11) audio features (MPEG-7 family, ZCR, MFCC, and combinations) by using the HMM and SVM classifiers. Extensive experiments have been conducted to demonstrate the effectiveness of these joint features for ES classification. Our experiments show that, the joint feature set ASFCS-H (Audio Spectrum Flatness, Centroid, Spread, and Audio Harmonicity) is the best representative feature set with an average F-measure value of 80.6%.
Frühwarnsysteme dienen zur möglichst frühzeitigen Information über eine sich anbahnende oder auftretende Gefahr, um Personen und Organisationen die Möglichkeit zu geben entsprechend darauf reagieren zu können. Die Konzeption eines Frühwarnsystems stellt komplexe Herausforderungen an die Systemarchitekten, hierzu liefert die vorliegende Arbeit ein Rahmenwerk für die Architektur zukünftiger Frühwarnsysteme.
Conference Paper
In our large library of annotated environmental recordings of animal vocalizations, searching annotations by label can return thousands of results. We propose a heat map of aggregated annotation time and frequency bounds, maintaining the shape of the annotations as they appear on the spectrogram. This compactly displays the distribution of annotation bounds for the user's query, and allows them to easily identify unusual annotations. Key to this is allowing zero values on the map to be differentiated from areas where there are single annotations.
Environmental and meteorological conditions are of utmost importance for the population, as they are strongly related to the quality of life. Citizens are increasingly aware of this importance. This awareness results in an increasing demand for environmental information tailored to their specific needs and background. We present an environmental information platform that supports submission of user queries related to environmental conditions and orchestrates results from complementary services to generate personalized suggestions. The system discovers and processes reliable data in the Web in order to convert them into knowledge. At runtime, this information is transferred into an ontology-structured knowledge base, from which then information relevant to the specific user is deduced and communicated in the language of their preference. The platform is demonstrated with real world use cases in the south area of Finland, showing the impact it can have on the quality of everyday life.
Conference Paper
Environmental data are considered of utmost importance for human life, since weather conditions, air quality and pollen are strongly related to health issues and affect everyday activities. This paper addresses the problem of discovery of air quality and pollen forecast Web resources, which are usually presented in the form of heatmaps (i.e. graphical representation of matrix data with colors). Towards the solution of this problem, we propose a discovery methodology, which builds upon a general purpose search engine and a novel post processing heatmap recognition layer. The first step involves generation of domain-specific queries, which are submitted to the search engine, while the second involves an image classification step based on visual low level features to identify Web sites including heatmaps. Experimental results comparing various visual features combinations show that relevant environmental sites can be efficiently recognized and retrieved.
Conference Paper
The ACM International Workshop on Multimedia Analysis for Ecological Data (MAED'12) is held as part of ACM Multimedia 2012. MAED'12 is concerned with the processing, interpretation, and visualization of ecology-related multimedia content with the aim to support biologists in their investigations for analyzing and monitoring natural environments, with particular attention to living organisms and pollution effects.
Full-text available
It is common practice to present environmental information of spatial nature (like atmospheric quality patterns) in the form of pre-processed images. The current paper deals with the harmonization, comparison and reuse of Chemical Weather (CW) forecasts in the form of pre-processed images of varying quality and informational content, without having access to the original data. In order to compare, combine and reuse such environmental data, an innovative method for the inverse reconstruction of environmental data from images, was developed. The method is based on a new, neural adaptive data interpolation algorithm, and is tested on CW images coming from various European providers. Results indicate a very good performance that renders this method as appropriate to be used in various image-processing problems that require data reconstruction, retrieval and reuse.
Conference Paper
Full-text available
It is common practice to disseminate Chemical Weather (air quality and meteorology) forecasts to the general public, via the internet, in the form of pre-processed images which differ in format, quality and presentation, without other forms of access to the original data. As the number of on-line available Chemical Weather (CW) forecasts is increasing, there are many geographical areas that are covered by different models, and their data could not be combined, compared, or used in any synergetic way by the end user, due to the aforementioned heterogeneity. This paper describes a series of methods for extracting and reconstructing data from heterogeneous air quality forecast images coming from different data providers, to allow for their unified harvesting, processing, transformation, storage and presentation in the Chemical Weather portal.
Full-text available
Atmospheric quality (AQ) is among the most significant and well-studied concepts of environmental quality. Yet the question of perception, interpretation and communication of AQ information remains open, and needs to be addressed in an effective way in order to design better abatement strategies for the protection of human health and new information services for quality of life support. Contemporary AQ information dissemination methods and tools can now make use of various telecommunication channels for pull and push service provision. These may include internet for e-mail notification, World Wide Web for detailed AQ related information, SMS for early warning services, WAP and J2ME applications in mobiles and PDAs for enhanced graphical and informative content on the move, street panels (VMS), and voice services for personal communication support. In addition, location-based services are also emerging in the field of personalised, quality of life services. The present paper addresses atmospheric (air and pollen) information dissemination, its legal background, and contemporary information systems and technologies. The paper also draws conclusions concerning the design and the functions of AQ in-formation services, thus providing input to research institutes, authorities and companies working in the area of environmental information dissemination
Full-text available
It is common practice to publish environmental information via the Internet. In the case of geographical coverage information such as pollutant concentration charts and maps in chemical weather forecasts, such data are published as web-resolution images. These forecasts are commonly presented with an associated value-range pseudocolor scale, which represents a simplified version of the original data, obtained through dispersion models and related post-processing methods. In this paper, the numerical and signal processing performance of a method to reconstruct numerical data from the published coverage images is evaluated by comparing the reconstructed data with the original forecast data
Conference Paper
Full-text available
Citizens are increasingly aware of the influence of environmental and meteorological conditions on the quality of their life. This results in an increasing demand for personalized environmental information, i.e., information that is tailored to citizens’ specific context and background. In this demonstration, we present an environmental information system that addresses this demand in its full complexity in the context of the PESCaDO EU project. Specifically, we will show a system that supports submission of user generated queries related to environmental conditions. From the technical point of view, the system is tuned to discover reliable data in the web and to process these data in order to convert them into knowledge, which is stored in a dedicated repository. At run time, this information is transferred into an ontology-based knowledge base, from which then information relevant to the specific user is deduced and communicated in the language of their preference.
Full-text available
The COST ES0602 action provides a forum for benchmarking approaches and practices in data exchange and multi-model capabilities for chemical weather forecasting and near real-time information services in Europe. The action includes approximately 30 participants from 19 countries, and its duration is from 2007 to 2011 ( Major efforts have been dedicated in other actions and projects to the development of infrastructures for data flow. We have therefore aimed for collaboration with ongoing actions towards developing near real-time exchange of input data for air quality forecasting. We have collected information on the operational air quality forecasting models on a regional and continental scale in a structured form, and inter-compared and evaluated the physical and chemical structure of these models. We have also constructed a European chemical weather forecasting portal that includes links to most of the available chemical weather forecasting systems in Europe. The collaboration also includes the examination of the case studies that have been organized within COST-728, in order to inter-compare and evaluate the models against experimental data. We have also constructed an operational model forecasting ensemble. Data from a representative set of regional background stations have been selected, and the operational forecasts for this set of sites will be inter-compared and evaluated. The Action has investigated, analysed and reviewed existing chemical weather information systems and services, and will provide recommendations on best practices concerning the presentation and dissemination of chemical weather information towards the public and decision makers.
Conference Paper
Full-text available
The authors present an automated method for the digitization, vectorization and storage of land record maps. Map data is input into a vision system as raw data using a solid-state camera. After noise removal, the processed data is converted to vector data by an efficient algorithm. This point data that presents the key points can be stored on peripheral storage devices for future reconstruction and modification of the map.< >
A European chemical weather forecasting portal is presented in this paper that has been developed within the COST (European Cooperation in Science and Technology) ES0602 action, “Towards a European Network on Chemical Weather Forecasting and Information Systems”. The portal includes an access to a substantial number (currently 21) of available chemical weather forecasting systems and their numerical forecasts; these cover in total more than 30 regions in Europe. This portal can be used, e.g., to find out, which services are available for a specific domain, for specific source categories or for specific pollutants. The portal currently expands its functionalities to allow for a harmonized presentation and inter-comparison of the various available forecasts, as well as for the computation of model ensemble predictions. It provides functions for obtaining relevant supplementary information, e.g., using the Model Documentation System of the European Environmental Agency. The new portal is an open access system, through which chemical weather forecasts can be added to the system, and the predictions can be accessed, analysed and inter-compared. Such a single point of reference for the European chemical weather forecasting information has previously not been in operation. We present the characteristics of the new portal, and discuss how this activity complements the GEMS and PROMOTE air quality forecasting portals.
Conference Paper
Raster map images (e.g., USGS) provide much information in digital form; however, the color assignments and pixel labels leave many serious ambiguities. A color histogram classification scheme is described, followed by the application of a tensor voting method to classify linear features in the map as well as intersections in linear feature networks. The major result is an excellent segmentation of roads, and road intersections are detected with about 93% recall and 66 % precision.
Conference Paper
The separation of overlapping text and graphics is a challenging problem in document image analysis. This paper proposes a specific method of detecting and extracting characters that are touching graphics. It is based on the observation that the constituent strokes of characters are usually short segments in comparison with those of graphics. It combines line continuation with the feature line width to decompose and reconstruct segments underlying the region of intersection. Experimental results showed that the proposed method improved the percentage of correctly detected text as well as the accuracy of character recognition significantly.