Content uploaded by Chen Zhang
Author content
All content in this area was uploaded by Chen Zhang on Sep 06, 2019
Content may be subject to copyright.
Cloud Environment for Disseminating NASS
Cropland Data Layer
Chen Zhang†‡ , Liping Di∗† ‡, Zhengwei Yang§, Li Lin†‡, Eugene G. Yu†,
Zhiqi Yu†‡, Md. Shahinoor Rahman† ‡, Haoteng Zhao† ‡
†Center for Spatial Information Science and Systems, George Mason University, Fairfax, VA 22030, USA
‡Department of Geography and Geoinformation Sciences, George Mason University, Fairfax, VA 22030, USA
§Research and Development Division, USDA National Agricultural Statistics Service, Washington, DC 20250, USA
Email: {czhang11, ldi*}@gmu.edu, Zhengwei.Yang@nass.usda.gov, {llin2, gyu, zyu4, mrahman25, hzhao22}@gmu.edu
Abstract—Cropland Data Layer (CDL) is an annual crop-
specific land use map produced by the U.S. Department of
Agricultural (USDA) National Agricultural Statistics Service
(NASS). The CDL products are officially hosted on CropScape
website which provides capabilities of geospatial data visual-
ization, retrieval, processing, and statistics based on the open
geospatial Web services. This study utilizes cloud computing
technology to improve the performance of CropScape applica-
tion and Web services. A cloud-based prototype of CropScape
is implemented and tested. The experiment results show the
performance of CropScape is significantly improved in the cloud
environment. Comparing with the original system architecture of
CropScape, the cloud-based architecture provides a more flexible
and effective environment for the dissemination of CDL data.
Index Terms—Cloud computing, Cropland Data Layer, Cyber-
infrastructure, Interoperability, OGC Web Services
I. INTRODUCTION
As the only field-level crop planting map for the entire con-
tiguous United States (CONUS) at present, the Cropland Data
Layer (CDL) product of the U.S. Department of Agriculture
(USDA) National Agricultural Statistics Services (NASS) is
widely adopted in agro-geoinformatics. Since the first release
in 1997, the CDL products have been used to support many
agricultural applications such as land use land cover change
[1], crop area estimation [2] crop loss assessment [3], flood
mapping [4]–[7]. The national cropland maps, crop frequency
data layers, cultivated layer, and confidence layer of CDL can
be downloaded from the official website of USDA NASS [8].
However, it is a labor-intensive task for users to explore the
national-scale crop planting map using the desktop Geographic
Information Systems (GIS) software. To facilitate the use
of CDL data, CropScape, a Web-based GIS application for
agricultural geospatial data visualization and analytics, was
released by USDA NASS and Center for Spatial Information
Science and Systems of George Mason University in 2011
[9]. It provides the interactive visualization and retrieval of
all historical CDL maps as well as a variety of on-the-fly
geospatial functionalities such as crop acreage statistics and
map creation in PDF. More importantly, CropScape offers
the Open Geospatial Consortium (OGC) standards-compliant
geospatial Web services for disseminating all CDL data and
geoprocessing over the web [10]. According to the statistics of
∗corresponding author
Google Analytics, more than 208,000 users around the world
have visited and interacted with CropScape as of May 2019.
During the past decade, the volume and variety of Earth
Observation (EO) data is experiencing the explosive growth
[11]–[14]. This trend has boosted the development of advanced
computing and Cyberinfrastructure (CI) in Earth science do-
main [15]–[18]. As an important technology in CI, cloud
computing has been used to support geospatial applications
and services [19]–[22]. For example, [23] and [24] presented
the cloud-based CIs for flood monitoring and analysis, [25]
developed a Web service-based application for demographic
information modeling in the cloud. Meanwhile, the feasibility
of integrating OGC Web Services with cloud computing tech-
nology has been explored, [26] introduced a cloud environment
for processing EO data with the OGC Web Services, [27] de-
scribed a framework for implementing the Earth science model
as OGC Web Processing Service in the cloud environment.
This study aims to facilitate the dissemination of CDL data
using cloud computing technology. We immigrate the Crop-
Scape application and Web services to the cloud environment.
All data, applications, and Web services of the CropScape are
distributed as virtual machine (VM) and managed by the cloud
platform. The remainder of the paper is organized as follow.
Section II describes the high-level system architecture of
cloud-based framework for CropScape. Section III compares
the performance of CropScape Web portal and Web services in
the cloud computing environment and the current environment.
Conclusion and future works are given in the section IV.
II. SY ST EM DESIGN
The original system architecture of CropScape was de-
scribed in [9] and [28], which contains application layer,
service layer, and data layer. The application layer refers to
GIS applications and clients that support OGC Web services.
The service layer includes standard geospatial Web services
such as Web Map Service (WMS), Web Feature Service
(WFS), Web Coverage Service (WCS), and Web Processing
Service (WPS). All raster data and vector data are stored in
the data layer. As of July 2019, the CropScape data catalog
covers full volume of CDL data from 1997 to 2018, the crop
frequency layer from 2008 to 2018, the crop mask layer from
Preprint submitted to 8th International Conference on Agro-Geoinformatics (July 2019)
Fig. 1. High-level architecture of CropScape in cloud environment
2014 to 2018, U.S. boundary layers, global land cover layer,
and a few auxiliary layers.
Based on the original system architecture, we developed a
cloud-based framework to enable CropScape applications and
services in the cloud environment. Figure 1 shows the high-
level architecture of the proposed framework. Two new layers,
the Infrastructure-as-a-Service (IaaS) layer and Platform-as-a-
Service (PaaS) layer, are added to this framework. The IaaS
layer is the foundational layer of the cloud environment. It
consists of the physical cloud infrastructures, which is built
with the server cluster, and offers the computing resources
(e.g. CPU, memory, disk, bandwidth). The PaaS layer mainly
manages virtualization, operating system, middleware, and
runtime. In this study, the IaaS service and PaaS service
are offered by GeoBrain Cloud (https://cloud.csiss.gmu.edu),
a private cloud platform operated by Center for Spatial In-
formation Science and Systems of George Mason University.
The GeoBrain Cloud uses Apache CloudStack as the manage-
ment framework to manage the network, storage, computing
resources, and VMs in the server clusters.
In the proposed framework, the data layer, service layer,
and application layer could be distributed in multiple VMs.
For example, all CDL data are stored in the data VMs, all
Web services are hosted in the service VMs, the CropScape
Web portal is hosted in the application VM. Meanwhile, the
OGC standards compliant design of CropScape system makes
the framework interoperable with other GIS applications and
Web service clients.
TABLE I
EXA MPL ES O F CROPSC AP E WEB SERVICES IN CLOU D ENVIRONMENT
Web Service Operation Request Example
WMS for CONUS CDL GetMap https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms cdlall?SERVICE=WMS&
VERSION=1.1.1&REQUEST=GetMap&LAYERS=cdl 2015&TRANSPARENT=
true&SRS=EPSG:102004&BBOX=-3987459.135,168311.354, 4472862.725,4177587.
947&FORMAT=image/png&WIDTH=800&HEIGHT=400
WCS for CONUS CDL GetCoverage https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms cdlall?service=wcs&version=1.0.0&
request=getcapabilities
WMS for COUNS Boundary GetMap https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms conustiger.cgi?LAYERS=conus
counties overview%2Cconus states overview&SRS=EPSG%3A102004&FORMAT=
image%2Fpng&SERVICE=WMS&VERSION=1.1.1&REQUEST=GetMap&STYLES=
&EXCEPTIONS=application%2Fvnd.ogc.se inimage&BBOX=-5228593,- 1874202.5,
5229977,5446796.5&WIDTH=300&HEIGHT=210
WFS for COUNS Boundary GetFeature https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms conustiger.cgi?service=wfs&
version=1.1.0&request=GetFeature&typeName=conus counties&maxFeatures=100
III. EXP ER IM EN TS
A. Implementing CropScape in Cloud Environment
The cloud-based CropScape prototype Web portal can be
accessed at https://cloud.csiss.gmu.edu/CropScape. Figure 2
shows the screenshot of the CropScape Web portal in the cloud
environment. The Web service endpoint of the cloud-based
CropScape prototype is https://cloud.csiss.gmu.edu/cdlserver.
Table I lists some examples of CropScape Web services in the
cloud environment.
Fig. 2. CropScape Web portal in cloud environment.
There are few differences in both software and hardware
between the current environment and the cloud environment.
CropScape is originally deployed in two servers which han-
dling jobs sent from the CropScape Web portal and Web
services respectively. The CropScape Web portal is hosted on
the physical server with 6 cores CPU and 48GB memory.
The CropScape Web services are hosted on the server with
12 cores CPU and 32GB memory. In the cloud environment,
the CropScape instance is allocated 8 cores CPU and 16GB
memory. On the other hand, the original CropScape is running
on the Windows Server system, the PaaS service offered by
the GeoBrain Cloud is based on Linux system. Additionally,
we utilized MapServer software to serve data in the OGC Web
services and Geospatial Data Abstraction Library (GDAL) to
power geospatial functionalities. The version of these software
in two environments is different. Table II summarizes the main
specification of current environment and cloud environment.
B. Performance Test of CropScape Web Portal
To test the performance of CropScape in the cloud envi-
ronment, firstly, we compared the loading time of the portal
of prototype (https://cloud.csiss.gmu.edu/CropScape) with the
current Web portal (https://nassgeodata.gmu.edu/CropScape).
We developed a script to automatically visit the portal and
record the response time of each visit. Based on the result of
1,000 visits for each portal, the result of performance test is
illustrated as Figure 3. It can be seen from the test result that
the average response time of cloud-based prototype is 1.3s. As
a control, the average response time of the original CropScape
Web portal is 3.4s.
Fig. 3. Average response time of CropScape Web portal (lower is better).
TABLE II
SPE CIFI CATI ON OF TE ST ENVIRONMENT
Server of CropScape Web portal Server of CropScape Web services VM in cloud environment
CPU 2.1Ghz, 6 Cores 3.47Ghz, 12 Cores 2.0Ghz, 8 Cores
Memory 48GB 32GB 16GB
System Windows Server 2012 R2 Windows Server 2008 R2 Ubuntu 16.04
MapServer v6.0.3 v6.0.3 v7.0.7
GDAL v2.0.4 v1.11.1 v2.2.2
TABLE III
EXAMPLES OF CROP SCAP E WEB SE RVIC ES USE D IN PERFORMANCE TE ST
Request Cloud Environment Current Environment
CONUS
CDL
https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms cdlall?SERVICE=
WMS&VERSION=1.1.1&REQUEST=GetMap&LAYERS=cdl
2018&TRANSPARENT=true&SRS=EPSG:102004&BBOX=
-3987459.135, 168311.354,4472862.725, 4177587.947&FORMAT=
image/png&WIDTH=2000&HEIGHT=1000
https://nassgeodata.gmu.edu/CropScapeService/wms cdlall.cgi?
SERVICE=WMS&VERSION=1.1.1&REQUEST=GetMap&
LAYERS=cdl 2018&TRANSPARENT=true&SRS=EPSG:
102004&BBOX=-3987459.135,168311.354, 4472862.725,4177587.
947&FORMAT=image/png&WIDTH=2000&HEIGHT=1000
State
CDL
https://cloud.csiss.gmu.edu/cdlserver/cgi-bin/wms cdl ia?SERVICE=
WMS&VERSION=1.1.1&REQUEST=GetMap&LAYERS=cdl
2018 ia&STYLES=&SRS=EPSG:4326&BBOX=-97,40.2, -90, 43.7&
WIDTH=2000&HEIGHT=1000&FORMAT=image/png
https://nassgeodata.gmu.edu/CropScapeService/wms cdl ia.cgi?
SERVICE=WMS&VERSION=1.1.1&REQUEST=GetMap&
LAYERS=cdl 2018 ia&STYLES=&SRS=EPSG:4326&BBOX=
-97, 40.2,- 90,43.7&WIDTH=2000&HEIGHT=1000&FORMAT=
image/png
Fig. 4. Average response time of Web Map Service for 2018 CONUS CDL
(lower is better).
C. Performance Test of CropScape Web Services
The second experiment tested the performance of CDL Web
services. We compared the response time of WMS requests
in the cloud environment and current environment. Table III
shows the examples of the WMS requests for the CONUS
CDL and state-level CDL. Figure 4 and Figure 5 illustrates
the average response time for loading 2018 CONUS CDL map
and 2018 CDL map of Iowa state respectively.
IV. CONCLUSION AND FUTURE WOR KS
This study utilized the cloud computing technology to
facilitate the disseminating CDL data. Based on the current
system architecture, we added the PaaS layer and IaaS layer to
enable the cloud-based framework for CropScape application
Fig. 5. Average response time of Web Map Service for 2018 CDL of Iowa
state (lower is better).
and its Web services. The experiment result shows the average
response time of CropScape Web portal in the cloud environ-
ment is much lower than in the current server. Meanwhile,
the performance of CropScape Web Service is significantly
improved in the cloud environment.
Currently, the WMS, WCS, and WFS of CropScape has
been implemented in the GeoBrain Cloud. The cloud-based
WPS of CropScape is still under development. The prototype
of cloud-based CropScape described in the experiment section
is based on the single VM. The performance of the current
implementation would be affected while the service is over-
loaded. This issue could be solved by distributing multiple
instances in the cloud and dynamically allocating computing
resources to the application and Web services. In the future,
we will investigate the scalability of the proposed architecture
and test the performance in real use case.
ACK NOW LE DG EM EN T
This study is supported by USDA NASS.
REFERENCES
[1] T. J. Lark, R. M. Mueller, D. M. Johnson, and H. K. Gibbs, “Mea-
suring land-use and land-cover change using the U.S. department of
agriculture’s cropland data layer: Cautions and recommendations,” In-
ternational Journal of Applied Earth Observation and Geoinformation,
vol. 62, pp. 224–235, Oct. 2017.
[2] X.-P. Song, P. V. Potapov, A. Krylov, L. King, C. M. Di Bella,
A. Hudson, A. Khan, B. Adusei, S. V. Stehman, and M. C. Hansen,
“National-scale soybean mapping and area estimation in the United
States using medium resolution satellite imagery and field survey,”
Remote Sensing of Environment, vol. 190, pp. 383–395, Mar. 2017.
[3] M. S. Rahman, L. Di, E. Yu, L. Lin, C. Zhang, and J. Tang, “Rapid
Flood Progress Monitoring in Cropland with NASA SMAP,” Remote
Sensing, vol. 11, no. 2, p. 191, Jan. 2019.
[4] L. Lin, L. Di, E. G. Yu, L. Kang, R. Shrestha, M. S. Rahman, J. Tang,
M. Deng, Z. Sun, C. Zhang, and L. Hu, “A review of remote sensing
in flood assessment,” in 2016 5th International Conference on Agro-
Geoinformatics, 2016.
[5] M. S. Rahman, L. Di, R. Shrestha, E. G. Yu, L. Lin, C. Zhang, L. Hu,
J. Tang, and Z. Yang, “Agriculture flood mapping with Soil Moisture
Active Passive (SMAP) data: A case of 2016 Louisiana flood,” in 2017
6th International Conference on Agro-Geoinformatics, Aug. 2017, pp.
1–6.
[6] L. Lin, L. Di, J. Tang, E. Yu, C. Zhang, M. S. Rahman, R. Shrestha, and
L. Kang, “Improvement and Validation of NASA/MODIS NRT Global
Flood Mapping,” Remote Sensing, vol. 11, no. 2, p. 205, Jan. 2019.
[7] R. Shrestha, L. Di, E. G. Yu, L. Kang, Y. Shao, and Y. Bai, “Regression
model to estimate flood impact on corn yield using MODIS NDVI and
USDA cropland data layer,” Journal of Integrative Agriculture, vol. 16,
no. 2, pp. 398–407, Feb. 2017.
[8] USDA NASS Research and Science, “Cropland Data Layer Releases,”
https://www.nass.usda.gov/Research and Science/Cropland/Release/,
2019.
[9] W. Han, Z. Yang, L. Di, and R. Mueller, “CropScape: A Web service
based application for exploring and disseminating US conterminous
geospatial cropland data products for decision support,” Computers and
Electronics in Agriculture, vol. 84, pp. 111–123, 2012.
[10] W. Han, L. Di, and Z. Yang, “Developing geoprocessing service for
Cropland Data Layer thematic map creation,” in 2013 Second Interna-
tional Conference on Agro-Geoinformatics (Agro-Geoinformatics), Aug.
2013, pp. 572–576.
[11] S. Nativi, P. Mazzetti, M. Santoro, F. Papeschi, M. Craglia, and
O. Ochiai, “Big Data challenges in building the Global Earth Observa-
tion System of Systems,” Environmental Modelling & Software, vol. 68,
pp. 1–26, Jun. 2015.
[12] M. Chi, A. Plaza, J. A. Benediktsson, Z. Sun, J. Shen, and Y. Zhu, “Big
Data for Remote Sensing: Challenges and Opportunities,” Proceedings
of the IEEE, vol. 104, no. 11, pp. 2207–2219, Nov. 2016.
[13] L. Di, “Big data and its applications in agro-geoinformatics,” in
2016 IEEE International Geoscience and Remote Sensing Symposium
(IGARSS), Jul. 2016, pp. 189–191.
[14] A. Karmas, A. Tzotsos, and K. Karantzalos, “Geospatial Big Data for
Environmental and Agricultural Applications,” in Big Data Concepts,
Theories, and Applications, S. Yu and S. Guo, Eds. Cham: Springer
International Publishing, 2016, pp. 353–390.
[15] J. L. Schnase, T. J. Lee, C. A. Mattmann, C. S. Lynnes, L. Cinquini,
P. M. Ramirez, A. F. Hart, D. N. Williams, D. Waliser, P. Rinsland, W. P.
Webster, D. Q. Duffy, M. A. McInerney, G. S. Tamkin, G. L. Potter,
and L. Carriere, “Big Data Challenges in Climate Science: Improving
the next-generation cyberinfrastructure,” IEEE Geoscience and Remote
Sensing Magazine, vol. 4, no. 3, pp. 10–22, Sep. 2016.
[16] L. Di, Z. Sun, and C. Zhang, “Facilitating the Easy Use of Earth
Observation Data in Earth System Models through CyberConnector,”
in AGU Fall Meeting Abstracts, 2017.
[17] Z. Sun, L. Di, H. Hao, X. Wu, D. Q. Tong, C. Zhang, C. Virgei, H. Fang,
E. Yu, X. Tan, P. Yue, and L. Lin, “CyberConnector: A service-oriented
system for automatically tailoring multisource Earth observation data to
feed Earth science models,” Earth Science Informatics, vol. 11, no. 1,
pp. 1–17, Mar. 2018.
[18] S. Wang and M. F. Goodchild, “CyberGIS for Transforming Geospatial
Discovery and Innovation,” in CyberGIS for Geospatial Discovery and
Innovation, ser. GeoJournal Library, S. Wang and M. F. Goodchild, Eds.
Dordrecht: Springer Netherlands, 2019, pp. 3–10.
[19] P. Yue, H. Zhou, J. Gong, and L. Hu, “Geoprocessing in Cloud
Computing platforms – a comparative analysis,” International Journal
of Digital Earth, vol. 6, no. 4, pp. 404–425, Jul. 2013.
[20] K. Evangelidis, K. Ntouros, S. Makridis, and C. Papatheodorou,
“Geospatial services in the Cloud,” Computers & Geosciences, vol. 63,
pp. 116–122, Feb. 2014.
[21] Z. Li, C. Yang, Q. Huang, K. Liu, M. Sun, and J. Xia, “Building
Model as a Service to support geosciences,” Computers, Environment
and Urban Systems, vol. 61, pp. 141–152, Jan. 2017.
[22] C. Yang, Q. Huang, Z. Li, K. Liu, and F. Hu, “Big Data and cloud com-
puting: Innovation opportunities and challenges,” International Journal
of Digital Earth, vol. 10, no. 1, pp. 13–53, 2017.
[23] Z. Wan, Y. Hong, S. Khan, J. Gourley, Z. Flamig, D. Kirschbaum,
and G. Tang, “A cloud-based global flood disaster community cyber-
infrastructure: Development and demonstration,” Environmental Mod-
elling & Software, vol. 58, pp. 86–94, Aug. 2014.
[24] M. M. Morsy, J. L. Goodall, G. L. O’Neil, J. M. Sadler, D. Voce,
G. Hassan, and C. Huxley, “A cloud-based flood warning system for
forecasting impacts to transportation infrastructure systems,” Environ-
mental Modelling & Software, vol. 107, pp. 231–244, Sep. 2018.
[25] L. Lin, L. Di, C. Zhang, L. Hu, J. Tang, and E. G. Yu, “Developing
a Web Service Based Application for Demographic Information Mod-
eling and Analyzing,” in 2017 6th International Conference on Agro-
Geoinformatics, 2017, pp. 14–18.
[26] C. Zhang, L. Di, Z. Sun, E. G. Yu, L. Hu, L. Lin, J. Tang, and M. S.
Rahman, “Integrating OGC Web Processing Service with cloud comput-
ing environment for Earth Observation data,” in 2017 6th International
Conference on Agro-Geoinformatics, 2017.
[27] C. Zhang, L. Di, Z. Sun, L. Lin, E. G. Yu, and J. Gaigalas, “Exploring
cloud-based Web Processing Service: A case study on the implementa-
tion of CMAQ as a Service,” Environmental Modelling & Software, vol.
113, pp. 29–41, Mar. 2019.
[28] W. Han, Z. Yang, L. Di, B. Zhang, and C. Peng, “Enhancing Agricultural
Geospatial Data Dissemination and Applications Using Geospatial Web
Services,” IEEE Journal of Selected Topics in Applied Earth Observa-
tions and Remote Sensing, vol. 7, no. 11, pp. 4539–4547, Nov. 2014.