Towards Cloud Native Continuous Delivery: An Industrial Experience Report
Davide Taibi, Kari Syst¨
Tampere University of Technology
Email: davide.taibi@tut.ﬁ, kari.systa@tut.ﬁ
Abstract— Several companies suffer from long delays be-
tween implementing and delivering software features.
Continuous Integration, Delivery, and Deployment tools and
practices are increasing their popularity since they can help
companies decrease delivery costs and delays, and reduce bugs
and errors in the delivery processes thanks to automation.
Companies have difﬁculties in implementing Continuous
Integration, Delivery, and Deployment. There is information
available on the subject, but collecting, studying, and distribut-
ing that information can be very costly.
The CI and CD pipelines are examples of well founded
Cloud Native applications that run in a variety of different
conﬁgurations and provide ﬂexibility and speed for companies,
but require elastic execution platforms and knowledge of
modern cloud technologies for optimal use.
In this experience report we present the lessons learned dur-
ing the implementation of a cloud-based Continuous Delivery
tool stack. Our experience includes the design and implemen-
tation of a pure cloud and hybrid cloud based Continuous
Integration and Delivery solutions at Vincit.
In 2015 Vincit, a Finnish software development company,
started a Continuous Integration and Delivery project as
part of the Finnish N4S project  with a the goal of
providing software developers build and test automation tools
that support complex Continuous Delivery pipelines which
contain multiple build, test, and deployment targets with
Continuous Delivery practices can save time and reduce
costs in projects. Development teams can ensure that soft-
ware has been tested and ready for production when working
and veriﬁed builds are rolled from integration pipelines
into staging environments. Managers and sales personnel
can happily tell the customer that software is ready to be
deployed when the customer asks to see the latest sprint
result in action. Customers do not have to wait for a week
or a month of deployment delays .
Continuous Delivery can be implemented in stages, sup-
porting ﬁrst the processes where the needs for automation
and orchestration are the greatest, or just implementing the
parts which offer the most returns for invested money and
time. The degrees of implementing automation have been
deﬁned in the Continuous Delivery Maturity Model which
offers guidelines for what to implement in what order to be
more efﬁcient , , .
We had previous experience in build and test automation
solutions from running tools such as Hudson, Jenkins, and
Travis CI, which we currently use in our repertoire as well.
However, the creation of conﬁguration pipelines and delivery
was still complex in 2015. Since then, all these projects
have seen tremendous improvement in terms of features and
maturity, and nowadays, are much more viable for advanced
Continuous Delivery usage.
We consider Continuous Delivery to be a Cloud Native
application as it requires elastic computing capacity and
the capacity requirements change rapidly. This has ﬁnancial
implications since it is hard to estimate the short- and long-
term ﬁnancial implications of implementing systems. Con-
tinuous Delivery pipelines also require advanced monitoring
In this paper we report the lessons learned while im-
plementing a customized open-source Continuous Delivery
system. The different phases we look into include research,
requirements deﬁnition, design, implementation, and refac-
toring of a whole Continuous Delivery software system on
top of pure cloud infrastructure. This purely cloud based
system architecture later mutates into a hybrid cloud archi-
tecture, and we brieﬂy discuss transforming from pure cloud
to hybrid cloud solutions.
The paper is structured into distinct sections that discuss
design, implementation and evaluation of a Continuous De-
livery system. Section 2 reports Background and Related
Works. Section 3 describes the different possibilities for im-
plementing a CD platform. Section 4 analyzes the difference
between self-hosted and cloud platforms. Section 5 discusses
cost of systems while Section 6 discusses the choice of
Continuous Delivery tools. Section 7 reports the case study
with an analysis of how the different steps of the CD pipeline
have been implemented. Section 8 evaluates the results of the
case study, and ﬁnally, Section 9 draws conclusions.
II. BACKGROU ND
Continuous Integration  is related to the frequent auto-
matic software integration, which commonly means building
and testing changed source code. Frequency means that
software is built and tested periodically or, for example, on
every version control commit.
Continuous Delivery  includes Continuous Integration
and making sure that the software is always conﬁgurable
and deployable. This requirement is usually satisﬁed with an
automated staging environment deployment.
Continuous Deployment includes always automatically
deploying software to production when it is committed to
version control system branches corresponding to production
environments and qualiﬁed by the automatic tests to be
Continuous Delivery and Deployment are often used in
the same context and can be mistaken with each other, but
in academic context differentiating the terminology is impor-
tant . Figure 1 shows the relations between Continuous
Integration, Delivery, and Deployment.
Fig. 1. Relations of Continuous Integration, Delivery and Deployment
Companies cannot hence implement Continuous Deploy-
ment without Continuous Integration and Delivery systems.
For the sake of clarity, we deﬁne different terms adopted
in this work:
•Continuous Integration is a group of practices that aims
to improve software development quality and speed
with build and test automation in order to improve
reproducibility and to remove the chance of errors
from manual build step execution or environment state
mutation in build processes. Continuous Integration is
a subset of Continuous Delivery;
•Continuous Delivery is a group of practices that includes
Continuous Integration and adds to them the automated
end-to-end testing and delivery of software in such
a way that software builds are stateless, reproducible,
and proven deployable across target environments and
platforms. Continuous Delivery makes delivering recent
software iterations to production at any given time feasi-
ble and aims at guaranteeing deployability. Continuous
Delivery is a super-set of Continuous Integration, and a
subset of Continuous Deployment. Continuous Delivery
does not include automatically deploying software to
•Continuous Deployment is a group of practices that
includes the aforementioned Continuous Integration and
Delivery practices but adds to them the practice of
automatically deploying software to production envi-
ronments. Continuous Deployment ideally removes the
need of manual production environment updates and
aims to roll-forward only deployments. Continuous De-
ployment is a super-set of Continuous Delivery.
III. HOW CA N CONTINUOUS DELI VERY B E
There are many tools for Continuous Integration and
Delivery written in many languages illustrated in Table I.
OPE N-SOURCE CONTINUOUS INTEGRATION AND DELI VERY TO OLS
Software Implementation Published Maintainer
Buildbot Python 2003 Mitchell et. al
GoCD Java 2007 ThoughtWorks, Inc.
Jenkins Java 2011 Kawaguchi et. al
Travis CI Ruby 2011 Travis CI, GmbH
Strider CD Ruby 2012 Radchenko et. al
GitLab CI Ruby 2012 GitLab, Inc.
Drone Go 2014 drone.io
Java is one of the most commonly used languages for
CD tools, powering services such as Jenkins  and GoCD
. There are also alternatives for Java powered platforms
such as Buildbot  written in Python, Travis CI 
written in Ruby, and Strider  and Drone  written in
Node.js. These are just few examples of the tools available
for implementing Continuous Integration. Many of the listed
alternatives are available as partly or fully open-source
Equally, many architectural models exist for Continuous
Integration and Delivery systems. The systems range from
simple examples running bash or Python scripts to multi-
tiered enterprise solutions that can be hosted in multiple data
centres. For example, the simplest of build systems can be
implemented in a few hours on top of Buildbot on a single
computer. Some platforms such as Travis CI or Snap CI 
require a multiple machine set up just to operate on-premise.
Platforms and tools such as Make and Buildbot can be
perfectly viable for implementing Continuous Delivery for a
software product, but the set up of the pipeline from devel-
opment to production server deployments with conﬁguration
management and source code builds can be more difﬁcult
to master and scale. Some speciﬁcally tailored software
platforms intended for building a speciﬁc technology solution
such as Azure Pipelines for Microsoft products can be much
easier to utilize, because they might support the scenarios
that one can commonly run into when setting up Continuous
Developers have a myriad of options for adopting Contin-
uous Delivery into their work and project ﬂows which can
be categorized as:
•managed SaaS solutions such as Travis CI;
•hostable enterprise solutions such as Azure DevOps, or;
•self-hosted open-source solutions such as GoCD.
Each of the aforementioned options can be valid, depend-
ing on the current and future situation and conditions in the
company. Smaller companies should prefer to use lightweight
managed solutions and avoid over-committing to one path
unless there is a clear need for a heavyweight system.
Larger enterprises might need multiple different systems to
support their operations. In each case, an understanding of
the different alternatives and their service and cost models
is necessary in making the right choice.
IV. DECIDING ON CLO UD PLATF ORM S VER SUS
SEL F-HOSTED SOLUTIONS
One important factor in choosing the right alternative is the
solution’s extensibility. If in 5 years time we need a feature
that is not implemented, what would we do? Many of the
current systems do not offer extensibility. Travis CI offers
access to its deployment tools, but a lot of platforms offer
no access to their inner workings or source code, and cannot
be modiﬁed at all. We already knew some requirements for
the Continuous Delivery system we wanted to implement in
our Minimum Viable Product (MVP).
The ﬁrst of our requirements was extensibility. Another
one was the ability to support cross-platform builds. We
wanted the same tool to be usable on macOS, Windows,
and Linux environments. The last important requirement was
that we could host the Continuous Delivery service in private
or public data centers. Having someone else host the service
was simply too rigid of an option. Our customers have a need
for ﬂexibility, so we wish to offer them as many options as
Hence we decided that we wanted to invest in an open-
source solution that we could extend and program ourselves,
and hopefully host ourselves, if needed. In the open-source
front there are a few options that have a community around
them offering support and tool-sets to each other. Narrowing
the search down, we found ourselves facing yet another
decision: choosing the right open-source option for our
V. COST MODE LS F O R CONTINUOUS DEL IVE RY
Different Continuous Delivery solutions have different
cost models, which are of essential knowledge when making
management decisions regarding the implementation and
lifespan of Continuous Delivery systems. Understanding of
the cost implications of the different implementation choices
is an important decision criterion.
Operating expenses and capital expenses are two main
elements of the cost models , . In short, capital
expenses are multiple-term costs that are tied to a system
for a long time, such as data centre investments and system
vendor acquisition costs, network infrastructure acquisition
and initial large licence purchases. Operating expenses are
single-term costs that are tied to running the system at a
certain load, for example network transfer and electricity
costs and manual maintenance labor. Using an operation-
ready SaaS platform has a pay-per-use cost model.
Implementing a platform with on-premise hardware or
cloud hardware can have very different cost models which
can include servers acquisition, management, power, cooling,
backups, maintenance, licences, and an assortment of other
things, which can be very hard to predict.
Most cloud service providers bill for networking, CPU,
RAM, and storage capacity. For example, Amazon Web Ser-
vices bills for network components such as Internet and VPN
connectivity and outbound trafﬁc, server resource usage such
as CPU cores and memory, and storage capacity. In addition
to this, for the users of proprietary operating systems, such
as Windows or Red Hat Enterprise Linux, licence fees apply
on per-machine basis. In case of proprietary Continuous
Delivery tools, a licence is also applicable. , 
On-premise computing capacity, in addition to the cloud
platform components, adds the cost for power, cooling,
and staff work. Moreover, on-premise systems also require
resources for handling with power outages, loss of data and
other similar issues.
From these concepts we deﬁned to ﬁxed and ﬂoating costs.
•Fixed costs are the baseline costs that are tied to the
running of the system and rarely change: data centers
and equipment are examples of ﬁxed costs.
•Floating costs are costs that change in the lifespan of
The ability and willingness to pay large ﬁxed costs and
make purchases up-front affects the choice of service and
hosting model. If a company can predict capacity needs in
detail and has liquidity, then an upfront investment can be
wise. If capacity needs are not static and change over time,
making optimal investment choices can be hard. With cloud
computing platforms capital is not tied to ﬁxed investments,
and risks are reduced. 
VI. CH O OS I NG T HE CONTINUOUS DEL IVE RY TOO L S
One of our requirements that was realized and reﬁned
during the project was the ability to support and specify de-
pendencies for build steps and different projects . Com-
plex dependency management is important when building,
for example, a microservice architecture or complex multiple
tier software where one wants to deﬁne the build, test, and
deployment pipeline as a dependency graph. For example,
it might be necessary to build the backend and frontend
software ﬁrst, and, then test their component integration, and
ﬁnally test the end-to-end functionality of the system.
Considering the different requirements regarding tooling
support for multiple platforms, languages and tools we
decided to look further into options that offered script based
and non-opinionated architectures. The most prominent of
these systems was GoCD. GoCD has most of the things we
wanted our tool to have. It is:
•open-source and has a permissive licensing;
•platform agnostic and runs anywhere where Java is
•non-opinionated and runs anything you can script to run
via system shell;
•scalable, both horizontally and vertically, and lastly;
•has a stable user community and good documentation.
All these factors combined, the only issue we had with the
project was its lack of an established plugin ecosystem, such
as the one in Jenkins. Jenkins CI has a myriad of different
extensions and supports most common tools because of
its age and community. GoCD was, in 2016, in middle
of implementation of some very central features such as
dynamic build agent provisioning. Small delays, however,
are things that we were willing to deal with when investing
into long-term tooling.
VII. IMPLEMENTING CONTINUOUS DEL IV ERY
We implemented the CD system in AWS cloud based
computer system. Our implementation work for the private
cloud at Vincit began by creating a network layout.
We created a VPC (Virtual Private Cloud) in Frankfurt
with /16 CIDR block that was compatible with our existing
network layout, and created three different subnets in that
network segment. Our subnets consisted of a /24 manage-
ment subnet, a /24 GoCD server subnet, and a /24 GoCD
agent subnet. Once we had our network layout deﬁned, we
set up a VPN gateway to it and opened a ticket to our ISP
(Internet Service Provider) requesting that our ofﬁce network
be connected via our router with VPN to the AWS network
and routing policies be conﬁgured. This took about two
weeks and a few failed conﬁguration attempts from our ISP,
but after the wait we had our networks deﬁned and were able
to connect to the AWS cloud via private connection from our
ofﬁce. During this waiting period we started setting up our
virtual server infrastructure and software components into
AWS to avoid downtime in the whole process.
We started our EC2 (Elastic Compute Cloud) virtual
server conﬁguration by searching for the Ubuntu LTS AMI
(Amazon Machine Image) from the AWS Marketplace ,
which houses software that can be run on the AWS. Most
Linux distributions can be found on the Marketplace free of
charge as they have permissive licensing schemes.
After ﬁnding and launching our Ubuntu instances, we
continued by conﬁguring them with SSH keys and setting up
secure connectivity with them. After successfully connecting
to the instances we installed updates and provisioned the
instances with Salt. Salt then proceeded to automatically
install Sensu and GoCD software to the nodes. At this point
we had the architecture illustrated in Figure 2.
The system seemed to work in the beta testing environ-
ment, and we had everything running smoothly. Builds were
executed on the GoCD agents and we were managing nodes
with an integrated Salt solution. Our Salt scripts would install
packages and whole programming environments required in
builds, and fetch and conﬁgure the SSH keys and conﬁgura-
tions needed to interact with source code repositories.
We were also tracking the raw node statistics with Sensu,
which was running on every agent node.
Analyzing the degree of system usage is easy on most
IaaS cloud platforms. Most IaaS platforms are virtualized
and offer access to the virtualization system’s CPU usage
AWS offers numerous statistics of an instance that can be
gathered and stored for an arbitrary period of time.
Some of the statistics that AWS offers via its proprietary
CloudWatch system for an EC2 virtual machine instance
are: 1) CPU usage; 2) disk read and write statistics, and; 3)
network device usage. Memory usage statistics are not pro-
vided by the virtualization platform, but can be additionally
monitored with reporting scripts running in the virtualized
guest operating system. 
After running the service for a while we realized that
we were running workers that were not used during the
night time since all our developers were out-of-ofﬁce. This
meant that we were running idle computing capacity but
paid for the full capacity. Because AWS supports capacity
scaling with ASGs (Auto Scaling Groups) and Autolaunch
Conﬁgurations, we created an automatically scaling cluster
that could increase capacity in the morning and decrease
capacity in the night time. The system would run zero
instances in the night and 2-4 instances between 6AM and
8PM, local time. This would total to 40% less running
time for worker instances, which reduced the total costs of
four worker instances and two management and instances
by over 25%. Since, we earlier saw that the EC2 running
costs constitute for about 80% of our overall costs, we could
reduce our overall AWS costs by about 20%.
Automatic scaling requires that each time an instance is
started, all necessary software and conﬁguration is installed
to it. We conﬁgured our Linux instances to run a boot-
strapping script that installs a Salt Minion  to a node
each time a cluster machine is brought up, and Salt, our
orchestration tool, would conﬁgure the node as a GoCD
agent after that. All-in-all, our whole bootstrapping for the
instance constituted to a simple shell script that can easily
be modiﬁed to install Salt on any Linux distribution. It
also can be reconﬁgured largely on the same principles
to bootstrap a node that is running macOS or Windows
for Salt conﬁguration. This removes the need for manually
conﬁguring computers. More details on the script can be
found in the associated thesis work .
VIII. EVALUATI ON RE SULTS
The systems implemented and discussed in this paper
offered data and subjective experiences that we would like to
further discuss and evaluate. Some interesting technical as-
pects of the system were cost factors, the technical evolution
of the tooling, improvements of processes and other tooling
we made with implementing Continuous Delivery tools, and
measurements of the systems.
A. Cost and labor factors
Work wise, we have invested about two weeks of time
in different stages of design and meetings throughout the
project. One person has also worked on the project and on
a Master thesis describing it for about three months . In
grand total, we have about four months of work invested
in research and development and the implementation of the
system. It might be possible to implement a Continuous
Delivery system from scratch for a small amount of projects
much faster, but overall, the effort of studying the tooling
and theory associated to software automation, orchestration,
metrics and data gathering, and different details such as cloud
platform speciﬁcs is quite time consuming.
In early 2016 we were paying for a few medium sized
nodes in AWS. We had a master build node, an orchestration
node and a monitoring node for GoCD in AWS. These
amounted to three ﬁxed cloud computing nodes. Rest of the
ﬂeet was ﬂexible build capacity that we were running as
needed. In addition to the computation capacity we are also
Fig. 2. Pure Cloud Continuous Integration and Delivery System Architecture
paying for approximately 200GB of ﬁxed storage capacity
and low network transfer costs. We did not have ﬁxed
dedicated capacity tied to the build system, but acquisition
and utilization of extra capacity was fairly easy. This setup
is illustrated in Figure 2.
The total cost of the system was in the few hundreds of
dollars per month range for the pure cloud solution. Most
notable individual cost of the system was an AWS VPN
gateway for a hybrid connection that cost USD 100 per
month, with the computing capacity costing between USD
100 to USD 300 per month, totaling to a well less than USD
500 per month in all cases for a small build ﬂeet.
Maintenance and administration work has required an
average of one to two day investment of work per month
on the tools, resulting in notable continuous running costs.
These costs, are however, hard to deﬁne and attribute, as
some could be associated to project setup, and at least some
of the work used on maintenance of the tools would be
required in a managed tool such as Travis CI because projects
need to be kept up-to-date and build failures investigated.
B. Evolution of the technology stack
Later in 2016 and 2017 we moved into a hybrid and then
onto a on-premise hosting model due to hosting structure
changes, and moved the GoCD master node into a ﬁxed
virtual machine and inside a Docker container. The build
agents are run as Docker containers, too. This seems to offer
excellent price-to-performance ratio due to lowered over-
heads. Containerization also corrects many of the problems
of running multiple build agents on the same physical or
virtual server. All agent workspaces are isolated because the
GoCD agent Java processes run inside containers, having
very limited and closely controlled access to the host system
resources aside from ephemeral ﬁle system and network
access. This has worked well for us for well over a two
At the time of writing in late 2018 we are running Jenkins
as our primary and GoCD as our secondary internal build
system with both doing Continuous Delivery for projects.
Jenkins has a larger amount of projects, slaves, and users due
to its more complete feature set. GoCD also has a distinct
group of projects, but most new projects are set up in Jenkins.
This is because we have previous history with Jenkins, and
its development for both core and and plugins have moved
forward with a faster pace than in GoCD. Most reasonable
technical requirements are also met by the existing plugins.
Jenkins has also introduced a large feature called pipelines
in 2016 and 2017. Pipelines enable writing build deﬁnitions
in either declarative or scripted manner, and enable shareable
libraries for deﬁning common build operations. Pipelines also
enable easy parallel build steps and allow granular build
dependency deﬁnition in free-form graphs.
A major note we would like to make is that both Jenkins
and GoCD have access control and identity management
limitations. In our use they are currently only offered to
our developers internally in company intranet due to security
considerations. Both could be opened up to the Internet
with some restrictions, but little is to be gained in terms of
usability, and system security will need to regularly audited.
and GoCD both do feature LDAP based login backends,
allowing easy directory integration. Both also offer varying
quality SSO solutions based on SAML 2.0, OAuth 2.0 and
OpenID Connect 1.0. These plugins, however, are commu-
nity driven and rarely if ever audited security-wise, and the
responsibility of ensuring their security often falls to the
administrators of the tools.
C. Perceived and measured improvements
Our build duration dropped by approximately 25% due
to moving our builders to the AWS cloud and having them
less loaded. This is largely connected to a single builder
node only processing a single project and not taking any
additional load. We were able to select the correct build
machine sizes for various projects and select the optimal
amount of resources to host our systems, making it possible
to ﬁnely tune the offered build capacity to the needs of our
developers. This is also possible with our on-premise build
system that is based on virtualization.
Failure rates due to system errors have reduced, because
we are only executing a single build on a single system
or container, and are not introducing conﬂicts, caching
problems, or computing resource exhaustion into the build
processes. These are all things that are fairly expensive to
debug, because a person has to go and investigate the build
and system logs to determine an indeterministic reason for a
build failure. The exact reduction in errors is not transparent,
but early data suggests we have solved some of our concur-
rency, virtualization, and container based problems, moving
from platform problems to build node or job conﬁguration
errors. The latter are much easier to locate and ﬁx.
It also seems that we improved our build tooling on many
parts during the project. GoCD supports resource tagging,
build environment speciﬁcation, heterogeneous builders,
management of project dependencies, and other features that
are hard to ﬁnd in traditional build tools. We have not faced
any performance issues or instability from the tool.
In addition to improving our systems technically we have
introduced the concept of push-button deliveries and high
deployability. Only some projects are using push-button
deployments on GoCD, but we have implemented similar
features using Travis CI for Continuous Delivery with the
AWS Elastic Beanstalk platform using the dpl tool by Travis
CI . We are currently introducing push-button delivery
to new projects, which has reduced the need for manual
deployments and saved work time in projects.
All-in-all, the perceived improvements are considerable.
The concrete measurable improvements which will save
our customers money will hopefully come apparent in the
upcoming months and years. Quantifying the project results
is hard at this stage, because we do not have extensive
data available yet. Many of the beneﬁts we have achieved
were not expected to be immediately available though, and
will accumulate in time when an increasing number of
projects adopt the Continuous Delivery methodology and
gain conﬁdence in rapidly available customer deliverables
and increased deployment rates.
In addition to us implementing continuous improvements
to the build tooling since 2016, we have gained a consid-
erable amount of knowledge in the Continuous Integration,
Delivery, and Deployment domain. The increased knowledge
has steadily improved the processes and tools in use, and
will hopefully enable us to move forward with software
development practices as well.
D. Metrics, data, and information
Getting features implemented and delivered to the cus-
tomer with less work, fewer errors, shorter development
cycle, and less downtime is the main thing that automation
enables . To improve the rate of delivery we hope to
implement a comprehensive measuring system that could
give us insight on deltas between development, deployment,
and activation times. We want to measure features and
releases done per month in addition to other system statistics
such as build durations and frequencies. We also want to
make this information transparent to software development
teams and customers .
At this time the collection and evaluation of metrics and
data is hard due to the use of multiple different build systems
that build different kinds of software projects. The effort
of closely measuring and analyzing the systems and their
differences has not been yet undertook, but is an essential
step in developing more robust and sophisticated tooling.
Metrics and data enable improved decision making pro-
cesses which are based on scientiﬁc methods. Most build
and deployment tools offer some built-in data visualization
and metrics, but few offer simple APIs for exporting the
said metrics into usable formats. The implementation of a
metrics and measurement service that integrates into different
projects, their services, hosting platforms, and other tools is
therefore a task that will require more research efforts.
In this paper we report on our experience in implement-
ing a Continuous Delivery system based on open-source
technologies. We ﬁrst described the different possibilities
to implement a Continuous Delivery platform, highlighting
pros and cons and we compared cost between hosted and
cloud solutions. In our case study we then implemented
a complete cloud-based Continuous Delivery pipeline, and
then we migrated into a hybrid solution. It is interesting to
notice how and when we maximized the beneﬁts between
cloud and hosted solutions, and how we integrated them.
The same beneﬁts could be transferred to any Cloud
Native application, not only Continuous Delivery. Companies
could learn from our experience when and why it could be
beneﬁcial to migrate parts of their systems into the cloud.
We believe that a pure cloud hosting and containeriza-
tion offers the most ﬂexibility for implementing Continuous
Integration, Delivery and Deployment systems. The current
trends in open-source build systems seem to moving towards
so called Cloud Native Continuous Integration. Quite re-
cently CloudBees and Kohsuke Kawaguchi have been driving
the Jenkins development into a more containerized direction,
even opting to redesign existing solutions and dropping
backwards compatibility to move development faster going
Future work include the deﬁnition of a cost model to
support companies in understanding the most suitable alter-
natives between different cloud providers and a in-premise
solution. We will also analyze the reasons why companies
are migrating to Cloud Native pipelines , of patterns
and anti-patterns of the Cloud Native pipelines, following
the previous approaches applied in , , . Moreover,
we are also planning to support companies in using data
collected from the CD platform to predict different software
characteristics. As an example, it we are planning to extend
data-driven models for software reliability , maintenance
 also considering dynamic measures  and suitability
of Cloud Native patterns .
 A. H¨
akli, ”Implementation of Continuous Delivery Systems”
Master Thesis. Tampere University of Technology, 2016.
 Need4Speed Project. DIMECC N4S-Program: Finnish Software Com-
panies Speeding Digital Economy.
 P. M. Duvall, S. Matyas, and A. Glover, Continuous Integration:
Improving Software Quality and Reducing Risk. Addison-Wesley
 J. Humble and D. Farley, Continuous Delivery: Reliable Software
Releases Through Build, Test, and Deployment Automation. Addison-
 V. Lenarduzzi, D. Taibi, ”MVP explained: A systematic mapping
study on the deﬁnitions of minimal viable product.” 42nd Euromicro
Conference on Software Engineering and Advanced Applications,
SEAA 2016, 112-119. doi:10.1109/SEAA.2016.56
 ThoughtWorks et al., ”Go Continuous Delivery,” https://www.go.cd/,
retrieved May 13, 2016.
 C. Caum, ”Continuous Delivery Vs. Continuous Deployment. What’s
the diff?” https://puppet.com/blog/continuous-delivery-vs-continuous-
deployment-whats-diff, retrieved Sept 2018.
 D. Taibi, V. Lenarduzzi, ”On the deﬁnition of microservice bad
smells.” IEEE Software, 35(3), 56-62. doi:10.1109/MS.2018.2141031
 A. Rehn, T. Palmborg, and P. B¨
ostrom, ”The Continuous Delivery Ma-
turity Model,” http://www.infoq.com/articles/ Continuous-Delivery-
Maturity-Model, Tech. Rep., 2013, retrieved Oct 2018.
 E. Minick, ”Continuous Delivery Maturity Model,”
maturity-model/, Tech. Rep., 2014, retrieved Oct 2018.
 P. Bahrs, ”Adopting the IBM DevOps approach for continu-
ous software delivery: Adoption paths and the DevOps maturity
model, ” https://www.ibm.com/ developerworks/library/d-adoption-
paths/, Tech. Rep., 2013, retrieved Oct 2018.
 K. Kawaguchi et al., ”Jenkins,” https://jenkins.io/, retrieved Oct 2018.
 D. J. Mitchell et al., ”Buildbot,” http://buildbot.net/, retrieved Oct
 Travis CI, GmbH, ”Travis CI,” https://travis-ci.com, retrieved Oct
 I. Radchenko et al., ”Strider Continuous Delivery,”
http://stridercd.com/, retrieved Oct 2018.
 drone.io, ”drone.io,” https://drone.io/, retrieved Oct 2018.
 ThoughtWorks, Inc., ”Snap CI,” https://snap-ci.com/, retrieved Oct
 Microsoft, Inc., ”Team Foudation Server,”
retrieved Oct 2018.
 D. Taibi, A. Janes, V. Lenarduzzi, ”How developers perceive smells
in source code: A replicated study.” Information and Software Tech-
nology, 92. 2017. doi:10.1016/j.infsof.2017.08.008
 ThoughtWorks, Inc., ”Fan-out & Fan-in,” https://www.go.cd/videos/
go-fan-out-fan.html, retrieved Oct 2018.
 A. Damodaran, Applied Corporate Finance: A User’s
Manual, 4th ed. John Wiley and Sons, 1999. [Online].
Available: http://people.stern.nyu.edu/ adamodar/New Home
 Amazon, Inc., ”How AWS Pricing Works,”
https://d0.awsstatic.com/whitepapers/aws pricing overview.pdf,
Tech. Rep., 2016, retrieved Oct 2018.
 ——, ”AWS Marketplace,” https://aws.amazon.com/marketplace/, re-
trieved Oct 2018.
 ——, ”AWS CloudWatch,” https://aws.amazon.com/cloudwatch/, re-
trieved Oct 2018.
 D. Taibi, V. Lenarduzzi, C. Pahl, ”Architectural patterns for microser-
vices: A systematic mapping study.” 8th International Conference on
Cloud Computing and Services Science, CLOSER, 2018.
 SaltStack, Inc., ”SaltStack architecture for system command and con-
trol,” https://saltstack.com/saltstack-architecture/, retrieved Oct 2018.
 Docker, Inc., ”Docker,” https://www.docker.com/, retrieved Oct 2018.
 L. Lavazza, S. Morasca, D. Taibi, D. Tosi, ”On the deﬁni-
tion of dynamic software measures.” International Symposium on
Empirical Software Engineering and Measurement (ESEM) 2012.
 K. Haase, ”Dpl deployment tool,” https://github.com/travis-ci/dpl, re-
trieved Oct 2018.
 K. Kawaguchi, ”Jenkins: Shifting Gears,”
https://jenkins.io/blog/2018/08/31/shifting-gears/, retrieved Oct
 D. Taibi, V. Lenarduzzi, C. Pahl, ”Processes, motivations, and
issues for migrating to microservices architectures: An empir-
ical investigation.” IEEE Cloud Computing, 4(5), 22-32. 2017
 P. Rosati, F. Fowley, C. Pahl, D. Taibi, T. Lynn, ”Making the cloud
work for software producers: Linking architecture, operating cost
and revenue.” 8th International Conference on Cloud Computing and
Services Science (CLOSER 2018), 2018-January 364-375
 S. Aldawood, F. Fowley, C. Pahl, D. Taibi, X. Liu, ”A coordination-
based brokerage architecture for multi-cloud resource markets.” 4th
International Conference on Future Internet of Things and Cloud
Workshops, W-FiCloud 2016, 7-14. doi:10.1109/W-FiCloud.2016.19
 L. Lavazza, S. Morasca, D. Taibi, D. Tosi ”An empirical in-
vestigation of perceived reliability of open-source Java programs.”
ACM Symposium on Applied Computing (SAC), 1109-1114.
 V. Lenarduzzi, A.C. Stan, D. Taibi, D. Tosi, G. Venters, ”A dynamical
quality model to continuously monitor software maintenance.” 11th
European Conference on Information Systems Management, ECISM