The latest Accelerate State of DevOps Report from DORA focuses on the importance of the database and shows that integrating it into DevOps avoids time-consuming, unprofitable delays that can derail the benefits DevOps otherwise brings. It highlights four key practices that are essential to successful database DevOps ...
As development speed has become a competitive advantage, the DevOps team has sought to enable continuous integration and continuous delivery (CI/CD).
CI/CD helps developers overcome enormous strains on the software development and testing process. Over the years, software development has migrated from the enterprise level, where there are vast resources, to smaller development shops that race to keep pace with demand generated from billions of smartphones and other mobile consumer devices.
One of the biggest challenges to a successful software development cycle is the management and integration of disparate information. Think about a group project where everyone divides and conquers on tasks with a goal of integrating their work at a later date. At this point, information can come in different formats, languages and styles.
Before CI/CD, developers worked on specific software modules and merged with others later for integration and testing, which would take weeks or months to address bugs and issue fixes for a final release. We all know this process as the waterfall method. With continuous integration (CI), developers now merge changes into the main branch of the software on a daily basis. Programmers can protect against potential risks by automating testing, which helps prevent integration problems commonly associated with the waterfall method. Continuous delivery (CD) promotes a steady flow of software releases through automation. For some applications, deployment to customers may take place almost immediately after tests are completed.
For the CI/CD process to be successful, it must be fast and efficient. Any potential roadblocks that delay any part of the process increase cycle times and slow down delivery.
Typically, those who manage the IT infrastructure are seen as being responsible for providing the resources needed to support the CI/CD process. But it's also the job of operations to make sure infrastructure facilitates agile development processes that enable CI/CD with minimal bottlenecks in the pipeline.
Operate at VM and Container Level
Development workflows typically run on VMs or in containers, and so automation should be built at the same level of granularity. That requires the right storage, which can greatly accelerate CI/CD workflows by doing away with legacy logical unit numbers (LUN) or volumes — operating in VMs and containers instead.
The key is to automate everything directly at the VM, vDisk, or container level to remove complexity and make it possible to leverage storage-related processes to accelerate your CI/CD pipeline. Snapshots, cloning and replication — common features of all-flash storage — can also be factored in to save time and storage capacity.
Remove Capacity and Performance as a Constraint
It's not uncommon for storage capacity and performance to become bottlenecks in the CI/CD pipeline. If the development team has to sit and wait for processes to complete or for capacity and performance to become available, it can affect software build times and slow down testing.
By applying flash storage at the VM and container level, you can automatically prevent CI/CD workloads from interfering with each other and bogging things down. This allows you to spin up tens, hundreds, or even thousands of test workloads in parallel, and you never have to worry about load balancing or the placement of different types, sizes, or numbers of VMs inside LUNs.
Gains in performance efficiency can also be achieved by offloading operations. For example, if an application includes databases, each person on the development team needs to work with and test against their own database. Given the size of modern databases, if you try to give every developer a physical copy, storage capacity will quickly become an issue, not to mention the time spent waiting for physical copies to complete.
VM-level cloning allows you to have as many environments with full database copies as you need and to create new environments almost instantaneously.
Synchronize VMs to Accelerate Provisioning of Development and Test Environments
It's important to refresh test workloads with the latest data from production as often as possible. One of the best ways to avoid bugs and integration problems is to perform integration tests with production data. However, to accomplish this testing, you have to refresh the data and make multiple copies and rebuild the environment, including all the necessary VM settings. This process can be cumbersome, time-intensive and error-prone.
But by synchronizing VMs, you can accomplish refreshes and streamline the process without rebuilding the entire environment. As a result, there's less disruption and you can refresh regularly.
By factoring the above into your infrastructure design and purchasing decisions, you can help ensure your DevOps teams never run out of resources and spend their time focused on code, not infrastructure.