DevOps transformation projects are increasingly appealing to established enterprises. InfoQ spoke to three businesses focused on DevOps consulting to understand how they are helping businesses make the change.
At QCon New York, Eric Brewer described how advancing from continuous delivery to fast and stable continuous evolution requires a discrete construction step to define an immutable model of the system. Brewer’s compute infrastructure design team uses Helm to construct and safely validate new deployment models, prior to attempting real deployment, although the concepts are technology agnostic.
RightScale published the results of its survey report highlighting the devops trends in the industry. Docker, Puppet and Chef dominate the tools market, with Docker adoption rising in the enterprise.
On April 27, 2016 Microsoft announced a new personal workflow tool called Flow. The platform is an “IFTTT like” tool that focuses on the automation of personal tasks by orchestrating work across popular SaaS based services.
In order to implement DevOps, individuals and organizations must prepare for the culture shift, new tools, and automation. This consensus has evolved during years of debate concerning what exactly DevOps means and how to use it. There are many voices in the discussion, and even with some areas of consensus, many points are far from agreement.
Android Studio 2.0 comes with several new features and improvements: Instant Run, integration with a Google service for testing on real devices, faster emulator, faster builds, GPU profiler and debugger, support for deep linking and others.
At QCon London 2016 Peter Alvaro and Kolton Andrus shared lessons learned from a fruitful collaboration between academia and industry, which ultimately resulted in the creation of a novel method for automating failure injection testing at Netflix. Core learnings included: work backwards from what you know; meet in the middle; and adapt the theory to the reality.
The world is naturally chaotic, and we should both plan for and test that our systems can handle this chaos, Rachel Reese claimed at the recent QCon London conference describing how Jet, an e-commerce company launched in July 2015, work with microservices and chaos engineering.
On February 25th, 2016 Microsoft announced updates to their Operations Management Suite (OMS). The updates, in this particular iteration of the service, are focused on the security and audit portions of the suite and target the user experience, additional capabilities and features.
Peter Thorngren, from Volvo Trucks, explains how the future world of smart trucks and autonomous transportation systems rely deeply on continuous delivery techniques like virtualization, test automation and continuous integration.
RightScale just released the results of their annual “State of the Cloud” survey which identifies trends in cloud adoption and usage. The key findings? Organizations are investing in both public and private clouds, security is no longer the number one challenge to adoption, Docker and Ansible are growing in popularity, and central IT departments are taking on a greater role in decision making.
At the Agile Practitioners 2016 conference Huib Schoots talked about testability. He stated that low testability, anything that makes our software hard to test, slows teams down, and explored how testability can be increased.
InfoQ interviewed Boris Modylevsky about the importance of measuring code quality and how measurements can be used to improve quality, integrating static code analysis in continuous integration, testing coverage and test automation, and the benefits that continuous integration with integrated code analysis and test coverage can bring.
InfoQ ran a DevOps research question during Q4 of 2015, to find out which practices contribute the most to a healthy DevOps culture. The results show there are no predominant practices as DevOps initiatives are highly contextual.
InfoQ recently sat down with Steve Smith and discussed the ideas behind his recent blog post “End-to-End Testing Considered Harmful”. Smith talked about release testing being a form of ‘risk management theatre’, discussed the benefit of unit and acceptance testing, and stressed the value of monitoring at runtime versus the typically fragile and slow-running implementation of end-to-end testing.