Zsolt Fabok presents several methods that can be used to find areas which are worth testing so that organizations do not have to spend more effort on testing than what is absolutely necessary.
Eric Minick discusses continuous delivery challenges in the enterprise where large projects, distributed teams or strict governance requirements have resulted in increased automation efforts throughout the life cycle.
The panelists present various approaches to API automation, sharing from their experiences.
John Penix describes the test automation system and the supporting build system infrastructure used by Google.
Zsolt Fabok provides guidance on selecting those sections of code that are most likely to profit from automated testing and leaving out those where chances for errors are low.
Paul Grenyer discusses why and how to create a Walking Skeleton - an implementation of the thinnest possible slice of real functionality that we can automatically build, deploy and test end-to-end.
Dave Hart introduces the “developer in test” role more testing at the unit level and adding a level of testing between unit and system, and providing testing frameworks for regression system testing.
Jonathan Lipps introduces Selenium, a functional testing framework, discussing and demoing how Selenium is used in the automated testing stack, then shares some gotchas and best practices.
Yehoram Shenhar and Alistair McKinnell present a way of doing testing having every team member involved in planning, estimating, and defining tests, testability being an architectural system attribute.
Doug Hiebert discusses the principles and objectives behind automated testing, TDD, Unit and Integration Testing, using asserting and mocking to write tests, and static analysis.
Daniel Spiewak and Aaron Bedra take a look at code verifying starting with Tony Hoare’s paper on testing(1969), type theory, and language-integrated proof systems.