Conference Presentations

Achieving Meaningful Metrics from Your Test Automation Tools

In addition to the efficiency improvements you expect from automated testing tools, you can-and should-expect them to provide valuable metrics to help manage your testing effort. By exploiting the programmability of automation tools, you can support the measurement and reporting aspects of your department. Learn how Jack Frank employs these tools with minimal effort to create test execution
status reports, coverage metrics, and other key management reports. Learn what measurement data your automation tool needs to log for later reporting. See examples of the operational reports his automation tools generate, including run/re-run/not run, pass/fail, percent complete, and percent of overall system tested. Take with you examples of senior management reports, including Jack's favorite, "My Bosses' Boss Test Status Report"-names will be changed to hide the guilty. Regardless of the

Jack Frank, Mosaic Inc
Legal Compliance in Quality Assurance

In many industries, we must comply with state or federal statutes, government regulations, and other legal standards. The Sarbanes-Oxley Act (SOX) has brought a new awareness to these issues within testing. So, how do you incorporate legal compliance into your QA and test efforts, and how do you get the information you need to do the job well? Elle Ringham, who deals with these important issues every day at Fidelity National Financial, shares her knowledge and experiences. First, she helps you understand the types of legal compliance bodies that can affect applications and offers research methods for industry specific areas of legal compliance, including internal sources, websites, and search strategies. Then, she discusses the artifacts and metrics needed to be maintained

Elle Ringham, Fidelity National Financial
System Implementation Details: Understand, Test, Ignore

Join John Lambert to see examples of implementation details that cause problems in GUI-based applications, APIs, and Web systems. Find out how you can develop skills that will help you test your current product and will transfer to entirely different products in the future.

John Lambert, Microsoft Corporation
Test Improvement for Highly Reliable NYSE Trading Systems

With billions of dollars changing hands every day, financial trading systems demand extremely high accuracy and reliability. So, how do you improve test process performance in the areas of time to market and efficiency and at the same time reduce failures? Over the last three years, using process and project measurement data as a guide, SIAC has focused on doing exactly that. Steve Boycan highlights the key elements of the process changes that have led to SIAC's current performance: the use of a rigorous requirements engineering process; controlled parallel and iterative work flows; changes to the level of abstraction in test documentation; emphasis on test planning, analysis, and design; causal analysis; and improving the test team's skills.

Steve Boycan, SIAC
Software Inspections for the Common Folk

Many have tried to implement software inspections in commercial development shops ... and failed. Common reasons include “not enough resources," "takes too long," and "makes people uncomfortable." Ron Yun demonstrates how ADP has implemented both "full” and "lite" inspection practices that reduce resources, automate many aspects of the process, and help take the personality issues out of the process. With their practices, ADP completes inspection meetings quickly and has eliminated all manual forms. They have merged key inspection information with data from their defect tracking system to determine progress toward the goal of finding defects as early as possible in development. Their automation tools track the identified issues, provide a permanent audit trail of the inspection, and automatically generate metrics to report status and gauge effectiveness.

Ronald Yun, ADP
It's Too Darn Big: Test Techniques for Gigantic Systems

Structuring test designs and prioritizing your test effort for large and complex software systems are daunting tasks, ones that have beaten many, very good test engineers. If you add concurrency issues and a distributed system architecture to the mix, some would simply throw up their hands. At Microsoft, where Keith Stobie plies his trade, that is not an option. Keith and others have reengineered their testing, employing dependency analysis for test design, model property static checking, "all pairs" configuration testing, robust unit testing, and more. They employ coverage to successfully help select and prioritize tests and make effective use of random testing including fuzz testing security. Finally, models of their systems help them generate good stochastic tests and act as test oracles for automation.

  • Test checklists for large, complex, distributed systems
Keith Stobie, Microsoft Corporation
Show Me the Money

Turn to The Last Word, where software professionals who care about quality give you their opinions on hot topics. This month, read how adding gauges to your software can show stakeholders how well it is meeting their goals.

Jeff Patton's picture Jeff Patton
A Systematic View of User Acceptance Testing

Acceptance testing is a vital and specific form of testing whether you are tasked with rolling out an enterprise application package, releasing a major system enhancement, or developing acceptance tests in an agile development project. In addition, acceptance tests can give some teeth to service level agreements and software acquisition contracts. However, most treat acceptance testing as the same activity as system testing-but done by different staff. That is wrong! Because acceptance testing is not about bug hunting and breaking the software, you need a different strategy. With over 25 years of experience covering acceptance testing for all types of systems from safety critical control systems to standard financial applications, Geoff Quentin shares his views on how to do acceptance testing correctly.

Geoff Quentin, QBIT Ltd
Free Test Tools are Like a Box of Chocolates

You never know what you are going to get! Until you explore, it can be hard to tell whether a free, shareware, or open source tool is an abandoned and poorly documented research project or a robust powerhouse of a tool. In this information-filled presentation, Danny Faught shows you where open source and freeware tools fit within the overall test tool landscape. During this double session, Danny installs and tries out several tools right on the spot and shares tips on how to evaluate tools you find on the Web. Find out about licensing, maintenance, documentation, Web forums, bugs, and more. Discover the many different types of testing tools that are available for free and where to find them. Danny demonstrates examples of tools that you can put to use as soon as you get back to the office.

Danny Faught, Tejas Software Consulting
Using Personas to Improve Testing

Too often testers are thrown into the testing process without direct knowledge of the customers' behaviors and business process. As a tester, you need to think and act like a customer to make sure the software does-in an easy-to-use way-what the customer expects. By defining personas and using them to model the way real customers will use the software, you can have the complete customer view in designing test cases. Get the basics of how to implement customer personas, their limitations, and ways to create tests using them. See examples of good bugs found using personas while learning to write bug reports based on them.

  • What you need to know to develop customer personas
  • Use customer personas for designing test cases
  • The types of bugs found by using personas but missed by other techniques
Robyn Edgar, Microsoft

Pages

CMCrossroads is a TechWell community.

Through conferences, training, consulting, and online resources, TechWell helps you develop and deliver great software every day.