Assess Your Agility

Suppose you’ve been using XP for a few months. How can you tell if you’re doing it properly? The ultimate measure is the success of your project, but you may wish to review and assess your approach to XP as well.

To help you do this, I’ve created a quiz that focuses on five important aspects of agile development. It explores results rather than specific practices, so you can score well even after customizing XP to your situation. If you aren’t using XP at all, you can also use this quiz to assess your current approach.

This quiz assesses typical sources of risk. Your goal should be to achieve the maximum score in each category—which is well within the grasp of experienced XP teams. Any score less than the maximum indicates risk, and an opportunity for improvement.

To take the quiz, answer the following questions and enter your scores on a photocopy of the blank radar diagram (Figure 4-2). Don’t give partial credit for any question, and if you aren’t sure of the answer, give yourself zero points. The result should look something like Figure 4-1. The score of the lowest spoke identifies your risk, as follows:

  • 75 points or less: immediate improvement required (red)

  • 75 to 96 points: improvement necessary (yellow)

  • 97, 98, or 99: improvement possible (green)

  • 100: no further improvement needed

Example assessment

Figure 4-1. Example assessment

Note

The point values for each answer come from an algorithm that ensures correct risk assessment of the total score.[12] This leads to some odd variations in scores. Don’t read too much into the disparities between the values of individual questions.

To see the XP solution for each of these questions, cross-reference the sections listed under “XP Practices” in Tables 4-1 through 4-5.

Self-assessment chart

Figure 4-2. Self-assessment chart

Self-Assessment Quiz

Table 4-1. Thinking

Question

Yes

No

XP Practices

Do programmers critique all production code with at least one other programmer?

5

0

Pair Programming

Do all team members consistently, thoughtfully, and rigorously apply all the practices that the team has agreed to use?

75

0

Pair Programming; Root-Cause Analysis; Retrospectives

Are team members generally focused and engaged at work?

5

0

Energized Work

Are nearly all team members aware of their progress toward meeting team goals?

4

0

Informative Workspace

Do any problems recur more than once per quarter?

0

5

Root-Cause Analysis; Retrospectives

Does the team improve its process in some way at least once per month?

5

0

Retrospectives

Table 4-2. Collaborating

Question

Yes

No

XP Practices

Do programmers ever make guesses rather than getting answers to questions?

0

75

The XP Team

Are programmers usually able to start getting information (as opposed to sending a request and waiting for a response) as soon as they discover their need for it?

4

0

Sit Together

Do team members generally communicate without confusion?

4

0

Sit Together; Ubiquitous Language

Do nearly all team members trust each other?

4

0

The XP Team; Sit Together

Do team members generally know what other team members are working on?

1

0

Stand-Up Meetings

Does the team demonstrate its progress to stakeholders at least once per month?

4

0

Iteration Demo; Reporting

Does the team provide a working installation of its software for stakeholders to try at least once per month?

1

0

Iteration Demo

Are all important stakeholders currently happy with the team’s progress?

3

0

Reporting; Iteration Demo; Real Customer Involvement

Do all important stakeholders currently trust the team’s ability to deliver?

3

0

Trust; Reporting

Table 4-3. Releasing

Question

Yes

No

XP Practices

Can any programmer on the team currently build and test the software, and get an unambiguous success/fail result, using a single command?

25

0

Ten-Minute Build

Can any programmer on the team currently build a tested, deployable release using a single command?

5

0

Ten-Minute Build

Do all team members use version control for all project-related artifacts that aren’t automatically generated?

25

0

Version Control

Can any programmer build and test the software on any development workstation with nothing but a clean check-out from version control?

25

0

Version Control

When a programmer gets the latest code, is he nearly always confident that it will build successfully and pass all its tests?

5

0

Continuous Integration

Do all programmers integrate their work with the main body of code at least once per day?

4

0

Continuous Integration

Does the integration build currently complete in fewer than 10 minutes?

4

0

Ten-Minute Build

Do nearly all programmers share a joint aesthetic for the code?

1

0

Coding Standards

Do programmers usually improve the code when they see opportunities, regardless of who originally wrote it?

4

0

Collective Code Ownership; Refactoring

Are fewer than five bugs per month discovered in the team’s finished work?

1

0

No Bugs

Table 4-4. Planning

Question

Yes

No

XP Practices

Do nearly all team members understand what they are building, why they’re building it, and what stakeholders consider success?

25

0

Vision

Do all important stakeholders agree on what the team is building, why, and what the stakeholders jointly consider success?

25

0

Vision

Does the team have a plan for achieving success?

4

0

Release Planning

Does the team regularly seek out new information and use it to improve its plan for success?

2

0

Release Planning

Does the team’s plan incorporate the expertise of business people as well as programmers, and do nearly all involved agree the plan is achievable?

3

0

The Planning Game

Are nearly all the line items in the team’s plan customer-centric, results-oriented, and order-independent?

4

0

Stories

Does the team compare its progress to the plan at predefined, timeboxed intervals, no longer than one month apart, and revise its plan accordingly?

4

0

Iterations

Does the team make delivery commitments prior to each timeboxed interval, then nearly always deliver on those commitments?

4

0

Iterations; “Done Done”; Slack; Estimating

After a line item in the plan is marked “complete,” do team members later perform unexpected additional work, such as bug fixes or release polish, to finish it?

0

25

“Done Done”

Does the team nearly always deliver on its release commitments?

3

0

Risk Management

Table 4-5. Developing

Question

Yes

No

XP Practices

Are programmers nearly always confident that the code they’ve written recently does what they intended it to?

25

0

Test-Driven Development

Are all programmers comfortable making changes to the code?

25

0

Test-Driven Development

Do programmers have more than one debug session per week that exceeds 10 minutes?

0

3

Test-Driven Development

Do all programmers agree that the code is at least slightly better each week than it was the week before?

25

0

Refactoring; Incremental Design and Architecture

Does the team deliver customer-valued stories every iteration?

3

0

Iterations; Incremental Design and Architecture

Do unexpected design changes require difficult or costly changes to existing code?

0

3

Simple Design

Do programmers use working code to give them information about technical problems?

1

0

Spike Solutions

Do any programmers optimize code without conducting performance tests first?

0

3

Performance Optimization

Do programmers ever spend more than an hour optimizing code without customers’ approval?

0

3

Performance Optimization

Are on-site customers rarely surprised by the behavior of the software at the end of an iteration?

4

0

Incremental Requirements

Is there more than one bug per month in the business logic of completed stories?

0

3

Customer Tests

Are any team members unsure about the quality of the software the team is producing?

0

1

Exploratory Testing; Iteration Demo; Real Customer Involvement



[12] Each question has a red, yellow, or green risk level. A zero score on any question leads to a total score no better than the corresponding color. Questions with scores between 25 and 75 are “red” questions, questions with scores between 3 and 22 are “yellow” questions, and questions with scores of 1 or 2 are “green” questions. Changing the risk level of one question requires reweighting the remainder: all red questions must total 75 points, all yellow questions must total 22 points, and all green questions must total 2 points. To preserve proper scoring, there may be no more than three red questions and seven yellow questions.

Get The Art of Agile Development now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.