I’ve written an article for InfoQ on engineering team speed, inspired by my Why Is Everything So Slow? presentation that I’ve given at Agile On the Beach and Lean Agile Scotland this year. Find out what slows down engineering teams and how to avoid getting the question ‘Why Is Everything So Slow?’
The full article is only available on InfoQ at the moment but I hope to also post it here soon. So head over to InfoQ to read more.
tl;dr No-one really likes performance reviews. But giving your team members good, actionable feedback and setting clear direction with them is very important. In my team we changed how we did performance reviews.
Note: This is the last post in a series on performance management and review. It’s worth reading the other three first:
As I mentioned in the last article, we ran the new process for four checkins, i.e. one cycle and then sought feedback from the team. They told us that:
They Liked
Focused sessions are clearer beforehand
Keeping track of things and getting constant feedback
It’s easier to remember what’s happened in a month
Makes the annual review much easier
More productive
Constructive, thought provoking guidance
They’d Like To Change
Some sessions are too similar or too regular
Some sessions are too close together
The process is too time consuming so could it be shortened?
No 360 feedback
There’s nowhere to document the managers feedback
So we now had some feedback to work with. Overall the experiment had been a success and was worth continuing with, albeit with some small changes. So what did we change?
New Themes and Timeline
Introducing a ‘Check In’
When we rolled out the initial version of the improved process we wanted to keep things simple. This meant having a small number of different themed sessions (the Atlassian process has six different themes – we had used four of them). One thing that was missing was a session where the team member could just discuss anything that they wanted. This session became the Check In.
It’s a general catch up with no agenda. As with every themed session the team member also discusses how they think they have performed and gives themselves a feedback score, and the manager then gives the team member feedback and a score themselves. Any differences are then discussed, just like in any other session.
Team members are also encouraged to put any important agenda items to discuss into the checkin form beforehand so that the manager has time to prepare.
The session is targeted at 30 minutes only, as usual.
360 Perspective
The second themed session that we introduced was 360 Perspective. Feedback had shown that team members would appreciate getting feedback from not only their manager but also their peers. The company performance management process mandated this should happen once per year.
Feedback is important but most people do not like or feel comfortable giving feedback. There’s multiple reasons for this:
Feedback is not requested or given frequently, meaning that people do not get used to giving or receiving it.
There is mistrust about what the feedback will be used for. Again, one reason for this can be because it is not received regularly enough.
Feedback is requested to be given anonymously, meaning that only the manager knows who has provided it. In some cases this is correct; for example in the case of performance issues, but generally anonymous feedback only amplifies the mistrust.
Feedback that is given is either positively or negatively biased rather than being balanced.
The 360 Perspective session gives the opportunity for the team member to collect feedback prior to the session, add that to the form, and then discuss with their manager during the session. Managers are of course free to also collect feedback for discussion as well, but all feedback requests should be open and visible to all rather than anonymous.
In the future we hope to make feedback gathering and discussion a team activity.
A New Timeline
Our team members had also told us that they thought that some of the sessions were too similar. On further investigation this was because sessions that had similar themes (Love&Loathe and Removing Barriers for example) were too close together, resulting in similar topics being discussed in each. Since new themes were also being introduced then there was also the opportunity to change the timeline and space things out better to remove potential duplication and a perception that some or all of the process was not valuable enough.
The new timeline not only spreads things out better but it also aligns key activities such as 360 Perspective with the expectations of the yearly, company wide, performance management process. This enables both processes to effectively co-exist.
This was also a good opportunity to remind everyone of the need to make sessions short and punchy, therefore avoiding making the whole process too time consuming.
Improvements To The Form
As well as introducing the new themes we also made changes to the form (get in touch if you want a copy). We added a clearer section for managers to leave feedback and the results of each session so that the form built up into a month by month record of achievement for each team member.
By doing so we enabled a very simple end of year discussion, as mandated by the companies yearly performance management process. A quick discussion to collate the month by month feedback, combined with feeding back the average of the monthly performance scores is all that is necessary. In order to set expectations as we go through the year then the form also now includes a predicted rating and graph of each team members progress.
So there’s no big performance review session that no-one looks forward to, no surprises for team member or manager at review time, and no concerns or fears about the outcome.
So What Have We Learnt?
Feedback is better when it is timely and bite sized. By introducing a monthly, targeted process then we enabled timely and frequent feedback, thus removing the anxiety around performance management. A key aspect of feedback is to amplify the positive and the more that this can be done the better.
Regular and specifically themed sessions enable managers to coach team members more effectively. They enable to process which becomes known and trusted by both manager and team member, not a yearly session that no-one looks forward to or enjoys.
Regular, short term goals which are set by both the manager and team member create buy-in. Regular reviews and visible progress help motivate and encourage team members to grow and improve.
Having each team member propose a feedback score every month enabled far richer and more valuable discussions around their performance than the previous process whereby only the manager told them how they thought they had performed. Many companies, including Atlassian, have taken this further and removed rigid and strict performance scores. Having to work this process within a yearly process where a score was required meant that this is not an option for us. Yet.
Encouraging team members to prepare for sessions and document their thoughts in one place not only helps make the sessions short, but it also builds up a record of their achievements through the year.
Keep iterating on the process and incrementally improving it. We’re not done with improving performance reviews – the next step is to look at how company objectives can be built into the process, taking the Google OKR concept. The intention is to then be able to use the process not only to help each team member improve, but also to promote consistent and challenging goals across the whole team.
Don’t get hung up about tools to support your process. We’ve run this as an MVP for almost one year now and it’s been very valuable. We use a Google Sheet we designed ourselves to document feedback, objectives and thoughts.
If you want to change something then do it. Figure out how to work within whatever processes you are required to do so, but don’t let that stop you implementing something new
Further Reading
If you want to learn more about introducing a more regular and valuable performance feedback and review process then the links below are worth reading:
tl;dr No-one really likes performance reviews. But giving your team members good, actionable feedback and setting clear direction with them is very important. In my team we changed how we did performance reviews.
Note: This is the third post in a series on performance management and review. It’s worth reading the other two first:
In part two of the series I talked about how my team and I took the performance review and feedback system that is used at Atlassian and adapted it to suit our groups context. I talked about the themed ‘checkin’ sessions and how we got the whole group working to a common timeline for review and feedback, and the advantages that brought.
We what happened? How were the changes received and what was the feedback from the teams?
Selling the Idea
Once we were clear on our direction, we had agreed as a leadership team how we wanted the process to work and what themes we were going to use, then we prepared to sell our idea to the rest of the team. I strongly believe that one should not dictate changes that affect people’s career development or relationship with their manager, and so it was critical to us that the team members were bought into the ideas of the changes and our reasoning for proposing them.
Feedback is better when it’s timely and so we want to start a process where that timely feedback enables higher quality discussions about you
As a group we agreed to run the process for one cycle and then review the results. If, after that, the teams were happy to continue then we would continue and if not then we would stop and revert back to the company-wide yearly cycle.
Kicking It Off
The issue with selling an idea to a group is that often, in private, people will express their reservations more freely. So it was key to ensure that managers then spent time with their team members and gave them the opportunity on a one-to-one basis to discuss the new process. This was also a good opportunity for managers to explain the process in more detail and seek additional buy-in.
The checkin sessions themselves are intended to be 30 minutes long in order to ensure that the process does not take too much time every month. It ensures that only the valuable things are discussed, and the sessions can be to the point and targeted at what matters. However, since this was a new process then we ensured that the first few checkin sessions ran to an hour so everyone could get the hang of it.
Writing It Down
Since checkin sessions were short and punchy – although not literally 🙂 – then it was critical that everyone prepared for them. The checkin’s process works if the team member comes to the session having already prepared what they want to talk about. In order for them to be able to do that then we needed somewhere that they could document their thoughts.
Being an MVP then we did not want to spend much on supporting tooling (there is an excellent tool called Small Improvements which supports this sort of process) so instead we designed our own form. It’s simple and includes questions that the team member should answer in order to prepare for the session. The idea is that they fill in the relevant tab before each session, the manager can then review and prepare based upon the information they’ve entered, and then the checkin session itself can be focused on the discussion and actions from that discussion, rather than the actual ‘thinking’ time. It allows sessions to be targeted and valuable.
We designed our own form to support the process. If you’d like to see a copy then get in touch. We used the questions below to drive discussions:
Another advantage of using the form was that each team member built up a record of their year, what they had done, and how they had performed. This meant that when we did need to provide yearly feedback into the main company process then it was simply a matter of extracting the information we already had in each form. It was also great to encourage team members to look back through their forms to see what they had achieved throughout the year.
Not Just Qualitative Feedback
By meeting with our team members every month for themed, targeted performance and career management discussions then it also meant that we had a great opportunity to give them quantitive feedback. Like a lot of companies, we use a four point scale and at the end of each year every employee receives a performance review score which has an impact on salary and promotion – standard stuff in most companies. Once per year with no indication in-between how they are performing relative to that four point scale.
This seemed unfair so we adapted the checkins process to also include giving the team member feedback each more on how they had performed against the scale and why. This helped solve multiple complaints against a yearly system:
One key aspect of feedback is to amplify the positive and managers could tell their team members about the good things they were doing.
Since the manager and team member were setting goals together as part of the checkins process then we wanted managers to act more like coaches. Having the team member measure themselves them this helped to amplify ownership of their goals and performance.
Team members got timely feedback that meant that they could adapt quickly if they weren’t on track.
Managers could give timely, targeted feedback with recent examples, in order to drive meaningful improvement. No more having to keep a years worth of notes or trying to remember what the whole team were doing a year ago.
However, merely getting a feedback score from the manager still presents a surprise and it was important that the manager was able to have a meaningful two-way discussion about performance. In order to drive this we asked each team member to give themselves a score, on the same scale, before the session. This enabled the manager to then understand how the team member felt they had performed, and it encouraged the team member to think about their achievements during that month. The discussion, held towards the end of the checkin session, could then be focused on any differences between the manager and team members interpretations of their performance.
Running The First Rounds
The first rounds went smoothly. One thing we learnt pretty quickly was that it took time to adapt to the process and so running the first couple of checkins for each team member as hour long rather than 30 minutes was definitely necessary. There was also some prompting required of some team members in order to get the forms filled in prior to the checkin session rather than during the session but this was to be expected given that the process was different and new.
People were surprisingly happy to provide a feedback score for themselves and by asking them to do so, we were able to make more meaningful and example based discussions on how they had performed.
Feedback
As we had promised, we ran the process for four checkins, i.e. one cycle and then sought feedback from the team. Broadly speaking it looked like this:
We Liked
Focused sessions are clearer beforehand
Keeping track of things and getting constant feedback
It’s easier to remember what’s happened in a month
Makes the annual review much easier
More productive
Constructive, thought provoking guidance
We’d Like To Change
Some sessions are too similar or too regular
Some sessions are too close together
The process is too time consuming so could it be shortened?
No 360 feedback
There’s nowhere to document the managers feedback
And So…
We now had some feedback to work with. Overall the experiment had been a success and was worth continuing with, albeit with some small changes. Overall we had:
Designed and rolled out a process which enabled timely, targeted, example based feedback to each team member
Enabled each team member to take responsibility for their own development and think about it each month, before the checkin session
Set clear, regular, expectations on people’s performance throughout the year
Made it far easier to deliver what was required by the companies yearly performance management process.
In the last article in this series I’ll explain what we did in order to improve the process to suit our context even better, and some of the key learnings we got from running the process over a longer time period.
tl;dr No-one really likes performance reviews. But giving your team members good, actionable feedback and setting clear direction with them is very important. In my team we changed how we did performance reviews.
Note: This is the second post in a series on performance management and review. If you haven’t read the first one on Improving Performance Reviews then I recommend you take a look at that first before reading on.
What Was Wrong?
As I mentioned in the previous article, a cycle with yearly performance reviews doesn’t work. Objectives become irrelevant or forgotten, the business and team landscape changes, and people simply don’t get comfortable being a part of, or good at, something that only happens once or twice a year. As a result the traditional performance reviews and objective setting cycles that a lot of companies use end up becoming de-motivating for both team members and managers and are just something “to get out of the way” each year before moving onto “real work”.
It doesn’t have to be like this.
So Why Not Change It?
Both me and my management team were becoming frustrated by being part of a performance management process with yearly performance reviews that we felt was not enabling us to get the best out of our teams. We wanted to change things. But, as part of a large company with pre-existing people processes, we did not have the mandate, nor the time, to initiate company wide change. So what we did was driven by a need to make both a positive change for our teams and also to still fit within the existing yearly cycle that the company uses. An attempt at getting the best of both worlds. And to fly a little under the radar while we proved the concept.
Let’s not forget, as Grace Hopper once said – “It’s easier to ask forgiveness than it is to get permission”.
So we made some incremental changes which we feel has had a positive effect on our team’ member’s career development, their motivation and our motivation too.
So What Did We Do?
When it comes to anything involving people then it pays to seek advice from as many different people and sources as possible. We looked at how a number of different companies were doing their performance management, companies ranging from those that did no formal sessions, to those that did a lot. There was one that really stood out – Atlassian. What Atlassian had done with performance management seemed like something that would fit with our thoughts on the types of changes we wanted to make.
I won’t explain why the Atlassian model is good in too much detail, because they have done a great job of that themselves. I suggest you read their posts which explain much more.
In short, what they had done was take the traditional model, split it into bite sized chunks, distribute these throughout the year and enable timely performance feedback. Exactly what we were looking for.
Now, it’s very easy to read blog post after blog post about other companies cultures and how awesome they are. I’ve seen it argued that a key measure that investors use when deciding which start-ups to invest in is the company culture, and so companies blog copiously about their perfect culture. Even if they aren’t exactly the whole truth. It’s always wise to treat these article with a pinch of salt and find somewhere else that has actually adopted similar changes. Fortunately we found that REED, a large UK based recruitment site had taken the Atlassian model and were using it day-to-day.
So we contacted REED and spent some time with their technology leadership team discussing how they had implemented the Atlassian model and what their experiences were. After those discussions we felt confident that the Atlassian model, while it would require some adaptations to suit our context, was usable, workable and something worth trying.
Start Small And Iterate
Given that we needed to work within the company performance management cycle and policy then we were able to take some parts of the Atlassian model but not all. The model overall splits into four different areas:
Rip apart the traditional performance review – and redistribute the good bits into more frequent sessions
Stop paying individual performance bonuses
Create bite sized chunks (of feedback and career review)
Performance is still evaluated – but no exact rating is given and a two axis feedback system is used to promote a more coaching based set of conversations.
Unfortunately we had no control over the payment of bonuses, and the performance evaluations (a.k.a a yearly performance ‘score’) needed to be fed back every year. Changing this just for my group was not possible and even if we had been able to do so, it would have presented a very confusing view to the team members who are also part of a wider technology team not using our new process. There was also the danger that if we started to evaluate our teams differently then that could have impacted how they were evaluated in comparison with their peers outside of the team, with potentially negative consequences. So we focused on implementing a system with bite sized chunks of feedback and review, together with ‘scoring’ using the same four level system that was implemented across the company.
MVP
The Atlassian model uses six different themed sessions called Checkins to drive performance and career development conversations. These are held monthly, targeted at taking about 30 minutes each and the discussions are led by the team member. Everyone has the same checkin at some point within the same month.
For our initial MVP of the process we decided to use four of the themes:
Focus Areas
Love and Loathe
Removing Barriers
Career Long Term
It was very important to explain the different checkin themes to the team, set expectations on what the team member needed to do to prepare and what the intended outcome was of each session. That way we could explain the value that could be gained from each session. We prepared cards that each team member could use to understand more.
Focus Areas
Love and Loath
Removing Barriers
Career Long Term
One Common Timeline
It was also important to ensure that it was clear to everyone when discussions needed to take place. By ensuring that everyone in the group (managers included) had the same themed checkin discussion each month then, as a leadership team, we were able to meet at the end of the month and identify any common themes that had arisen from discussions. For example, was there something that was blocking a lot of people? Could we allocate some budget or additional focus to doing more team building, changing a process, buying more equipment, etc that would benefit the team as a whole? Were there some stakeholders that multiple team members were having issues with connecting with? What was making the team members happy and therefore we should be doing more of?
By having a common dataset at the end of each month we were able to have far more meaningful and valuable conversations about change than we had been able to have before.
We used the timeline below to ensure a common focus across the team.
So What Happened?
In the next article in this series I’ll explain what happened. How did we roll out the new process to the team? How did we continue to evaluate performance in-line with our new monthly process and the companies expectations, and how did we ensure a common approach to preparing and recording the output of each of the checkin discussions? How was the new process received by the team?
tl;dr No-one really likes performance reviews. But giving your team members good, actionable feedback and setting clear direction with them is very important. Yearly performance reviews don’t encourage this to happen.
Note: This is the first part in a multipart series of posts on performance management. In my team I changed how we do performance management. Future posts explain what changed and what happened as a result. This post sets the scene.
So What’s Wrong?
It’s fair to say that the traditional way of doing performance reviews is exactly that, traditional. Many companies work on a yearly cycle, with managers setting their people some objectives at the beginning of each year then reviewing them at the end. It’s likely that it’s been that way for ever. Sometimes if you are lucky then they’ll also be a mid year review.
I have a problem with this way of managing performance. We try to teach our teams to think in the present, to inspect and adapt their ways of working and the feature sets of products themselves, regularly and iteratively. Yet when it comes to the people we do the opposite. It doesn’t make sense.
The Problems With Performance Reviews
I’ve seen the following problems with the yearly review cycle in a number of companies that I’ve worked in. The same concerns and feedback every time.
The Forgotten Objectives Problem
Firstly there’s a huge gap between when someone is set some objectives and when those objectives are reviewed. Objectives get forgotten. A person’s goals and motivations change. The business landscape changes. You get to the end of the year, review the objectives and discover that none of the objectives that a person has are relevant anymore. The logical action when faced with this situation is to just ignore the objectives and review someone based upon feedback and your own personal viewpoint. Your team member gets the same performance score they did last year and then the whole cycle starts all over again.
In this situation you should think about the message that this is sending to your team member. You’re essentially saying “objectives aren’t really that important”. Don’t be surprised when you don’t get the buy-in next time objective setting comes around.
The Disruption Problem
When companies are tied to yearly performance review cycles then early December and late January are the periods that managers hate. Believe me, it’s where you really get to be the bottleneck. There’ll be expectations from HR that you do all your reviews by a certain deadline. It’ll be the same deadline as all the other managers have. Everyone will be running around trying to do the same things. Free meeting rooms will be pounced on by the first person who sees free space in the calendar. Everywhere will be full of people in groups of two talking quietly.
Preparing for a performance review is extremely important and a manager can’t just rock up to one without having prepared. There’s a lot to be done. Feedback to be requested, notes to be collated, memory banks to be queried to remember what happened almost twelve month’s ago, feedback to be re-requested, notes to write, the review itself and subsequent follow-up. That’s just for one team member.
A yearly review cycle effectively takes the entire management population out of delivery work for two months a year. It creates bottlenecks. It’s one big performance management big bang. That’s not how we deliver software anymore.
It’s also disruptive for the team members. Suddenly everyone is being asked to provide feedback on everyone else, everyone is preparing for their own reviews at the same time and because those reviews only take place once a year then they take a long time to complete.
Feedback Is Better When It’s Timely
There is nothing more annoying than being told “well that thing you did 6 months ago – that wasn’t good”. You have no chance to change the situation and little chance to effectively learn from the feedback. The time will have past and the opportunity to change may well have gone as you’ve moved onto something new.
Yearly performance review cycles don’t of course preclude managers from giving regular feedback to their team members. But they do make it easier not to give that feedback, and not to seek it from others. They subtly encourage us to think that ‘proper’ feedback is only needed to be given once a year. That’s wrong.
The Infrequent Problem
When you do something regularly then it becomes easier. You learn how to do it, your brain becomes wired to do certain parts of the activity almost without thinking, and you lose your fear of something new. Mastering a new skill and gaining the confidence to do it well takes time and practice.
We don’t get to practice the art of performance reviews regularly in a yearly cycle. At best a manager may well have done about ten reviews a year for a couple of years at a company. Those reviews come round only once a year, by which time a lot of the skills are forgotten or take time to become natural again. You make mistakes, those mistakes de-motivate your team members and just as you feel like you’ve got your review mojo back, then it’s time to put it away again for another year.
This also applies to your team members. They get reviewed once per year. In fact it’s worse – they get fewer opportunities to get good at reviews than you do.
It’s human nature to fear change and to feel anxious and apprehensive about something that is new. A yearly performance cycle becomes something that neither manager nor team member looks forward to, precisely because it happens so infrequently that no-one get’s the chance to feel confident doing it.
The “I’ve Forgotten What I Did” Problem
When reviews are yearly then it you can find yourself trying to remember what both you and your team members were doing almost twelve months ago. Even with the best notes in the world, and the most fastidious team members, there will be points that are forgotten.
Feedback from other team members and stakeholders won’t focus on what was done in the past – when under pressure to provide feedback for multi people (which happens more under a yearly cycle) then people remember the recent past far better than something that happened months ago. And if they do remember something from months ago then it was probably an event they perceived as negative. Bad memories stick better than good ones. The feedback you get will come with an unintentional negative bias. Or it may not even be true at all.
There’s a Long Lead Time
I’m a strong believer is delivering software iteratively, seeking regular and timely feedback, and adapting as a result. In my career I’ve done this enough times to feel that this is the least risky way to deliver. We learn as we go along, and we change approach, feature set, timescales, etc when required.
One key metric I like for a team is lead time. How long does it take us, from the point at which we receive a request, to the point at which we deliver it? Enabling a suitably short lead time to enable the team to be nimble and adaptive is key to delivering valuable software. A yearly performance management cycle has a lead time of exactly that, one year. The opportunity to adapt is tiny.
So What Should We Do About It?
I believe the traditional yearly performance review cycle does not work. It does not enable timely feedback, does not exploit our desire to work in the present, it’s time consuming and it’s demotivating for both team members and managers.
In my team we decided to do things differently. We changed how we do performance feedback and performance reviews. The next article in this series will tell you what we did, how we did it and what happened.
I really loved being a part of WeTest 2016. I gave the opening keynote at both conferences, as well as a couple of other talks at sponsor events. If you came to any of them then this post contains some useful links you may want to read. It was great to meet so many engaged and knowledgeable testers in New Zealand.
The Opening
I was thrilled when I was offered the opportunity to travel over to New Zealand and talk at the WeTest 2016 conferences in Auckland and Wellington. I mentioned my conference speaking goal as part of my presentation at both events – this was that one day someone would invite me to speak in New Zealand. And Katrina Clokie, on behalf of WeTest did just that earlier in the year. It didn’t take long for me to agree to be a part of the conference and being offered the opening keynote was a real privilege.
Now there is one thing that’s clear and that is that it is a long way to New Zealand (28 hours to be precise from the UK). So I wanted to ensure that I made the most out of my short time there and to give a presentation that fitted in with the conference theme of “Influence and Inspire”. Given that I have made a move out of a pure test role over the last few years then this seemed like a logical theme for the presentation, and an opportunity for me to give my views, gained from outside of testing, on how testing is perceived. I also wanted to show the audiences that there’s important leadership roles within testing to be taken and made the most of.
Preparation and Arrival For WeTest
I can say without a doubt that WeTest was the best organised conference that I have spoken at and it’s testament to the effort put in by Katrina, Aaron, Shirley and Dan. All too often presenters at conferences aren’t treated brilliantly by organisers – either by having a lack of support leading up to the event, being expected to pay their own way or by needing to foot the bill for travel and accommodation then claim back. WeTest was different – they organised and paid for all travel up front and sent detailed itineraries for each speaker, plans on how the conference rooms were to be setup, as well as ensuring that the simple things like a taxi to pick you up off a long haul flight was pre-arranged. It made the whole experience easy and it made us speakers feel valued. Other conferences should take note.
Up and Away?
My presentation focused on my views of testing from outside of just testing and how I made a move into software management and why. It was part of a clearly well thought out programme and the other presentations, whether focused on more technical aspects such as mobile testing, or other leadership presentation, fitted together really well. You can see the programme here to get a feel.
Highlights
I thought all the presentations were good but a particular highlight for me was Adam Howard’s “Exploratory Testing Live” where Adam took the really brave step of doing live exploratory testing on the Trade Me website in front of a conference audience. Fair to say it went a bit better in Wellington than Auckland, where he found a bug early on that blocked a lot of the rest of the session but he coped really well and there was value in both sessions. We don’t see enough live testing at testing conferences, it’s common to get live coding at development ones but hasn’t seemed to catch on. Based upon Adam’s session then it should.
Also I really enjoyed Joshua Raine’s really personal story about “Conservation of Spoons” which he did as a noslides presentation. Deeply personal at times and spell binding. I won’t give you the details because I’m sure he’ll do this one again and to know the story would spoil it.
A New Approach to Q&A
As part of my presentation I thought I’d try out the new Q&A feature in Google Slides. It turned out that this was a good move because it enabled me to get questions from the audience as I talked and also to keep a record of all of them for later. If you speak at conferences I’d really recommend it.
For those of you who attended then here’s the questions asked, my responses, and a little more information.
What do you most miss about your testing focused roles?
I guess I miss being the expert with the safety of many years of experience of testing. I also make sure I keep connected to the testing community because it’s great and if I didn’t then that would be the biggest thing that I’d miss.
It’s hard to find a balance between adapting and burning out when trying keep up with new trends and technologies. Do you have suggestions on how to manage that balance?
I make sure I do one thing well; that together with Jerry Weinberg’s Lump Law – “If we want to learn anything, we mustn’t try to learn everything” help me to maintain a balance. Trying to find the one thing to focus on is sometimes hard, but by discussing with your stakeholders and your team then you can usually come to find out what is the most important.
Do you have any resources/suggestions for learning to speak ‘Dev’?
There’s not one clear answer to this – it depends on your context, languages and architecture that you are dealing with. When I started then I tended to use google a lot and base my searches on discussions I’d had with the dev’s in my team. I’d also ask them what they thought I should be learning and why. I’ve also found sites like Hacker News and Stack Overflow to be great for keeping up with what’s happening. Talking Code also comes recommended.
Was your career path in the same company or different? Would you suggest to change companies or progress in one?
It was in the same company and I think it’s much easier to make a change from one path to another when you are in the same company. You have that reputation to use and you are a known quantity. It’s also likely that you’d been working towards such a change as part of your personal development anyway so the company will be confident in your plans.
Since we are talking about change, at any given point during those career changes did you resist the change? How did you over come that? Were you specifically looking for change?
I talked about how it can be scary to make changes but trying something new for 6 months is key. You need to understand that you will go through a change curve just like anyone else and things will feel extremely uncertain as a result at the beginning. I knew what to expect since I’ve been through other changes.
In this case I wasn’t specifically looking for change, more for an opportunity.
Thank you for the nice talk If you had to recommend doing ONE PRACTICAL exercise to raise awareness with the team that WE ALL own quality – what would it be?
I like the idea of whole team testing and bug bashes are a great way of starting.
What do you think testing will be like in 10 years time?
I think we’ll see less traditional testing roles and the focus on automated checking will become more of a development activity. This may mean that there are less testing roles in total but good, exploratory, coaching testers will always have a place in teams. I think the biggest change will affect Test Managers, with far less of a need for them and a transition to coaching roles.
How do you avoid being typecast as just being a tester and overlooked for other career opportunities?
Show that you have an interest to others in your company. Work with your manager to map out your path to a new role and start to show that you are learning the skills that will be required in order to be successful at it.
In short – you own the responsibility for your own career.
A lot of people in this company are transitioning from Waterfall to Agile. Some of the people in the room might gain from you talking to how you see that transition working from the experience you bring, especially if they’re concerned about career development and management.
It does depend a lot on how the management structure adapts. In agile transitions that I have been a part of we’ve changed to autonomous, cross-functional teams with single managers, who have been supported by coaches for both agile, testing and development. This support network is key, as is the establishment of discipline based communities. A strong testing community for example, allows testers whose roles are changing to adapt.
Support from coaches allows not only testers to be supported but also means that someone with testing experience becomes responsible for establishing career paths, competency expectations, etc that can then be used by other managers who are not experienced in managing testers. It also helps maintain a ‘voice of testing’ towards senior management and to enable activities that help testers grow within an organisation.
And Repeat…
The first conference was in Auckland and we repeated the experience again in Wellington three days afterwards. This meant that the whole WeTest circus upped sticks and set off for Wellington for a repeat performance. I also did an “Understand Your Mobile Users” presentation at a sponsor the night before, plus the same WeTest presentation at a sponsors internal conference the day afterwards. Four presentations in one week was something new to me and actually pretty interesting to do as I got into the speed of things. It almost felt like being in a band on tour 🙂
Closing Thoughts
WeTest was great. I met some great people and learnt some new things. It was extremely well run by a passionate bunch of volunteers who knew how to treat their speakers well and how to organise a great programme for the conferences. The testing community in New Zealand is really engaged and it’s clear that there’s some really forward looking testers pushing the boundaries here. If you get yourself over to New Zealand then I’d certainly recommend it. It’s not that far. Really 🙂
I’ve been reading a lot about test management recently. There’s some excellent posts out there, in particular I’d recommend you look at this one from Katrina Clokie, explaining the changes that are required for test management to remain relevant in the world of Agile software development and continuous delivery.
She also links some other articles which I would definitely recommend you read, if you are interested in the subject.
So why am I interested? Well for starter I’m a Test Manager. Over the last couple of years I have seen my role change, from one of leading a separate, large test department, to one of managing testers across a number of project teams. It’s about to change again.
I’ve seen the challenges being a Test Manager in an Agile environment brings, in particular the difficulty in remaining relevant in the eyes of product and development managers, and the challenges of understanding enough about multiple areas in order to be able to support your team members. Being a Test Manager in a Agile environment can be isolating at times, particularly when the department is big, and the number of agile teams is large. It requires an ability to balance a lot of information, priorities, and tasks, across a number of areas. Stakeholder management and influence become key. Context switching comes as standard. Often it’s not much fun.
Through discussions with others, and looking at my own situation, I’m increasingly coming to the conclusion that the new ‘Agile Test Manager’ positions that Test Managers are moving/ falling into just don’t fit with the ways that teams want to work anymore. The team is more important than the manager, and, for example, choosing to keep discipline based management because it means testers are managed by testing ‘experts’ isn’t enough to justify it. Managers are not able to effectively support their people if they do not have the time and energy to keep fully in the loop with the team. As Test Managers get split across multiple teams, (primarily because having one Test Manager per Agile team is massive waste), then it becomes nearly impossible.
Continuous Delivery
Moving to Continuous Delivery complicates matters further. Giving a team complete autonomy to design, build and release it’s own code is an extremely motivating way of working. Do Test Managers fit in with this ? I’m not sure they do. Where independence and autonomy are key, management from someone from outside of the team just doesn’t fit, particularly when that management is only part-time.
Change Is Coming
So how do we change? Do Test Managers merely become people managers, desperately trying to understand what their people, spread across multiple teams, are up to? Are they there to help manage testing but not people?? What about the coaching and mentoring, the sharing of knowledge and expertise, and the personal development of testers?
As I see it I think we’re going to see a lot more of this sort of setup:
Engineering Managers, who line manage an entire team. They understand the people best because they work with them day-to-day. No need for handovers, no need for performance feedback requests to other managers at review time, and no need to waste time and effort with coordination. Engineering Managers manage the whole delivery process and people involved. They may have come from a background of expertise is a particular discipline, but now they need to be able to represent all. But crucially they are focused on the management of a team who own a particular product or component and so share a single focus with their team.
Test Project Managers, who manage larger testing projects/ programme’s and dedicated testing phases such as UAT and customer acceptance. No people to manage, just deliverables. This role is very dependant on the nature of the software/ hardware solution being delivered. It’s most likely not needed in a lot of companies.
Test Coaches, who help organisations deliver the optimum testing possible. This means through coaching, mentoring, advising and working with engineering managers and whole teams in order to help them optimise their testing effort. Similar to James Bach’s idea of Test Jumpers, but with more focus on providing advice, guidance and strategy. In smaller companies they are much more likely to be exactly like the idea of Test Jumpers. Call them Test Jumpers, Test Managers, Heads of Testing or whatever, but the key point is that they are test experts who have the mandate to support testers in multiple teams but do not manage them. They can assist with recruitment and personal development if required, but are not a particular person’s official manager, and may get involved more with recruitment and personal development process, rather than people.
What Next?
The dedicated Test Manager, who manages testers and testing is not a role I can see continuing for too much longer. It is a hangover from the past, when large, dedicated test teams needed management, and it simply does not fit with how a lot of teams work anymore.
But, and this is a big but, I work in web, web services and mobile. I’ve seen the push for Agile and the push for Continuous Delivery because it fits the nature of the projects and technology used in these areas. Team’s are lean and projects are short. Almost certainly this makes me biased.
I would be interested to know what you think. Do you think the traditional Test Management role is reaching the end of the road? Or is it alive and well, and relevant in the area that you work? Why not leave a comment below and get the conversation started.
I firmly believe that in order to be as effective as possible, testers need to engage with the software testing community. Learning from others, particularly outside of the companies where we work, makes us more rounded and better informed individuals. It enables us to inspire ourselves and our colleagues in ways that we could not otherwise.
Recently I’ve been wondering why more people do not engage with the community. What is stopping them, and how can we all help change this? We can explain how brilliant the wider community is, and we can give examples from our experience. We can send people to conferences and email round blog posts. What is that does not work?
What do we do about those within a team who do not want to interact? Those who do not see it as a good use of their time, and are not willing to spend time on community matters, even if that time is given to them by the company. Should we incentivise people to do so? At least in order to push them in the direction of the wider testing community, where hopefully they will get hooked? Or should we do the opposite? Is it a valid idea to make community engagement a part of people’s role description, and therefore penalise those who hold such positions and do not exhibit such engagement?
Or is there another way of persuading everyone that the software testing community is key to their personal development? I’d be interested to know what you think.
A typical testers career path can look like this – it’s rather rigid and set. There are similar paths to this for management roles, and often testers are pushed towards management because it pays better, even if they don’t like management.
But what if we forgot about roles and job descriptions?
But what if we forgot about roles and job descriptions? After all, why do we have to consider our roles as part of a career ladder? How do we fit within our roles? What makes us satisfied as testers? What satisfies those who work for us in testing?
It’s all about considering building blocks. How do we fit the pieces together? How do we build a different kind of model to help us understand ourselves and our testers? Well I think Maslow’s work can help us out here. It can give us a framework.
A Testers Hierarchy of Needs
Our first level – Maslow would call this the Physiological level. I call it acceptance. Acceptance of testing. So testing is just a job. Management view is ‘anyone can do it’. You are probably complying with work rules and basic processes to continue staying in job. There’s a lack of respect or support BUT there is testing happening. It’s probably manual checking only but it’s a start. So a positive step.
The next level is Maslow’s Safety level. Testing is mandated. There is a test team. Testers have peers. There is some safety – the job is seen as worthwhile and so unlikely to suddenly be moved to somewhere where people can press the buttons and check more cheaply. But testers are still universally seen as an annoyance. It’s about meeting expectations of peers and bosses, sometimes unspoken, to meet a moral and social code. It’s a learning level. As a tester you can learn testing, the company you work for is learning about testing. You play a key part in ensuring that learning happens.
One level up and it’s all about belonging. Testing, testers and the test team is seen as a key part of the company. It is taken seriously by all. It’s about meeting and exceeding more demanding expectations of peers and bosses, sometimes unspoken, to meet moral and social code. Consultation. Respect. Inclusion.
The Interaction level is next. The impact of testing reaches further than just the IT team. Consultation from the business. Reliance. Interaction. And overall a real sense of Esteem for the tester. A sense of worth. You are really adding value.
Then the top. Maslow’s self-actualisation level. Test Mastery.
• The skills and continual learning needed to succeed.
• Coaching others up the hierarchy.
• A responsibility to the wider testing community.
• The person must not only achieve the previous needs, but master them.
Recognition from all. Inside a company and outside. Respect and recognition from peers, thought leaders, and those that you really respect and value.
What affects your movement between levels?
Maslow never drew the theory of needs as a pyramid. In fact he never drew it as a diagram at all as far as I know. That was done in a later psychology text book. He felt that there could be overlap between states. You can move between states. Down as well as up. Changing roles, changing teams, life changes, can all affect movement between levels. Where you are on the hierarchy is just a snapshot in time.
I feel that, for a testers theory of needs, a pyramid is right. It shows the hierarchy best. But it could be a set of overlapping balloons, if you like that sort of thing.
So….where do you think you are on the pyramid?
Are you on a path to self-actualization. To Test Mastery? Do you think you are there already? What are you doing to stay there? What about the testers who are in your team? What about the company you work for?
TestBash 2.0 is less than 3 weeks away. I’ll be speaking on “A Testers Hierarchy of Needs”.
What motivates you to work and improve as a tester? Why do the testers in your team work well together? Or why don’t they? Have you ever wondered what motivates professional testers?
Maslow’s Theory of Needs seeks to describe the psychology of humans by way of a hierarchical model. From physiological needs up to self-actualization, it helps explain what motivates us and how we express that motivation. A Testers Hierarchy of Needs does the same for software testers.
If you manage testers and have a keen interest in ensuring that your team are motivated and work well together, or if you are a tester wondering what is needed to make your team great, then come along and discover more. I will present a framework which helps explain tester psychology, how internal and external factors can affect their motivations, and steps that can be taken to better motivate yourself and your team.
It would be great to see you there.
Keynote speaking, technical leadership, coaching, training and presenting