Proper Villains - Work Style
“High stakes? Okay. I’ll play. But it ain’t your game, and it ain’t your rules, and it ain’t your world — and I brought my crew.”
There’s no one-size-fits-all way to create team culture. Because teams are made up of individuals, any configuration of folks is going to present unique challenges. Adding or subtracting people will threaten existing culture. For this reason, if I’m trying to instill a generative culture on a team, I focus on four things:
Work Style — Find the power of We Don’t Care About That Shit
Staffing — Recruit for fit, not just horsepower
Tooling — What do we have to spend time and money on to succeed?
Identity — Bonding and picking each other up when it gets hard
I’ll have a separate post for each section, but will include links above as they are finished so you can navigate between sections easily. For today, we’ll focus on…
Work Style
Despite all the factory metaphors we use for development work, putting together a team isn’t building equipment. You can’t use the blueprint from the last team to build the new one, even if they are in the same company using the same technology stack. When you’re putting together a team, you need to think about it like product design — what is the outcome I am looking for? What are the nonfunctional requirements/constraints imposed on me by my org or industry? What is the minimum version I need to start testing my hypotheses about how this team will deliver?
I call this category “Work Style” because I believe the hardest part comes down to something squishy that is mostly captured by that phrase — put a little longer, this section is about how management designs the mission, agreements, and assumptions a team will be built upon. It’s the table stakes that need to be explicit, because they are the most likely things to create culture mismatches.
There are non-negotiables depending on company and business model, the afore-mentioned constraints and nonfunctional requirements. Does this team need to use Scrum because of company mandate? Do you need an SRE group that has consistent, well-defined on-call schedules? Are you a professional services group that bills on time and material?
Given the constraints you cannot fight, your job as a manager is to find a sustainable path toward the outcomes you need from this team. This sounds extremely formal, and I guess you could get real nuanced — lay out every possible positive feature of a team and stack rank them based on impact on your outcomes. For me, that’s usually unnecessary. Of course, given the opportunity, we all want teams that
Release often
Have predictable velocity
Have few bugs
Communicate well
Have ample time to focus via flow time
and all the rest of general wisdom about “good dev teams".
But it’s more important, in my mind, to figure out what pieces of an ideal work style you don’t care about for this team. I have had a lot of trouble learning to accept this — am I really okay saying the culture of my team, for instance, is based on not limiting context switching? Do I really want to institutionalize the idea that we’ll code bugfixes faster than a majority of users will see them? Am I actually cool with developers building tools to circumvent existing release processes?
It took coaching professional ultimate frisbee (www.strikeultimate.com) with defensive coordinator Dooby Helm to give me the language to get really comfortable with it. Without getting too deep into frisbee, when concerned players asked Dooby why a certain defensive strategy didn’t try to take away something usually seen as a threat, Dooby somewhat indignantly said:
We don’t care about that shit! The whole point is making them work with that, because we’re deciding what we’re attacking. They’re gonna take [that action] and we’re gonna get blocks because we know what they’re about to do next.
As a manager, be like Dooby. Decide what you’re attacking. Said another way, consciously decide what weaknesses you’re comfortable with. Design your work style around finding an advantage through strategically killing your darlings. Start from the desired team outcome instead of hefting the baggage of every trauma-driven assumption you have about what makes a good team. The next three articles in this series exist because doing so will piss off a lot of team members if you don’t hire, tool, and coach them for it.
But what does that even look like?
For an engineering example, one of my highest performing teams was designed as a ‘startup within a startup’. We needed a team that could deliver new features, often barely proofs of concept, to capture customers in a brand new market. The work to get those new clients may not even end up as long-term features, just enough to prove we could do something. We also needed to be responsive in a way that would not fly in the larger business — when you only have five clients, two of them facing a major bug is catastrophic. Two customers in the larger business would not even warrant a support ticket.
So the work style of this team had to be one that was intensely amenable to change. Change in priorities, change in mission, change in code. We needed engineers who weren’t precious about flow time. We needed the ability to make changes in production often, And we needed to be willing to throw away work when priorities changed or new information came up.
Ultimately, what we chose to attack was rigidity and consistency.
I know! I KNOW! From an agilist, that sounds idealistic — “we chose to attack consistency and gained agility!” Like most business blog posts, it sounds like it’s all highlight reel.
But there are downsides to this — moving fast, with sometimes sketchy requirements, means developers could (and did) build things we didn’t need, which led to a scramble. That sucks.
Spending two weeks hammering out a project with a bunch of heat from execs, only to find out a customer churned before it released? That sucks.
Having to justify every technical solution that skirts an existing process because we need it outside of regular release cadence? That sucks.
But listen to Dooby.
We don’t care about that shit!
We have chosen to not let those issues be the sorts of things that ruin our days, our weeks, our quarters. And that isn’t easy. That’s why this series is about establishing a culture, not how to implement a series of meetings or process flow diagrams or whatever.
“We don’t care about that shit” has to be in your team members’ bones. It has to be something (see: Identity) that you take perverse pride in getting burned by. If you name the pain you’re comfortable with up front, and get buy-in, teams will even brag about the failures you chose to allow, because it frees them up for the successes you really wanted.
For all that to work, though, you need the next three sections.
Just Deliver, Baby
Just Deliver, Baby
“When you have great coaches, and you have great players, and you have a great organization, you tell’m one thing– Just Win, Baby.”
In software, it is so easy to get bogged down in process and tooling and finding the perfect, best-optimized way to do something. Agile methodology is just as susceptible to navel-gazing as tech-stack choice and architecture. Worse, it affects HOW engineers get to work on the stuff they are having those other holy wars about.
So what we can do is drop the ego, assume we’re probably wrong about stuff, and chase the only outcome that matters: delivering software. The best way to work is the one that will work. It won’t be perfect. It will have contradictions. Work will settle un-done in the recesses of ambiguity. But if you’re being candid and honest, you’ll be able to respond to problems in a way that requires neither over-engineering a process nor ditching methodology altogether.
For the record, this is exactly the same mentality that led to the Agile Manifesto. It’s just the kind of thing that needs to be rediscovered, over and over, by groups of engineers once orthodoxy overtakes outcomes-thinking. Don’t be the voice of orthodoxy, be the voice of What Works.
That’s easier said than done– many of us bear the trauma of huge failures and frustrations from previous jobs. Many of us want perfect, clean systems that would account for every possible permutation if only humans didn’t have to interact with each other. Many of us think of process as a way for the man to get us down.
The secret to getting past it is, again, remembering the outcome: delivering software.
So, borrow an idea from Kanban. “Start where you are.” Don’t up-end your whole system. Along with the rest of your team, pick the pain point that is most getting in the way of your delivering software and construct an experiment to attack it.
If you don’t have a Where You Are to start from (say, you’re a new team), then you can use my preferred starting point.
How do you construct an experiment?
Experiments are the key to agility. No matter what tools you’re using and what process you adopt, you’re going to need to actually respond to stimulus and trauma. Let’s talk about the components of a well-constructed experiment, and then I’ll run through an example or two.
Define the problem. This is about what’s making it harder to deliver software. It could be “our release process is slow” or “the tests are so brittle we turn them off and then stuff breaks” or “we’re doing too much operational work” or anything, really, that is getting in the way of delivering software.
Define the preferred outcome. This is tightly coupled to the first item. What is the thing you want changed to try to solve the problem– a faster time from merge-to-deploy, or a lower percentage of failed tests, or fewer commented-out tests.
Define the metric. Again, coupled to the above – what does a ‘faster’ merge-to-deploy cycle mean? Days, hours, minutes? What percentage of tests fail and what is a reasonable expectation of what can be fixed? Can you dig into metrics like time-in-status, number of deployments, or anything concrete to help guide what the experiment should be?
Define the timebox. Knowing the outcome and metric, set an aggressive and realistic timeframe (usually no more than a couple weeks, but sometimes as much as a month) to test out the change. You may scale your metric based on what you reasonably believe you can get done in the timebox.
Define the attempted solution. Now that you know the details, you can propose possible solutions and try one. What if we skipped manual testing after product acceptance? What if we allowed teams to release on their own instead of waiting for a batch on a schedule? What if we spent Fridays doing a test-jam to get our coverage up?
Solutions can be technical or process-related, but the idea is that you take a small swing at making a pain-point better. If the experiment succeeds, you reassess and build on the success or move to the next pain point. Importantly: if the experiment fails, that is also a great learning. It’s also where the ‘ego-free’ idea comes in – if your favorite pet idea doesn’t move the needle, believe the data. It’s easy to indict the experiment when you get a result you don’t like– resist that instinct. Run more experiments. You may want to try a similar, but better-constructed, version later (probably not right away). This is good! Refining how you implement a new process before you commit to it ensures that the team understands the value of the change.
We all want to deliver software. We’re designing and adapting so that we can deliver software better.
An Example
A team I was working with was having an issue with ‘bursty’ velocity. Measured over a quarter, their delivery was roughly in line with what we would expect. But in any given week or two-week period (we didn’t do Sprints at this time, but tracked velocity on roughly that cadence), they’d sometimes deliver almost nothing, and other times deliver two or three times what their ‘average’ would come out to. This is not a problem in an absolute sense, but it made predictability nearly impossible and limited our ability to diagnose any other issues with the team’s flow and delivery.
Digging some metrics in Jira, we found that the ‘In Review’ status was a surprising bottleneck. Tickets were averaging around 9 business days in that status. For this team, that could mean no one was looking at the ticket, but it could also have meant that there was too much work that was failing review and ‘bouncing back’ to the initial engineer. Looking at comment histories (mostly spot-checking, not pulling detailed reports), both options seemed viable, but the feeling on the team was that people weren’t prioritizing code reviews as much as they could.
So we set a week long experiment–
Every time you break flow (for a meeting, to go to the bathroom, to eat lunch), you would come back and do a review. If we saw a reduction by a noticeable number of days, we’d consider it a success.
The end results were middling. We got the average number down to 7.5 days, which is a good percentage change, but the team felt that the experiment wasn’t sustainable– they would dread coming back to big PRs that would stop them from getting into good, deep flow time.
Notice that finding.
The team was dreading big PRs.
The next experiment we tried for a week?
Limit PRs to under 200 lines.
The result was a drop to less than 12 business hours in the ‘In Review’ status.
We made that limit a team norm. We wanted to enforce new process. The team treated it as a point of pride to drop that number until it was “around 100 lines.”
A ‘failed’ experiment led to one that actually worked, and it created an incentive to behavior that started chipping away at the ‘bursty’ nature of our work. We started thinking about shipping small pieces of work regularly.
Was this the ‘ideal’ size? Probably not.
Could a different experiment have led to a better outcome? Maybe!
But this one worked, accomplished something we wanted, and helped us deliver software better.
Many, many more experiments were necessary to get us to a ‘flow’ state, but this one solution set a pattern for the team really enjoying thinking about and experimenting with process.
That’s the essence of agility – owning your process, defining outcomes, and trying something. Especially if it’s a bold departure from what you would expect. Especially if the experiment goes against long-held assumptions. You’ll either validate what you believe or learn something new. Either way, it gives you the real information you need to improve and deliver.
How Bold?
When I say ‘bold departure from what you would expect,’ I mean it. Nothing is sacred, everything is up for experimentation– but you need to construct the experiment in a way that doesn’t screw over your coworkers or customers.
Want to require pair programming on every ticket?
Cool! What’s your measurement for ensuring delivery is still on track? Maybe you’re worried that some executive will catch wind of this and break it up out of a desire for ‘efficiency.’ The solution to that is simple:
Deliver software.
Honestly measure if the outcomes are better, and if challenged, provide that data. More likely, if you’re delivering software, you’ll give no one any reason to inspect the ‘how’ at all. And sometimes when they do? They’re looking to their highest performers to see what they can glean for the rest of the organization.
Do you want to try a No-Meeting-Thursday?
Do it! How will you measure success? How will you ensure responsiveness to emergent issues? What if someone on support complains that they can’t get you on “just a ten minute call on Thursday”?
Deliver software.
Show why that entire flow-time day improves your delivery. Make it more than just vibes. Demonstrate the through-line on ‘heads down on Thursday’ to ‘getting the last bugfix you needed out in less than a day.’
Want to eliminate story point estimates because refinement takes up too much time?
Okay! Do it! Just make sure folks who are relying on your estimates for forecasting are getting something that will satisfy what they need. Product, marketing, exec– they may have to be part of your experiment, too.
You will be shocked at how on-board other departments are if you include them in determining the outcome, defining the measurement and, you guessed it:
Delivering software.
A relentless pursuit of improvement in service to delivery will keep your team focused on the important parts of how they work. That pursuit will challenge your assumptions. It will likely piss you off. But when you come out the other side, your process will be custom-tailored to your team. And the things that start to pinch, the stuff that gets in your way? You’ll have the tools to get them out of the way too.
A “Where” to Start From
A “Where” To Start From
I would never assume to tell someone what their perfect agile structure will be. My entire philosophy on agile development is that teams own their process— and that you should start where you are.
All that said, sometimes you just need…something. If you want a heavyweight system, you can completely follow the Scrum guide. You can even use SAFe or LeSS or some other strangely-capitalized system.
If you do try to implement something with a name and a structure, do me a favor.
Implement all of it.
If you’re relying on the research and years of experience of someone trying to coach you, don’t dismiss whole pieces of it because they look like something you didn’t like before. Especially with Scrum, if you’re just picking and choosing pieces, and it fails, you will have no way of knowing if the problem was Scrum or a lack of followthrough. See Ron Jeffries’ excellent piece “We Tried Baseball and It Didn’t Work”.
For the rest of you— folks who are bought in on the way I’ve run teams, for instance—I’m going to lay out the basic structure I use when I am launching a team. It’s not perfect, but remember: it is a system designed to be changed by experimentation. This is a predictable starting point, not an installable-system.
Teams
A team consists of:
2-6 engineers (including SDETs/quality engineers)
A product owner
(Possibly) 1-2 designers
(Possibly) An agile coach
The team is the atomic unit of delivery-responsibility. Heroics may be called out occasionally, but the responsibility for delivery falls on the team as a whole. Notice that SDETs/QA have engineering responsibilities and vice-versa. The whole team is responsible for validation and automation. SDETs are subject matter experts the same way a Front-End developer may be an SME for React or Angular.
A team has a Charter, which consists of:
Values - The beliefs the team holds to. Faced with pressure, and in the absence of any other guidance, team members are expected to follow these values when making decisions.
Example: “We are educators-- it is as important that we can make others understand concepts as it is that we can implement them. We have to be constantly learning in order to better serve our teams. To this end, we do not do work where we cannot learn or improve experience.”
Norms - Work-specific items. Some are generalized, but most will be generated by experiments that get promoted to Norms.
Example: “When rolling out large changes, we bring the items to tech sync, dev jam, or an ad-hoc meeting for cross-team discussion so that our customers have a say in what we build and how."
Experiments - A list of the ongoing experiments that are trying to improve our process of Delivering Software.
Example: “We are enforcing a 100-line per PR limit for 1 week to see if it reduces our time-in-status for ‘In Review’, thus improving flow”
Basic SDLC
The basic software development lifecycle I prefer is roughly based on Kanban. I’ll write more articles over time on each phase, but the flow I start with is as follows. These phases can be roughly mapped to columns in a Kanban board on Jira or the like.
Backlog - Work is defined. Acceptance criteria is created and the “what are we gonna do” question is answered, generally in tickets.
Work Refined - Things that have been refined by the team. This means that the engineering and design team has talked with the Product Owner about the work, figured out how to break it into small segments, and ideally estimated how much effort / complexity / uncertainty / risk is involved.
Development - Work that is actively being developed. “Actively being developed” means not just writing code, but creating necessary technical design documentation and automated test code. Engineers validate that their code meets acceptance criteria by running it in a development environment before moving it on to review.
In review - Work is being reviewed by peers (generally, for me, through PRs in Github) and the engineer(s) who did the coding is/are responding to feedback. Code and automated tests are reviewed for accuracy, style, security, and performance. But also team members consider whether review feedback should be a blocker or just a note for a follow-up ticket. In general, I prefer teams to have two approvals on any PR, but that is easily negotiable per team.
PO Acceptance - Code is stood up in a pre-production environment in order for the product owner to validate that code has passed their written acceptance criteria.
Done - Code is merged and pushed to production, live for use by customers (pending potential feature flagging).
This system is very much a “Happy Path”.
Many intermediate states may be necessary for, first, primary work — namely if a ‘UX Review’ or ‘Manual QA’ status is needed— but also for waiting states like ‘Ready for Acceptance’ or ‘Ready to Deploy’. Teams are encouraged to add statuses to visualize their implicit queues so that metrics can be gathered (time-in-status). This helps identify where inefficiencies exist.
All that said, teams should assess if they actually need those intermediate states. Teams are expected to optimize flow through the whole system, meaning that in these categories, Developers aren’t only responsible for “Development,” and SDETs aren’t only responsible for “In Review”: the whole team must focus on getting work flowing through the system at a constant rate, even if that means working outside of their specialties.
Flow
The idea of ‘flow’ is taken from Kanban. In essence, it is a concept that optimizes how work moves through a system. Consider a car factory — it is not helpful to have a machine that stamps out and builds doors lead to a pile-up because the ‘construct chassis’ section of the factory is understaffed. The goal is to construct a whole car, and passing blame via “well I did MY part” is not good enough. It delivers no value.
In that way, software teams are all responsible for the lifecycle of their product features. Design, development, review, and deployment are responsibilities of the team as a whole. Optimizing the flow through every step in that process, not just the most natural for any member’s role, is the heart of agile development. Faced with “if only X, we’d be in great shape,” the team should find a way to change X, or, failing that, agitate for the ability to effect change in X with upper management.
CI/CD
If you’re using this as a basic guide, my hope is you’re not stuck maintaining legacy software with legacy build and deployment technology. As such, my recommendation is that you pursue true Continuous Integration / Continuous Deployment— namely, that code is regularly merged, and that each merge is automatically deployed to production.
This kind of system requires a comprehensive automated test suite that runs in the pipeline. It requires good feature-flagging so that half-finished software doesn’t break user experience. This is good! These are the kinds of things that are inexpensive to add early, and exponentially harder to bolt on over time.
The reason I prioritize CI/CD early is because an ability to continue to deliver code in near-real-time is difference-maker in claiming market-share and gaining confidence with sales and support. Engineering’s ability to deliver is directly tied to how easy it is to deploy and release code. You don’t need to get ready (to deploy code) if you stay ready (to deploy code).
General Values
The above is intended as a starting point, but you’re going to run into problems that aren’t covered. For those, I base decisions on some very basic values that help guide decision-making organization-wide:
We prefer smaller changes, released more frequently.
We prefer short feedback cycles — both while developing (within the team) and after release (with customers).
We prefer team-based ownership of work, not individual siloing. Our job is building and maintaining software that leads to outcomes, not slamming out code.
We prefer automating any onerous task — both within product features and in how we test and ensure quality.
We prefer experimentation with clear goals and measurement over pretending that we can know “the right” answer at the beginning.
We prefer building only that which differentiates us from competitors— we buy services and software that provide commodity value.