When the Data Finally Matches the Work: Building Evaluation Capacity across a Civic Engagement Network

Case Study
When the Data Finally Matches the Work
Building Evaluation Capacity across a Civic Engagement Network
Anecdotally, the work was real. Demonstrably, there was almost nothing to show for it. This case study traces how a civic engagement network of almost 100 nonprofits in Houston closed that gap; not by demanding better data, but by building the conditions where better data was possible.

The Starting Assumption

Houston is one of the most diverse cities in the country and one of the hardest places to sustain civic participation across a fragmented, under-resourced landscape. For nearly a decade, a network of almost 100 nonprofits had been doing the slow, relational work of civic engagement: mobilizing voters, building power, organizing communities. Anecdotally, we knew it was working. You could see the relationships deepening, the coordination improving, the activity compounding. And yet, we could not say, with any confidence, whether it was working. That gap between what we believed and what we could demonstrate was where this work began.

The driving metric is, and always has been, voter turnout. Funders, partners, and the network itself needed to know whether all of this organizing and mobilization was actually moving people to the polls. But proving that connection requires more than counting events or contacts. It requires being able to trace a path from a specific outreach effort to a specific voter showing up on election day. Without that, the work remains anecdotal; compelling to those inside it, invisible to everyone else.

The Challenge

The reason came down to data. Most partners were small, community-rooted organizations without dedicated data staff. Some tracked participation on paper. Others submitted aggregated summaries (total contacts, total events) that confirmed activity happened but couldn't tell us much else. Two problems, each unique enough to require different solutions. The first was foundational. The second was more specific, and it shaped everything else that followed.

1
Data Collection requires knowing what you're trying to measure and why.
The practice of continuous learning, both defining outcomes and designing data collection around a theory of change, requires time, support, and dedicated capacity that most small, community-rooted organizations simply don't have. Without that foundation, meaningful and useful data collection was beside the point. You can't measure what you haven't yet had the space to design. And when organizations are only ever asked to collect data for funders, with no investment in helping them use it for themselves, the result is predictable: data becomes something to comply with, not something to learn from.
2
The voter file is the accountability infrastructure.
It's public information: a record of who voted, when, and where. If you want to know whether your canvassing effort actually moved people to the polls, you need to match your outreach records to that file. That requires individual-identifiable data (names, addresses, contact information) collected consistently and in a format that can be linked. Five years before this work began, most partners weren't there. Aggregate counts were the norm, hiding the impact of their work.

Moving to collecting individually-identifiable data isn’t just a technical issue. It’s a relational one. The communities at the center of this work had historically (and currently) good reasons to be cautious about giving their information to anyone. Trust couldn’t be assumed. It had to be built into the infrastructure itself. You can't ask communities with legitimate reasons to distrust data collection to simply trust you because your intentions are good. The goal I kept coming back to: improve data quality without eroding community trust. Those two things had to move together.

Designing with Intention

One of the first decisions was whether to build this capacity in-house or bring in a dedicated partner. It wasn't a simple call. There were real debates about which direction made more sense. Ultimately, what drove the decision was a desire to protect what the team was already providing partners, direct campaign support. Adding evaluation capacity building to their plate risked diluting the support organizations were already counting on. Keeping it separate meant both could be done well.

The RFP did as much framing work as it did scoping work. I named our evaluation values explicitly: participatory methods, equity at the center, data as a tool for learning rather than a compliance requirement. Prospective partners were held accountable to those values during selection. The document was setting a tone, not just defining a scope.

Out of all the proposals and interviews, Pivot Data Design was the right fit. Before any curriculum was built, we both wanted partners to have input into the final product. Pivot spent six months getting to know the team and the network conducting partner interviews, attending meetings, learning how organizations actually operated rather than how they describe themselves on paper. It was important for the curriculum to be shaped by the people who would actually live with it. If we skipped that phase and went straight into instruction, we’d build something technically sound that nobody trusted or found useful. Relationships before curriculum wasn’t a soft preference. It was how we protected the integrity of the whole program.

Alongside the program design, I invested in the data infrastructure and community trust work that would help make meaningful collection possible, but that work deserves its own telling.

The program that launched in June 2023, Harnessing the Power of Data, was designed so that learning happened in multiple directions at once.

Asynchronous Learning
Online lessons delivered through Teachable gave participants exposure to core concepts before going hands-on. In pedagogy, this is the flipped classroom model. Participants arrived at workshops already familiar with the material, ready to apply it rather than encounter it for the first time.
In-Person Workshops
Hands-on sessions put the concepts to work. With the foundational knowledge already in place, workshop time could go deeper, focused on practice, troubleshooting, and building confidence with real data.
One-on-One Coaching
Monthly coaching sessions gave organizations space to apply what they were learning to their specific context. Not every organization was working on the same problems, and coaching made the learning responsive to where each one actually was.
Progress Partners
Organizations were paired with each other for year-long peer learning relationships. Knowledge didn't only flow from facilitator to participant. It moved laterally across the network, which is where some of the most durable learning happened.
Partner Stipends
Stipends were built into the budget from the start. If organizations were being asked to invest real staff time in learning, the program needed to signal that the investment was reciprocal.

The design was deliberate. Each component was designed and reinforced the other. What happened next showed whether the logic held.

Observable Shifts

Within the first year, we felt a noticeable shift across the network. The gains were clear, and visible in multiple ways:

By the numbers
  • Data quality rose 32 percentage points, from 55% to 87%.
  • Near-universal data submission replaced a baseline where many organizations had submitted nothing at all.
  • 97% of participants found the workshops beneficial.
Across the network
  • Team enthusiasm and confidence in working with data increased significantly.
  • Data collection became more streamlined and efficient across organizations.
  • Partners improved how they communicated their impact to the communities they serve.

These weren’t isolated wins. They were network-wide shifts that showed the design was working. The more important change was in how the data was actually working for the organizations, not just the funders.

Behind the network-wide numbers were individual organizations doing something different. Two organizations shared what the shift looked like in practice.

Alief Votes, a young organization with big ideas, came into the program with energy and ambition. Through the coaching process, they rebuilt their logic model and honed in on their theory of change. Youth voter turnout targets weren't just met; they were exceeded by more than 700 percent. "When we got the data from that, we were like, wow, maybe we're doing something right…now we’re seeing, OK, how can we look at the data, capture this moment that we’ve activated, and then start to think of different types of ways to engage either returning voter or new voters." said Abby Triño.

For Bridges to Empowerment, the conceptual shift was as meaningful as the technical one. "A logic model is basically a power map," said Koretta Brown, describing how her organization came to see data not as a funder requirement but as a tool for telling their own story. They began producing videos anchored to their data: more focused, more pointed, more clearly connected to outcomes. And now the organization has a formidable database of community members and voters with whom they have established relationships it can continuously engage through meetings and community events.

This is data functioning as it should: not as a reporting requirement, but as a vehicle for community voice and organizational power.

Lessons to Carry Forward

The case for evaluation capacity building is easy to make in the abstract. What's harder is getting specific about what it actually requires, and honest about what it can and can't solve.

1. Start with the accountability question, not the training question. What does your organization actually need to know, and what infrastructure does that require? For us, the voter file matching requirement was specific and consequential. Naming it clearly shaped everything downstream: what curriculum Pivot built, what standards I set, what I asked of partners. Vague evaluation goals produce vague results.

2. Treat the RFP as a design document. The values you articulate in the scope of work set the tone for the entire engagement. I wrote our evaluation principles into the RFP (equity, participation, data as learning rather than compliance) and those principles shaped who responded and how the work unfolded. That document was doing more work than I realized when I wrote it.

3. Trust and data quality are not competing priorities, but they require separate interventions. Improving data standards without attending to community trust is a shortcut that tends to backfire. Organizations need to be able to collect better information in a way they can honestly defend to the people they serve. That work isn't glamorous, but it's load-bearing.

What didn't change is also worth naming directly.

Knowledge and skill are one type of organizational capacity. Time, budget, and staffing are another. This program excelled at addressing the first. The second remains a structural gap. Many organizations now understand what good data looks like; not all of them have the personnel to sustain it without continued support. That's not a failure of the program. It's the limit of what a training investment can accomplish without a corresponding investment in human capital.

When a backbone organization sets a clear standard and then invests seriously in helping every organization in the network meet it, the whole network rises. Not because reporting improves, but because organizations learn to use evidence on their own terms, to understand what they're producing, to adapt, and to tell their story with the kind of clarity that builds power.

Previous
Previous

From Participation to Cohesion: Measuring Affiliate Health through Engagement Pathways and Network Structure

Next
Next

When the Data Isn’t the Problem