Measuring What Matters: From Theory to Data
Measuring What Matters: From Theory to Data
How to operationalize the framework into actual measurements that drive policy decisions.
The Measurement Challenge:
You can’t manage what you don’t measure.
But measuring digital equity is complex:
- Too narrow: Count broadband subscriptions → Miss aspiration and skills gaps
- Too broad: Measure everything → Drown in data, no actionable insights
- Wrong focus: Measure inputs (money spent) → Not outcomes (equity achieved)
Your framework solves this: Six Dagg Compass components map directly to the three-stage pathway, grounded in 25 years of development research.
The Dagg Compass Framework:
Six Components (Dagg et al., 2023):
- Contexts - Demographics, socioeconomics, geography
- Governance - Policy, power structures, institutions
- Connectivity - Infrastructure, devices, access
- Skills - Literacy, training, competencies
- Application - Use cases, relevance, adoption
- Outcomes - Achievement, inclusion, equity
Citation: Dagg, S., Giles, K., Knight, G., McHenry, K., & Schrubbe, M. (2023). Digital Opportunities Compass. Merit Network and University of Michigan.
Mapping Compass to Researchers:
The Brilliant Connection:
| Compass Component | Framework Stage | Researcher Theory | What It Measures |
|---|---|---|---|
| Connectivity | Opportunity | Sen (resources) | Can people ACCESS digital? |
| Application | Aspiration | Appadurai (navigation) | Do people SEE VALUE in digital? |
| Skills | Growth Mindset | Dweck (beliefs) | Can people LEARN to use digital? |
| Outcomes | Digital Equity | Sen (functionings) | Are people ACHIEVING equity? |
| Contexts | (All stages) | Hampton & Bauer (validation) | What local factors matter? |
| Governance | (All stages) | Policy implementation | Is policy enabling or blocking? |
This is the measurement operationalization of the theory!
Component 1: Contexts (Demographics)
What to Measure:
Purpose: Understand WHO we’re serving and WHAT conversion factors they need (Sen).
Key Metrics:
Demographics:
- Age distribution (older adults need different support)
- Income levels (affordability barriers)
- Education attainment (baseline literacy)
- Language diversity (content relevance)
- Disability status (accessibility needs)
Geography:
- Rural/urban classification (infrastructure challenges)
- Population density (deployment economics)
- Distance to services (transportation access)
Socioeconomics:
- Employment status (relevance of digital skills)
- Industry composition (local use cases)
- Housing characteristics (device availability)
Data Sources:
American Community Survey (ACS):
- Census.gov API
- 5-year estimates for small areas
- Annual updates
Local Data:
- School district records
- Library usage statistics
- Community needs assessments
Why This Matters:
Sen’s conversion factors: Same infrastructure → Different outcomes based on context.
Example:
Rural county, median age 58, median income $38K:
- Need: Affordable devices, relevant applications (telehealth, remote work)
- Wrong approach: Deploy fiber, assume adoption
- Right approach: Infrastructure + navigators + affordability programs
Urban county, diverse languages, 22% poverty:
- Need: Multilingual content, digital literacy, public access
- Wrong approach: English-only training
- Right approach: Culturally relevant navigators + community partnerships
Hampton & Bauer evidence: Demographics predict which gaps appear (K-12 Michigan data).
Component 2: Governance (Policy)
What to Measure:
Purpose: Assess whether policy ENABLES or BLOCKS digital equity pathway.
Key Metrics:
Policy Existence:
- Digital equity plan adopted? (Yes/No)
- Funding allocated? (Budget amount)
- Accountability measures? (Who’s responsible)
Policy Quality:
- Theory-grounded? (Cites Sen, Appadurai, Dweck, Toyama)
- Evidence-based? (Uses Hampton & Bauer findings)
- Measurement framework? (Dagg Compass implementation)
Implementation:
- Timeline adherence (milestones met)
- Budget execution (funds deployed vs. allocated)
- Stakeholder engagement (community input)
Outcomes:
- Policy impact on equity (before/after comparison)
- Barrier removal (identified obstacles addressed)
Data Sources:
Government Documents:
- State broadband office plans
- County digital equity strategies
- Municipal ordinances
Budget Data:
- Appropriations
- Expenditures
- Grant awards
Stakeholder Surveys:
- Community perception of policy effectiveness
- Navigator feedback on barriers
Why This Matters:
Good policy amplifies interventions; bad policy blocks them.
Example:
Measure: Navigator program funding
- County A: $500K allocated, $450K deployed (90% execution)
- County B: $500K allocated, $120K deployed (24% execution)
Analysis:
- County A: Policy enables action
- County B: Policy exists but implementation blocked
Investigation reveals:
- County B has procurement barriers (18-month approval)
- Solution: Streamline procurement policy
This is policy evaluation, not just policy existence!
Component 3: Connectivity (Opportunity)
What to Measure:
Purpose: Assess Sen’s “resources” stage—do people have OPPORTUNITY to access digital?
Key Metrics:
Infrastructure Availability:
- % covered by 100/20 Mbps or better (FCC standard)
- % covered by 1 Gig symmetrical (future-ready)
- Technology type (fiber, cable, fixed wireless, satellite)
Infrastructure Quality:
- Actual speeds (M-Lab tests) vs. advertised
- Latency (critical for video, gaming, remote work)
- Reliability (downtime, packet loss)
Adoption:
- % subscribing to broadband
- % with home devices (computer, not just smartphone)
- Reasons for non-adoption (survey: affordability, irrelevance, skills)
Affordability:
- Cost as % of median income
- ACP/Lifeline enrollment rates
- Device financing availability
Data Sources:
# FCC Broadband Data Collection (BDC)
import requests
fcc_api = "https://broadbandmap.fcc.gov/api/public/"
response = requests.get(f"{fcc_api}coverage", params={
'county_fips': '26153', # Schoolcraft County, UP
'speed_down': 100,
'speed_up': 20
})
coverage_data = response.json()
# M-Lab Speed Tests (Actual Performance)
from google.cloud import bigquery
query = """
SELECT
county,
AVG(download_mbps) as avg_download,
AVG(upload_mbps) as avg_upload,
STDDEV(download_mbps) as download_variability
FROM `mlab-public-data.ndt.unified_downloads`
WHERE state = 'MI' AND date >= '2024-01-01'
GROUP BY county
"""
# ACS Adoption Data
from census import Census
c = Census("YOUR_API_KEY")
adoption = c.acs5.state_county(
('B28002_001E', 'B28002_004E'), # Total households, with broadband
'26', Census.ALL # Michigan, all counties
)
Example Measurement:
Schoolcraft County (UP) Assessment:
Infrastructure Availability:
- 100/20 coverage: 68% (FCC BDC, June 2024)
- Fiber coverage: 22%
- Fixed wireless: 41%
- Satellite (Starlink): 95%
Infrastructure Quality:
- Median download: 47 Mbps (M-Lab, below 100 Mbps standard)
- Median upload: 8 Mbps (below 20 Mbps standard)
- Latency: 38ms (good for most uses)
Adoption:
- Broadband subscription: 71% (ACS 5-year)
- Home computer: 78%
- Smartphone only: 14% (device gap)
Affordability:
- Median cost: $89/month
- Median income: $43K/year
- Cost burden: 2.5% (above 2% threshold)
- ACP enrollment: 31% (good, but program ending)
CONNECTIVITY SCORE: 0.58 (moderate, infrastructure primary barrier)
Sen interpretation: Resources (infrastructure) exist but incomplete. Conversion factors (affordability, devices) also gaps.
Component 4: Skills (Growth Mindset)
What to Measure:
Purpose: Assess Dweck’s “growth mindset” stage—can people LEARN digital skills?
Key Metrics:
Digital Literacy:
- Northstar Digital Literacy assessments
- Basic skills (email, web browsing, file management)
- Advanced skills (privacy, security, online safety)
Training Engagement:
- % completing digital skills programs
- Hours of training per learner
- Retention rates (repeat participation)
Self-Efficacy:
- Confidence surveys (Likert scale: “I can learn new digital skills”)
- Mindset assessment (fixed vs. growth beliefs)
- Attribution style (success due to effort vs. luck)
Applied Skills:
- % using digital for employment
- % using digital for education
- % using digital for healthcare, government services
Data Sources:
# Northstar Digital Literacy Platform
# https://www.digitalliteracyassessment.org/
# Example aggregate data request
northstar_metrics = {
'total_assessments': 1250,
'basic_skills_pass_rate': 0.68,
'advanced_skills_pass_rate': 0.42,
'average_attempts': 1.8, # Shows persistence (growth mindset)
'improvement_rate': 0.73 # Pass on retry (learning)
}
# Community Survey (Custom)
survey_questions = [
"I can learn new digital skills if I try. (1=Disagree, 5=Agree)",
"When I struggle with technology, I see it as an opportunity to learn.",
"My digital skills can improve with practice.",
"I feel confident helping others with technology."
]
# Training Program Data
training_metrics = {
'enrolled': 450,
'completed': 338, # 75% completion (good retention)
'avg_hours': 12,
'post_training_confidence': 4.2 # Up from 2.8 pre-training
}
Example Measurement:
Chippewa County (UP) Skills Assessment:
Digital Literacy (Northstar):
- Basic skills: 64% proficiency
- Advanced skills: 38% proficiency
- Assessment participation: 850 adults (12% of population)
Training:
- Library program: 320 completions/year
- Community college: 180 completions/year
- Navigator-led: 275 completions/year
- Total: 775 completions (growth mindset pedagogy used)
Self-Efficacy (Survey, n=520):
- "I can learn digital skills": 3.8 / 5.0 (moderate confidence)
- "Technology frustrates me": 3.2 / 5.0 (some fixed mindset)
- "I help others with tech": 2.9 / 5.0 (lower, opportunity here)
Applied Skills:
- Used digital for job search: 58%
- Used telehealth: 41%
- Filed taxes online: 67%
SKILLS SCORE: 0.61 (moderate, training working but more needed)
Dweck interpretation: Growth mindset emerging (training completion high), but fixed mindset beliefs persist for some. Need reframing pedagogy.
Component 5: Application (Aspiration)
What to Measure:
Purpose: Assess Appadurai’s “aspiration” stage—do people SEE VALUE in digital?
Key Metrics:
Use Diversity:
- Number of different use cases adopted
- Frequency of use (daily, weekly, monthly)
- Meaningful use (employment, education, health, civic) vs. entertainment only
Perceived Relevance:
- Survey: “Digital helps me achieve my goals” (Likert scale)
- Survey: “I see pathways through digital” (Appadurai navigation capacity)
- Survey: “I know people who succeeded with digital” (role models)
Navigator Engagement:
- % who have interacted with digital navigator
- % who report navigator helped identify relevant use case
- Satisfaction with navigator support
Aspiration Thickness:
- Can articulate 3+ ways digital helps life (thick map)
- Sees clear steps to goals (navigation capacity)
- Has role models who used digital successfully (cultural capability)
Data Sources:
# Custom Community Survey (Appadurai-grounded)
survey_design = {
'use_diversity': [
"Which digital activities do you do regularly? (Check all)",
"[ ] Email, [ ] Social media, [ ] Online banking, [ ] Job applications",
"[ ] Telehealth, [ ] Online courses, [ ] Government services, [ ] Remote work"
],
'perceived_relevance': [
"Digital technology helps me achieve my personal goals. (1-5)",
"I see clear pathways to improve my life through digital. (1-5)",
"I know people like me who succeeded using digital. (1-5)"
],
'aspiration_thickness': [
"List 3 ways digital technology could help you in the next year:",
"(Open-ended, coded for specificity)"
],
'navigator_impact': [
"Have you worked with a digital navigator? (Y/N)",
"Did the navigator help you see new possibilities? (1-5)"
]
}
# Usage Analytics (Library, Community Centers)
usage_data = {
'job_applications_assisted': 245,
'telehealth_connections': 180,
'online_learning_enrollments': 320,
'government_services_completed': 410
}
Example Measurement:
Luce County (UP) Application Assessment:
Use Diversity (Survey, n=380):
- Email: 78%
- Social media: 68%
- Online banking: 52%
- Job applications: 34% (lower, opportunity)
- Telehealth: 28% (lower, relevant for rural)
- Online courses: 19% (lower, aspiration gap)
- Average use cases: 3.2 (moderate diversity)
Perceived Relevance:
- "Digital helps my goals": 3.4 / 5.0 (moderate, could improve)
- "I see pathways": 2.9 / 5.0 (LOWER—Appadurai aspiration gap)
- "I know role models": 3.1 / 5.0 (moderate, more models needed)
Navigator Engagement:
- Worked with navigator: 28%
- Navigator helped see value: 4.3 / 5.0 (high impact when used!)
- Would recommend navigator: 89%
Aspiration Thickness (Open-ended, n=380):
- Could name 0-1 ways: 32% (thin maps)
- Could name 2-3 ways: 51% (moderate maps)
- Could name 4+ ways: 17% (thick maps)
APPLICATION SCORE: 0.52 (moderate-low, ASPIRATION GAP evident)
Appadurai interpretation: Thin aspirational maps (can’t see pathways). Navigators effective when used, but only 28% engaged. This is Hampton & Bauer “unclear value” gap!
Component 6: Outcomes (Digital Equity)
What to Measure:
Purpose: Assess Sen’s “functionings”—are people ACHIEVING equity through digital?
Key Metrics:
Achievement:
- Employment rate improvement (attributed to digital skills)
- Educational attainment increase (online courses completed)
- Health outcomes (telehealth utilization, health literacy)
- Civic engagement (online civic participation)
Inclusion:
- Gap reduction between groups (Gini coefficient)
- Participation rates by demographic (age, income, race)
- Barrier removal (accessibility, affordability)
Digital Equity Index:
- Composite score (0-1) combining all components
- Trend over time (improving, stable, declining)
- Comparison to state/national benchmarks
Economic Impact:
- Income growth (attributed to digital skills)
- Business creation (digital-enabled)
- Remote work adoption (geographic mobility)
Data Sources:
# Outcomes Measurement Framework
# Employment Outcomes
employment_impact = {
'job_seekers_using_digital': 340,
'jobs_obtained': 87, # 26% success rate
'attributed_to_digital_skills': 68, # 78% attribute to training
'avg_wage_increase': 4200 # Annual, compared to pre-training
}
# Education Outcomes
education_impact = {
'online_course_enrollments': 520,
'completions': 380, # 73% completion
'credential_obtained': 210,
'degree_progress': 95
}
# Gini Coefficient (Inequality Measure)
from scipy import stats
def calculate_digital_gini(access_scores):
"""
Access scores: array of digital access scores (0-1) for each household
Returns: Gini coefficient (0 = perfect equality, 1 = perfect inequality)
"""
n = len(access_scores)
sorted_scores = np.sort(access_scores)
cumsum = np.cumsum(sorted_scores)
gini = (n + 1 - 2 * np.sum(cumsum) / cumsum[-1]) / n
return gini
# Example
baseline_gini = calculate_digital_gini(baseline_access) # 0.44
after_intervention_gini = calculate_digital_gini(after_access) # 0.37
improvement = baseline_gini - after_intervention_gini # 0.07 reduction (good!)
Example Measurement:
Regional Outcomes (14 UP Counties, 2-year intervention):
Achievement:
- Employment: 245 jobs obtained with digital skills attribution
- Education: 380 online course completions, 210 credentials
- Health: Telehealth utilization +42% (18% → 60%)
- Civic: Online government service use +38%
Inclusion (Gini Reduction):
- Baseline digital access Gini: 0.44 (2022)
- After intervention Gini: 0.37 (2024)
- Improvement: 0.07 reduction (significant!)
- Interpretation: Gap between highest and lowest access narrowed
Participation by Demographic:
- Age 60+: Broadband adoption 58% → 71% (+13 points)
- Income <$35K: Adoption 52% → 68% (+16 points)
- Rural: Adoption 64% → 78% (+14 points)
- Urban: Adoption 79% → 84% (+5 points)
- EQUITY: Rural/urban gap narrowed from 15 points to 6 points
Economic Impact (Survey attribution):
- Avg income increase (digital skills): $3,800/year
- New businesses (digital-enabled): 28
- Remote work adoption: 12% → 19% (+7 points)
- Estimated regional economic impact: $4.2M/year
OUTCOMES SCORE: 0.68 (GOOD—equity improving, gaps narrowing)
Sen interpretation: Functionings achieved! People using digital to do/be what they value. Conversion factors (navigators, training) working.
Integration: Compass Components → Bayesian Network:
How Metrics Feed the Prediction Model:
The Bayesian network uses measured Compass components as inputs:
# Bayesian Network Structure (from TrainingCompassBayesian.md)
from pgmpy.models import BayesianNetwork
from pgmpy.inference import VariableElimination
# Define model structure
model = BayesianNetwork([
# Contexts influence everything
('Contexts', 'Connectivity'),
('Contexts', 'Skills'),
('Contexts', 'Application'),
# Governance enables all stages
('Governance', 'Connectivity'),
('Governance', 'Skills'),
('Governance', 'Application'),
# Three-stage pathway (Sen → Dweck → Appadurai)
('Connectivity', 'Skills'), # Opportunity enables Growth Mindset
('Skills', 'Application'), # Growth Mindset enables Aspiration
('Application', 'Outcomes'), # Aspiration enables Equity
# Also direct path (infrastructure alone insufficient, but contributes)
('Connectivity', 'Outcomes'),
('Skills', 'Outcomes')
])
# Measured data populates conditional probability tables (CPTs)
# Example: P(Outcomes | Connectivity, Skills, Application)
# Query: What if we improve Connectivity to 0.70?
inference = VariableElimination(model)
result = inference.query(
variables=['Outcomes'],
evidence={'Connectivity': 0.70, 'Skills': 0.61, 'Application': 0.52}
)
print(result)
# Output: P(Outcomes=High) = 0.58
This is how measurement drives prediction!
Workflow:
- Measure Compass components (data collection)
- Input to Bayesian network (evidence)
- Query network for predictions (policy scenarios)
- Deploy intervention based on prediction
- Re-measure after 6 months (update model)
- Compare actual to predicted (learn)
- Refine CPTs with new data (improve predictions)
Data Collection Frequency:
Recommended Measurement Schedule:
| Component | Frequency | Rationale | Data Source |
|---|---|---|---|
| Contexts | Annual | Demographics change slowly | ACS Census |
| Governance | Quarterly | Policy implementation tracking | Government records |
| Connectivity | Semi-annual | Infrastructure deployment tracking | FCC BDC, M-Lab |
| Skills | Quarterly | Training program outcomes | Northstar, surveys |
| Application | Quarterly | Use patterns evolve with exposure | Surveys, usage analytics |
| Outcomes | Semi-annual | Equity changes require time | Composite measures |
Critical: Need both leading indicators (Skills, Application) and lagging indicators (Outcomes) to adapt policy quickly.
Real Example: EUP Connect Measurement:
User’s Lived Experience (From WHY_RESEARCH_IN_GRAPH_IS_BRILLIANT.md):
Situation: Five years as broadband project manager, eastern Upper Peninsula Michigan.
Challenge: How to measure if interventions working?
Solution: Implemented Compass-like framework (before Dagg published):
Year 1 Assessment (Baseline):
Connectivity (Infrastructure):
- Fiber coverage: 34% of locations
- 100/20 coverage: 58%
- Adoption rate: 67%
- Score: 0.51 (moderate-low)
Skills (Growth Mindset):
- Library training participants: 450/year
- Northstar assessments: Limited data
- Community survey confidence: 2.9 / 5.0
- Score: 0.48 (moderate-low)
Application (Aspiration):
- Use diversity: Low (2.8 use cases avg)
- Perceived relevance: 3.1 / 5.0
- Navigator access: None (didn't exist yet)
- Score: 0.44 (LOW—Hampton & Bauer gap)
Outcomes (Equity):
- Gini coefficient: 0.46 (high inequality)
- Rural/urban gap: 18 percentage points
- Score: 0.39 (low)
Year 3 Assessment (After Intervention):
Connectivity: 0.68 (+0.17) - Fiber deployment + ACP
Skills: 0.62 (+0.14) - Training programs expanded
Application: 0.58 (+0.14) - Navigator program launched (KEY!)
Outcomes: 0.61 (+0.22) - Gini to 0.39, gaps narrowing
MEASUREMENT VALIDATED:
- Application (aspiration) gap was real (Hampton & Bauer)
- Navigator program addressed it (Appadurai theory)
- Outcomes improved significantly (Sen functionings)
This is measurement-driven policy adaptation in action!
Common Measurement Mistakes:
Mistake 1: Measure Inputs, Not Outcomes
Wrong:
"We spent $5M on infrastructure!" (input)
"We trained 500 people!" (output)
Right:
"Infrastructure deployment (input) →
Connectivity score 0.45 → 0.68 (outcome) →
Equity score 0.38 → 0.54 (impact)"
Measure the ENTIRE pathway, not just activities.
Mistake 2: Measure Too Late
Wrong:
Deploy intervention → Wait 3 years → Measure outcomes
(Can't adapt if not working!)
Right:
Measure quarterly → Compare to predictions → Adjust strategy
(Continuous improvement)
Leading indicators (Skills, Application) predict lagging indicator (Outcomes).
Mistake 3: Measure Without Theory
Wrong:
"Let's measure 47 metrics because they're available"
(No framework, can't interpret)
Right:
"Dagg Compass operationalizes Sen, Appadurai, Dweck, Toyama →
Each component theoretically grounded →
Measurement has PURPOSE"
Theory drives what to measure and why.
Mistake 4: Single Metric Focus
Wrong:
"We'll measure broadband adoption rate. Done!"
(Ignores aspiration, skills, equity)
Right:
"Connectivity (Opportunity) + Application (Aspiration) +
Skills (Growth Mindset) → Outcomes (Equity)"
Need all three stages to measure complete pathway.
Measurement Checklist:
For Each Policy Intervention:
Before Intervention:
- Baseline Compass assessment (all 6 components)
- Data collection plan established
- Measurement frequency determined
- Bayesian network CPTs initialized with baseline data
- Predicted outcomes calculated
During Intervention:
- Quarterly leading indicator measurement (Skills, Application)
- Semi-annual lagging indicator measurement (Connectivity, Outcomes)
- Compare actual to predicted (variance analysis)
- Update Bayesian network with new evidence
- Adapt strategy based on data
After Intervention:
- Final Compass assessment
- Equity outcome measurement (Gini, gaps)
- Compare to baseline and predictions
- Document lessons learned
- Share data with research community
- Update knowledge graph for future predictions
If yes to all → Evidence-based, measurement-driven policy!
Bottom Line:
Measurement transforms theory into accountability.
The Dagg Compass operationalizes:
- Sen’s Opportunity (Connectivity component)
- Appadurai’s Aspiration (Application component)
- Dweck’s Growth Mindset (Skills component)
- Sen’s Equity (Outcomes component)
- Hampton & Bauer validation (all components)
This creates:
- Baseline: Where are we now?
- Gaps: What’s missing?
- Interventions: What to deploy?
- Predictions: What will happen?
- Monitoring: Is it working?
- Learning: How to improve?
From theory → to evidence → to measurement → to better policy.
This is 21st-century evidence-based digital equity practice.
See Also:
TrainingCompassDagg.md- Deep dive on Compass frameworkTrainingCompassBayesian.md- How metrics feed prediction modelTrainingCompassGini.md- Inequality measurement detailsTrainingCompassPolicy.md- Using metrics for policy decisions
Version: 1.0
Last Updated: November 2025
Part of: Project Compass (Merit Network) - Digital Opportunities Intelligence Network (DOIN) • Working draft