EDORA
Skip to content

EDORA Learn — Methods

← Back to Learning Center

Aggregation, Disaggregation, and Scale Effects

Every dataset can be viewed through different lenses—individual, program, county, or state. Aggregation simplifies, but it also silences. Understanding how scale alters meaning keeps analyses honest and comparisons fair.

Overview

Every dataset can be “zoomed” to multiple scales—individual, program, county, or state. Each zoom level tells a different truth. Aggregation simplifies and strengthens signals but hides nuance; disaggregation exposes difference but invites volatility and privacy risk. Understanding scale keeps comparisons fair and conclusions humble.

  • Concept. Scale determines meaning. Combining data across people or places can reverse patterns (Simpson’s paradox) or mislead about individuals (ecological fallacy). Always name the analytic level behind a statistic.
  • Connection to the Data Dictionary & Pipelines.Fields carry implicit scale tags:individual, program, geographic, ortemporal. Pipelines that merge or summarize data must document which level is preserved and which are collapsed.
  • How it shows up in dashboards & docs.Charts display scale badges—“Youth-level,” “Program-level,” “County-average.” Hover text clarifies denominators and time windows. Comparison pages include a “Scale Notes” section explaining how aggregation or smoothing was applied.

Takeaway: Scale choice is an analytic decision, not a convenience. Label it, justify it, and help readers see what’s gained and lost in translation.

Levels of Analysis

  • Individual level: Tracks person-based events or outcomes—highest resolution but privacy sensitive.
  • Program level: Aggregates by site, intervention, or provider; reveals operational differences.
  • Geographic level: County or state roll-ups useful for policy but risk masking local patterns.
  • Temporal level: Monthly, quarterly, or annual summaries show cycles but may blur short-term effects.

Common Pitfalls

  • Ecological fallacy: Inferring individual behavior from group averages (e.g., assuming every youth in a high-rate county is at higher risk).
  • Atomistic fallacy: Assuming relationships at the individual level hold for aggregates.
  • Simpson’s paradox: A pattern reversed when data are combined; aggregation changes the sign of association.

Equity and Representation

Disaggregation by race, gender, geography, or program uncovers disparities hidden in averages. However, small subgroup counts raise suppression and confidentiality concerns. Balance clarity with privacy by reporting both detailed and pooled results where safe.

  • Publish disaggregated tables alongside aggregated summaries.
  • Include notes on cell suppression and pooling logic.
  • Use small-area estimation only when validated and transparently documented.

Scale and Statistical Stability

Smaller units show more volatility—one event can swing a county rate dramatically. Analysts can stabilize through multi-year averages, Bayesian smoothing, or regional pooling, but every adjustment must be labeled so readers know the data’s true granularity.

Data & Methods

Scale decisions often drive disagreement between state and local reports. Two agencies can be accurate yet inconsistent if one reports youth counts and the other rates per 1,000 residents, or if their time windows differ. The method for transparency is to declare the level of aggregation, the denominator, and the temporal frame for every indicator.

Checklist for transparent scaling

  • Level of analysis. Individual, program, county, or state. Note which entities were combined or averaged.
  • Denominator alignment. Ensure rates use the same population base when comparing across regions or time.
  • Temporal window. State the span used (month, quarter, year) and whether data were smoothed or averaged.
  • Disaggregation. Show results by race, gender, geography, or program when safe to do so. Note suppression rules and pooling logic.
  • Stability adjustments. When small units fluctuate wildly, document any smoothing (multi-year average, Bayesian shrinkage, regional pooling) and label the adjusted metric accordingly.
  • Interpretation guardrails. Remind readers that group patterns do not imply individual risk, and vice versa.

Reusable metadata pattern

scale_level: individual | program | geographic | temporal
              denominator: "population under supervision" 
              time_window: "CY2024"
              aggregation_method: "mean | median | sum | rate_per_1000"
              disaggregated_by: ["race", "gender", "county"]
              suppression_rule: "n < 10"
              stability_adjustment:
                used: true
                method: "3-year rolling average"
                note: "Stabilizes small-county variation"
              comparability_notes: "Aligned denominators across state and county reports"
              display_label: "County 3-year average youth arrest rate"

Reading scale correctly

Before comparing numbers from different jurisdictions or years, verify that they share the same denominator, population base, and time frame. Differences in scale can mimic trends or disparities that don’t truly exist. When in doubt, visualize both the disaggregated and aggregated versions side by side.

Related

Transparency note: Always label the level of analysis and note when disaggregation or smoothing was applied. Scale shapes truth; honesty about scale preserves it.