Skip to content

What Drives Participation in Amateur Radio and Digital Communications?

A Market Research Proposal to Inform ARDC's Grantmaking Strategy

Prepared by: Jim Idelson
Date: March 2026
Status: Draft


1. Executive Summary

ARDC allocates significant resources to programs, technologies, and communities with limited systematic knowledge of the people those decisions are meant to serve. Who is in the amateur radio and digital communications ecosystem today? What brings people in, and what causes them to disengage? What messages land with which audiences? Where can a grant or program measurably change participation outcomes? These are not rhetorical questions. They are the information gaps sitting at the foundation of every strategic funding decision ARDC makes.

This proposal describes a phased, mixed-method market research program built to close those gaps. The research produces three concrete outcomes: a clear, data-grounded picture of who is in the licensed amateur radio community and what determines engagement at every stage of the licensee lifecycle; specific findings about where programs and grants can shift participation outcomes for licensed hams, digital communications practitioners, and the adjacent communities from which future experimenters will come; and evidence about how amateur radio and digital communications are perceived by people not yet part of these communities, and what narrative actually moves them.

The program is structured so ARDC receives actionable findings at each major milestone rather than at the end of a long wait. The first substantive deliverable, a Qualitative Findings Report, arrives well before the main survey data is collected. Every phase closes with a structured briefing and an approval gate before resources move to the next phase.

The research findings are the visible output. But this engagement also builds three durable organizational capabilities that change how ARDC operates long after the final report is delivered. The research program described here gives ARDC the intelligence, the frameworks, and the infrastructure to act on that reality with confidence.

Jim Idelson brings the research methodology background, the amateur radio community knowledge, and the established ARDC relationships to conduct this work in a way an outside firm could not replicate. This proposal describes what the research will do, how it will be executed, and what ARDC will have when it is complete.


2. ARDC's Goals & This Project's Role

ARDC is not operating in a vacuum. There is substantial anecdotal knowledge about the amateur radio and digital communications community — observations from ARDC's grant program, the experience of community insiders, patterns visible to VEs and club leaders and online forum participants. That knowledge is real and it should inform how this research is designed. What it cannot do is tell ARDC how representative those observations are, how much variation exists beneath the surface, or whether the patterns that seem obvious hold up across the full range of the community. The anecdotal picture is valuable as a starting point. Systematic, quantitative research is what turns it into a basis for confident decisions.

%%{init: {
  "flowchart": { "useMaxWidth": true, "nodeSpacing": 70, "rankSpacing": 120, "curve": "basis" },
  "themeVariables": { "fontSize": "19px", "nodePadding": 18 }
}}%%
flowchart LR
  subgraph WITHOUT["Without This Research"]
    direction TB
    W1["Partial signals<br/>(anecdote, applications,<br/>insider impressions)"]
    W2["Investment<br/>decisions"]
    W3["Outcomes<br/>unmeasured<br/>or unclear"]
    W1 --> W2 --> W3
  end

  subgraph WITH["With This Research"]
    direction TB
    R1["Research-grounded<br/>understanding<br/>of the community"]
    R2["Targeted<br/>investment<br/>decisions"]
    R3["Measured<br/>outcomes"]
    R4["Refined<br/>strategy"]
    R1 --> R2 --> R3 --> R4
    R4 -->|"learn and adjust"| R2
  end

  WITHOUT -. "This proposal" .-> WITH

Exhibit 2.1: Research Benefits for ARDC — Sharper granting focus and ability to quantify grant results

ARDC's grantmaking is itself a form of structured experimentation: fund things, learn what works, iterate. That cycle works well when it is informed by good data about what the community actually needs and where the leverage is. Without that foundation, the learn-experiment-do cycle still runs, but it starts from a weaker position. This research gives that cycle a firmer footing by answering the foundational questions first: Who is in the licensed amateur radio community, what drew them in, and what keeps them engaged or drives them away? Who is in the digital communications community adjacent to amateur radio, what motivates their participation, and what does ARDC's mission look like from where they stand? Who are the people within reach of amateur radio and digital communications who have not yet connected with either, and what would bring them in?

Those questions are not the only ones. ARDC has already identified specific areas it wants to understand better: how messaging affects perception of amateur radio among people unfamiliar with it; what role radio clubs and maker organizations play as entry points; how people move through the licensing process and what happens after; where the analog-to-digital transition creates opportunity or friction. These are addressed within the same integrated research program, not as separate studies layered on top, because the answers are interconnected.

Beyond supporting ARDC's internal decisions, this work creates an external asset for the field: a credible, repeatable evidence base on the nature and trends of participation in the amateur radio community.

Two questions about scope deserve direct acknowledgment. The first is the AR/DC boundary. ARDC's focus sits at the intersection of amateur radio and digital communications, but where that boundary sits in practice has been genuinely fuzzy. The Charter phase of this project will include an explicit conversation with ARDC about where the boundary sits for purposes of this research. That decision belongs to ARDC, made with full awareness of what each choice includes and excludes. The second is geography. The primary survey research is built on US infrastructure, in particular the FCC Universal Licensing System, and is explicitly scoped as US-focused. The expert interview program is global by design, however, and will include voices from international ham radio organizations and the broader digital communications world from the outset. For the formal survey workstreams, we are open to exploring opportunities to add non-US populations where appropriate sources and execution paths exist. Both scope questions are named here so they do not get papered over.

Beyond producing a body of findings, this engagement builds three organizational capabilities that persist long after the final report. The first is an evidence-based framework for evaluating grant proposals against demonstrated gaps in the ecosystem, along with a baseline that makes it possible to measure whether past and future grants actually moved the needle. The second is the foundation for proactive grantmaking, where ARDC identifies specific addressable gaps and designs grant initiatives around them, rather than evaluating only what arrives in the inbox. The third is a credible, data-grounded public presence as the authoritative source of knowledge about the amateur radio and digital communications community, positioning ARDC's grantmaking decisions in a way that builds trust inside and outside the community. Each of these is a direct response to a dimension of the grantmaking challenge ARDC faces today. They get fuller treatment later in this proposal.


3. Research Objectives: The Decision Map

The purpose of this research is not to produce interesting findings or add to the body of academic knowledge about amateur radio demographics. It is to produce findings that drive specific decisions. Every question in the research program connects to a decision ARDC needs to make: where to invest, which challenges to prioritize, which programs to strengthen, where to join efforts already underway, and where to scale things that are working well but have not yet reached their potential. This section maps those connections.

Locked: Decision Logic, Scope, and Success Criteria

This section is intended to be readable by sponsors and reviewers as a concise statement of what ARDC will learn and be able to decide. Unless explicitly revised and approved, this section is the decision lock for the proposal.

| Ever-Licensed Population Decision Area | Key Decisions ARDC Must Make | Evidence Needed | Primary Outputs | |---|---|---|---| | Lifecycle investment strategy | Where to invest across the license lifecycle | Engagement stage distributions, trajectory patterns, stage-level barriers/drivers | Prioritized lifecycle intervention map | | Retention and reactivation strategy | Which segments need retention, reactivation, or both | Inactivity signals, renewal intent, lapse factors, re-entry triggers | Segment playbook (retain/reactivate) | | Program targeting | Which channels and program models are most likely to move outcomes | Pathway and channel exposure data, participation continuity indicators | Program hypothesis set and test-ready priorities | | Evaluation readiness | How to measure and monitor whether grants are plausibly moving participation outcomes | Logic-model alignment, feasible indicators, baseline and trend metrics | Grant KPI framework, baseline, and recommended measurement plan | | Never-Licensed / Adjacent Decision Area | Key Decisions ARDC Must Make | Evidence Needed | Primary Outputs | |---|---|---|---| | Outreach and audience strategy | Which technical-adjacent audiences and messages to prioritize | Awareness/receptivity patterns, message-response signals, barrier patterns | Audience/message matrix by segment | | Ecosystem growth priorities | Where participation growth is most achievable in learn-experiment-do pathways | Hands-on experimentation frequency/depth, project participation indicators, continuation intent | Segment-prioritized opportunity map | | Portfolio design | Which pre-license and non-license pathways merit investment | Pathway participation patterns, motivation/barrier profiles, channel effectiveness | WS2 investment recommendations by pathway | | Partnership focus | Which partners can accelerate outcomes most quickly | Access routes, trust anchors, participation multipliers | Partner/channel strategy map | **Locked commitments** - **Ever-licensed success definition:** stronger enrollment-to-activity continuity, deeper engagement, and lower lapse risk (measured through renewal intent, inactivity signals, and reactivation triggers). - **Never-licensed/adjacent success definition:** increased awareness, experimentation depth, and participation across AR/DC pathways; conversion to licensed amateur radio is one useful metric, not the sole success criterion. - **General population stance:** a general population benchmark is optional and directional, used to calibrate awareness/receptivity baselines. Priority remains technical-adjacent and technical-practitioner segments where ARDC can act directly. - **Inference posture:** the ever-licensed population analysis is probability-oriented using the FCC ULS frame. Never-licensed/adjacent analysis is segmented and directional with explicit limits on inference and triangulation across segments. - **Method posture:** mixed methods with qualitative exploration, instrument design, cognitive testing/pilot, and quantitative fielding; no full launch without gate approval. - **Option structure:** official tiers remain Core, Plus, and Max. Expansion labels are Core Program, Technical Depth Expansion, and Longitudinal Extension. - **Timeline and investment posture:** phased ranges and decision-dependent cost bands are finalized after Charter and sample design decisions; no false precision early. - **Governance record:** material scope/definition/output changes are logged in **Appendix J: Decision Registry (and Change Log)**. The research is organized around four decision domains. Each domain represents a cluster of questions that, when answered, gives ARDC a clearer basis for action in a specific area of its grantmaking. Together they form a complete picture of the ecosystem ARDC is trying to strengthen. **Participation Mechanics** covers the fundamental questions about how people enter and move through the amateur radio and digital communications communities. Who gets licensed? What draws them in? Where do they get stuck or drop off? What distinguishes the person who becomes an active, lifelong participant from the one who passes the exam and disappears? In the amateur radio domain, the 10-year FCC license cycle creates a natural lifecycle with identifiable stages, each one an opportunity to understand what is working and what is not. Equally important is understanding the distribution of engagement patterns across the licensed population — how common it is for people to never meaningfully engage, to thrive long-term, to start strong and then drift, or to disengage entirely at different points in the journey. This domain provides the foundation for everything else: without understanding the participation cycle and the patterns within it, it is not possible to know where interventions can have the most effect. **Adoption and Retention Drivers** goes deeper into motivation. Getting someone licensed is not the same as getting them engaged. Getting someone engaged is not the same as keeping them there. This domain asks what accelerates or blocks deeper involvement as a radio operator, an experimenter, a builder, a software developer, or a community contributor. What role do clubs and maker organizations play? What technologies open doors versus create barriers? What changes in people's lives lead to disengagement, and are there interventions that could change that outcome? This is also the domain that probes the conditions associated with each engagement trajectory: what is consistently present when people become deeply active, and what is consistently present when engagement falters or ends? Radio clubs and maker organizations are a specific area of interest here: ARDC has identified them as potentially high-leverage pathways, and the research will test that directly. **Messaging and Perception** addresses what amateur radio and digital communications look like from the outside and from the inside. Among adjacent technical populations — STEM educators, maker community participants, technical hobbyists, and DC practitioners — awareness may be present but uneven, and conversion to sustained participation remains largely unresearched. This domain asks what would make a technically curious person lean in rather than scroll past, and what narratives work for which audiences. From inside the community, perception matters too. Amateur radio contains distinct sub-communities with deep expertise, and the research will look at whether movement across those communities is open or obstructed. A frequent pattern is that unwelcoming expertise creates invisible barriers for engaged newcomers who have not yet accumulated the knowledge to fit in. That dynamic, if it shows up in the data, is directly relevant to where ARDC invests in onboarding and mentoring programs. **Investment Levers** is where the decision-support orientation of this research becomes most explicit. The first three domains produce understanding; this one produces direction. Which lifecycle stages represent the highest-leverage investment opportunities? Which programs and grant categories have demonstrated community impact and which are operating on assumptions that have not been tested? Where are the gaps that no current grantee is addressing, and what would it take to lead rather than follow in filling them? Where are efforts already underway that ARDC could bring to scale? What would constitute evidence that grants are plausibly moving participation, retention, and digital adoption outcomes over time? This domain connects the research findings directly to ARDC's grantmaking calendar and its developing grant evaluation framework. The table below summarizes the four domains and the core research questions each one addresses. Appendix A (Questions and Measurement Map) provides the full mapping from these objectives to specific survey questions and qualitative inquiry themes. | Decision Domain | Core Research Questions | |---|---| | Participation Mechanics | Who gets licensed and through what pathways? What draws people in at each stage? Where does engagement break down, and why? What differentiates those who stay from those who lapse? | | Adoption and Retention Drivers | What accelerates or blocks deeper involvement as a ham, experimenter, or technologist? What role do clubs, mentors, and communities play? What barriers does the analog-to-digital transition create, and what opens it up? | | Messaging and Perception | What does amateur radio mean to people outside the community? What narratives work for which audiences? Are there internal barriers to movement within the hobby, and what sustains or reduces them? | | Investment Levers | Where can grants and programs measurably change outcomes? Which lifecycle stages and population segments represent the highest-leverage investment opportunities? What gaps are currently unaddressed, and what is already ready to scale? | --- ## 4. The Research Populations Understanding the amateur radio and digital communications ecosystem requires research across multiple distinct populations. These groups have different relationships with the hobby and the technology, different motivations, different access points, and different things to tell us. No single survey or recruitment method reaches all of them. This section describes who each population is, why they matter to ARDC's grantmaking strategy, and how they will be reached. The card map below is a quick orientation view of the five populations before the detailed definitions that follow.

Population 1: Current Licensees

Decision relevance: Understand engagement distribution within current licensees.

Signal examples: Activity level, mode diversity, renewal intent.

Primary access: FCC ULS mail-to-web.

Population 2: Former Licensees

Decision relevance: Identify reactivation opportunities and barriers to return.

Signal examples: Expiration drivers, return intent, re-entry friction.

Sub-population: Grace-period licensees (still eligible for fee-only reinstatement).

Primary access: Historical FCC ULS + data hygiene + mail-to-web.

Population 3: Adjacent and Affinity Communities

Decision relevance: Prioritize pre-license pathways with highest practical potential.

Signal examples: Awareness, receptivity, first-step participation.

Primary access: Partner networks + targeted outreach + screened panels.

Population 4: Digital Communications Practitioners

Decision relevance: Improve ARDC fit and impact in non-licensed DC communities.

Signal examples: Collaboration patterns, tooling barriers, support needs.

Primary access: Open-source communities + technical channels + practitioner panels.

Population 5: General Population Benchmark (Optional)

Decision relevance: Directional context for awareness and receptivity baselines.

Signal examples: Unaided awareness, interest, perceived relevance.

Primary access: Benchmark-oriented online panel module.

Exhibit 4.1. Population Segmentation Cards — five populations, decision relevance, and primary access method

**Population 1: Currently Licensed Amateur Radio Operators** All holders of a current FCC amateur radio license (administrative status: current licensee), from newly licensed Technicians to long-tenured Extra-class operators. This population provides the core picture of current licensees: who is in it, how engaged they are, what drew them in, and what has sustained or reduced that engagement over time. Administrative status and engagement level are not the same thing: some current licensees are highly active, some are occasional participants, and some are largely inactive. The data from this population will produce that distribution directly. Reach: FCC Universal Licensing System (ULS) — a publicly available federal database of all current US license holders with name and mailing address. Mail-to-web invitation model using ULS-sourced addresses. Supplementary lists from international licensing authorities or national amateur radio societies welcomed where available. **Population 2: Former Licensees** All holders of an FCC amateur radio license that has expired without renewal (administrative status: expired licensees). This includes both those still within the two-year grace period (reinstatement requires only a renewal fee) and those beyond it (reinstatement requires re-examination). In this proposal, the grace period group is treated as a sub-population of Population 2. This population surfaces the full picture of departure from the licensed community - what eventually led people to allow their license to expire, how that arc unfolded, and what, if anything, might have changed the outcome. Combined with Population 1, it completes the map of the licensed lifecycle from current to expired status. The grace-period sub-population is especially actionable because the window for low-friction return is narrow and closing. The population most valuable for understanding fresh, actionable disengagement is not defined by a single group or license status. People at or near the point of tipping away from continued participation show up across all of the populations in this section: licensees in the later stages of the active license cycle who are drifting, those newly lapsed who recently crossed the line, and members of adjacent groups who tried amateur radio and stepped back. The research design identifies this at-risk population through survey responses rather than sampling frame alone — engagement trajectory questions across all populations will surface people who are recent or imminent disengagers, and the analysis will build a profile of the conditions and experiences that characterize that moment. That profile is what informs targeted intervention design. Reach: Historical FCC ULS records retain lapsed license data with the same name and address fields as active records. Additional data hygiene steps required (address updates, deduplication). Same mail-to-web model as Population 1. Supplementary international lists welcomed where available. **Population 3: Adjacent and Affinity Populations** People who have not pursued a ham license but whose interests, activities, or contexts put them within reach of amateur radio and digital communications. Three sub-groups are of particular interest: *STEM and maker communities:* STEM educators and students, participants in robotics clubs and competitions, makers, hardware tinkerers, hackerspace and makerspace participants, and youth organization members with technology components. *Consumer and specialized radio users:* GMRS and FRS licensees and active users, and hobbyist radio users in other established communities including CB, marine radio, and aviation radio. *Emergency preparedness and community service volunteers:* CERT participants, volunteer communications support for public events (road races, parades, marathons), neighborhood preparedness networks, and people whose interest in communications capability in emergencies or high-stakes situations has developed outside amateur radio. Research with these groups identifies the populations most likely to find amateur radio and digital communications relevant, what they currently know and how they perceive these communities, and what would move them from adjacent interest to active engagement. This is the foundation for evidence-based outreach and recruitment strategy. Reach: Partner organizations across STEM education, maker networks, robotics programs, and emergency preparedness communities; online panels with interest-based screening; targeted outreach through digital channels where these communities are active. **Population 4: Digital Communications Practitioners (Non-Licensed)** Researchers, developers, and technical contributors working in the digital communications and open-source radio space — contributors to open-source radio software, SDR researchers in university and non-profit contexts, mesh networking experimenters, and developers of DC technology that ARDC funds or might fund — who engage with the technology primarily through technical collaboration and development rather than on-air operation. These are not aspirational recruits. They are active participants in the work that defines ARDC's mission at the DC/AR intersection. Research with this population surfaces what motivates their participation, what they need, and where ARDC investment can have the most impact — independent of whether their path ever leads to a ham license. The precise scope of this population will be confirmed with ARDC during the Charter phase. WS2 design will treat this segment as potentially global in scope, given the international nature of open-source digital communities. Reach: Open-source DC forums and communities, social channels where these communities are active, and online panels with technical screening criteria. **Population 5: General Population Benchmark (Optional/Directional)** The broader US adult population, with no assumed connection to amateur radio, digital communications, or maker culture. This module is optional and directional. This module can establish a directional awareness and perception baseline: what fraction of the adult population has heard of amateur radio, what associations it carries, and what might increase receptivity among technically inclined segments. It is intentionally soft-pedaled in this proposal because national-quality general population work is expensive to execute well and often less actionable for ARDC than technical-adjacent and practitioner segments. If included, scope will be calibrated during Charter to maintain proportional value. Reach: Online survey panels with benchmark-oriented sampling and explicit limits on inference. --- **A Note on Cross-Tabulation and Sample Size** Across all populations, the research is designed to support sub-group analysis — by license class, engagement level, geography, age, and other relevant dimensions. The minimum cell sizes required for reliable sub-group analysis determine the overall sample target for each population, and therefore a significant portion of research cost. The cross-tabulation plan must be finalized before sampling targets are set. This is one of the first items addressed in the Charter phase. --- **The Community Snapshot: Act on What We Know** The cross-sectional survey data from Populations 1 and 2 will produce, for the first time, a reliable, population-representative picture of where the licensed and previously-licensed amateur radio community stands today. Not an estimate built from online forum activity or grant applicants — an actual sample drawn from the complete pool of FCC license holders. That picture will show the distribution of engagement levels across the licensed population: who is deeply active, who is passively licensed, who has effectively stepped away, and who has departed entirely — and how large each of those segments actually is. Knowing that, say, 40% of licensed hams are essentially inactive is a very different prioritization input than knowing the figure is 15%. The snapshot will show how those proportions vary by license class, geography, time since licensing, and other dimensions, and will identify the segments that are thriving, the segments that are at risk, and the segments that have already been lost. This is actionable intelligence in its own right. ARDC and its grantees can design programs aimed at the segments where engagement is fragile. Investments in onboarding, mentoring, and community infrastructure can be targeted at the stages of the lifecycle where people need them most. The snapshot establishes a baseline against which future program outcomes can be measured — so that when ARDC invests in improving retention, there is a reference point to measure progress from. The community is what it is today. The snapshot helps us understand what we have to work with, and where the investment opportunities are in the near term. But here is the question the snapshot alone cannot answer: how do people get to these states in the first place — and can we reach them before they do? **Engagement Trajectories: Getting Ahead of the Problem** We can. Mainstream market research offers proven techniques for understanding how people change over time. Applied to the licensed amateur radio community, these techniques allow us to reconstruct the journey each respondent has taken — not with perfect precision, but with enough fidelity to identify the common patterns, the inflection points, and the conditions associated with each. Where the snapshot shows us the current distribution of engagement, trajectory analysis answers the deeper question: how did people get there? Which experiences early in the license lifecycle predict strong, sustained engagement later? At what point in the journey does disengagement typically begin? What conditions — a lack of mentoring, a bad club experience, a shift in life circumstances, a technology barrier — are consistently present when people fall away? And critically: where in the lifecycle could a well-designed program catch people before engagement falters, rather than after? This is the shift from understanding the current picture to acting on it proactively — designing programs that intercept people earlier, rather than investing in re-engagement only after disengagement has fully set in. We are appropriately cautious about what the trajectory analysis will deliver. Retrospective recall is imperfect, and how well the data supports trajectory reconstruction in this specific domain will depend on what the survey produces. The snapshot is the guaranteed primary output; the trajectory work is a complementary analytical layer we will build to the extent the data allows. What it produces — even partially — makes the program investment guidance sharper. And it lays the groundwork for something more powerful still: true cohort tracking over time, following the same individuals through the license lifecycle in future research waves, to test whether the programs ARDC invests in today are actually changing outcomes. *Exhibit 4.2: FCC License Lifecycle and Engagement Trajectory Frame — the ten-year license term, cohort positions, and conceptual trajectory types* > Key point: We measure engagement signals across the lifecycle and relate them to renewal / lapse outcomes.
%%{init: {
  "flowchart": { "nodeSpacing": 60, "rankSpacing": 75, "curve": "basis" },
  "themeVariables": { "fontSize": "20px", "nodePadding": 16 }
}}%%

flowchart TB

  %% Hard lifecycle anchors (neutral spine)
  A["Initial License<br/>Grant Date"]
  B["Renewal window opens<br/>(+ 9 yrs 9 mo)"]
  C["Expiration Date<br/>(+ 90 days)"]
  D["End of Grace Period<br/>(+ 2 years)"]

  %% Invisible spacers to depict the long mid-term span (A→B)
  S1[" "]
  S2[" "]
  S3[" "]
  S4[" "]
  S5[" "]

  %% No arrowheads for the long A→B span; arrow into the renewal window + milestone steps
  A --- S1 --- S2 --- S3 --- S4 --- S5 --> B --> C --> D

  %% Outcomes (terminators)
  O1["On-time Renewal"]
  O2["Grace Renewal"]
  O3["No Renewal<br/>License lost"]

  %% Trajectories (top→bottom time sequence)
  T5["Immediate Exit<br/>(no actual interest)"]
  T4["Early Exit<br/>(failure to launch)<br/>very high non-renewal risk"]
  T3["Drifting<br/>(minimal / spotty)"]
  T2["Steady<br/>(lower intensity)"]
  T1["Thriving<br/>(high engagement)"]

  %% Ordering hints (INVISIBLE)
  T5 ~~~ T4
  T4 ~~~ T3
  T3 ~~~ T2
  T2 ~~~ T1

  O1 ~~~ O2
  O2 ~~~ O3

  %% Tie trajectories to the start (conceptual membership)
  A -.-> T5
  A -.-> T4
  A -.-> T3
  A -.-> T2
  A -.-> T1

  %% Trajectories to outcomes (meaningful only)
  T1 -.-> O1

  T2 -.-> O1
  T2 -.-> O2

  T3 -.-> O2
  T3 -.-> O3

  T4 -.-> O3
  T5 -.-> O3

  %% ----------------
  %% Styling
  %% ----------------
  classDef anchor fill:#ffffff,stroke:#111111,stroke-width:2px;
  classDef red fill:#fff5f5,stroke:#c53030,stroke-width:2px;
  classDef orange fill:#fffaf0,stroke:#dd6b20,stroke-width:2px;
  classDef green fill:#f0fff4,stroke:#2f855a,stroke-width:2px;
  classDef invis fill:none,stroke:none,color:transparent;

  %% Apply classes
  class B,C,D anchor;
  class S1,S2,S3,S4,S5 invis;

  class A,T1,O1 green;
  class T3,T2,O2 orange;
  class T5,T4,O3 red;

Exhibit 4.2. FCC License Lifecycle and Engagement Trajectory Frame — the ten-year license term, cohort positions, and conceptual trajectory types

## 5. Research Program Overview The program is organized around two primary workstreams, each serving a distinct set of populations. The workstreams are intentionally staggered, and both are supported by an ongoing expert interview track that runs across the program. Sections 6 through 9 describe each component in full.
Program Start Staggered Overlap Program Integration

WS1 Lane — Current and Former Licensees

Area of focus: Establish a decision-grade baseline for current and former licensees using FCC ULS-based sampling, then convert findings into lifecycle, retention, and reactivation investment guidance.

WS2 Lane — Adjacent, DC, and Optional Benchmark

Area of focus: Prioritize adjacent and practitioner segments, build practical access paths, and produce actionable cross-segment findings with explicit confidence and inference limits; optional general benchmark remains secondary.

WS2 is staggered to start after core WS1 qualitative findings are in hand, then runs in parallel with later WS1 phases.

Exhibit 5.1. Staggered high-level program timeline: WS1 starts first, WS2 begins after core WS1 qualitative input, and both run in controlled overlap toward integrated decision outputs.

**Workstream 1 — The Licensed and Previously-Licensed Amateur Radio Community** WS1 is the first priority workstream. It serves Populations 1 and 2 (currently and previously licensed amateur radio operators) using the FCC Universal Licensing System as the sampling frame. It is the foundation of the entire research program: the community snapshot, the engagement trajectory analysis, and the investment lever findings for the licensed population all flow from this workstream. It begins first and produces the first major substantive deliverables. **Workstream 2 — Adjacent Populations, Optional General Population Benchmark, and the DC Community** WS2 serves Populations 3, 4, and 5: the adjacent and affinity groups, the non-licensed digital communications practitioners, and the optional general population benchmark module. It is staggered to begin after WS1 qualitative findings are in hand, because what we learn about the licensed community's experience — which pathways into the hobby were most effective, which adjacent communities produced the most engaged hams — informs how WS2 is designed and which sub-populations deserve the deepest attention. WS2 recruitment relies on online panels, partner organizations, and targeted community outreach rather than a federal licensing database. **Expert Interviews — A Parallel Thread** Structured interviews with key voices in the amateur radio and digital communications ecosystem run throughout the project. Interviews conducted early in the program serve a design function: they surface practitioner perspectives on the most important questions, identify gaps in the planned scope, and inform the research questions and instrument design for both workstreams before fielding begins. Later in the program, expert interviews enrich the interpretation of survey findings — adding context, nuance, and real-world grounding that survey data alone cannot fully provide. Expert interviews are not a separate workstream with their own fielding cycle; they are a running source of practitioner-level insight that cross-cuts both workstreams. At the Core level, expert interviews are scoped to voices most relevant to the licensed community; at Plus and Max, the program expands to include voices from adjacent communities and the DC world. Section 9 describes the interview program in detail. **Why This Design Works** This design is intentionally staged so ARDC gets reliable early signal, learns fast, and adjusts before major downstream spend. It allows ARDC to front-load qualitative discovery and instrument validation before large-scale fielding, reducing the risk of collecting expensive but low-decision-value data. It allows ARDC to use WS1 learning as an input to WS2 prioritization, so adjacent population work starts with better targeting, better access assumptions, and a clearer confidence posture. It allows ARDC to make decisions at defined gates, with substantive deliverables at each milestone, rather than waiting for one large end-of-project reveal. This design is intentionally adjustable. ARDC can tune both scope and depth within each workstream — for example by narrowing or expanding segment coverage, increasing or reducing methodological rigor in selected modules, or changing analysis depth by decision area — while preserving the overall architecture. If ARDC chooses to extend beyond the initial program, the same design also creates a clean transition path into continuity metrics and follow-on waves, turning one study into a durable, data-driven strategy and decision capability. --- ## 6. Phase 0: Planning & Charter Alignment Every research project that misses the mark does so for the same reason: the people doing the work did not fully understand what the people funding it actually needed until it was too late to change course. Phase 0 exists to prevent that. It is not a kickoff meeting. It is a structured process of genuine alignment that does not close until all the right people have been heard, the hard questions have been answered, and the path forward has been agreed to at the level of the organization that has to live with the results. **Pre-Charter: Listening Before Scoping** Before the formal Charter sessions begin, the process starts with individual 30-minute video conversations with a subset of ARDC board members. Not a presentation. Not a briefing. A listening session: what are you hoping to understand from this research? What decisions are you trying to make? What would success look like, and what would disappointment look like? What concerns do you have about a project of this scope? These conversations serve two purposes. First, they bring the real goals, expectations, and concerns of decision-makers into the room before any scope is fixed, not filtered through project contacts who may summarize, soften, or miss what matters most. Second, they create a record: when the final deliverables are presented, it will be possible to return to each of these goals by name, show what the research produced in response, and have a direct conversation about whether the project delivered. That closing loop is one of the highest-value things a research engagement can produce. **Stage 1 Charter: Aligning on Scope, Values, and WS1 Design** Following the individual discovery conversations, the Charter process moves into a series of working sessions with the full project team: ARDC project contacts, relevant board members, and the research team. Three to four sessions is typical. The first session surfaces the full range of goals and questions on the table; subsequent sessions work through the decisions that require homework, debate, or additional information; the final session closes with affirmative agreement on everything that has to be locked before work begins. The decisions made in Stage 1 are: *Research values.* Before anything else, ARDC and the research team explicitly agree on the values that govern the work: bias management standards, statistical quality thresholds, privacy and data security commitments, PII handling and de-identification protocols, data ownership and licensing terms, and the open data framework. These are not fine print; they are design decisions that affect every choice made downstream. Section 11 describes each of these in detail. *Scope.* The AR/DC boundary is confirmed. The five populations are reviewed and any additions, removals, or priority adjustments are made. The international scope question is answered. If included, general population benchmark parameters are confirmed as directional. *WS1 design brief.* The cross-tabulation and sub-group analysis plan for WS1 is finalized, because it determines sample size and cost. Instrument priorities are documented. The preliminary expert interview list is built. *Participation model.* What ARDC staff will be involved, at what stages, in what roles. Time commitments are made explicit so they are not a surprise later. The Stage 1 Charter process closes with a signed Charter document: a written record of what was agreed, signed by people with authority to commit ARDC's resources and direction. It is the reference document that governs the project and the thing we return to at every decision gate. **Stage 2 Charter: WS2 Alignment (Before WS2 Design Begins)** A focused WS2 scope alignment session occurs later in the program, immediately before WS2 instrument design begins, triggered by the completion of WS1 qualitative findings. At that point, the research team will have learned things from the licensed community that should inform which WS2 sub-populations deserve the deepest attention, which adjacent groups are most likely to be reachable, and what questions deserve priority in the WS2 instruments. Stage 2 is one to two meetings focused on WS2-specific design decisions. Its outputs feed directly into Section 8. Stage 2 has its own sign-off requirement; it is a minor gate (see Section 10). --- ## 7. Workstream 1: The Licensed Amateur Radio Community WS1 is the methodological core of this program. It addresses the populations with the clearest defined access path and the research questions most directly relevant to ARDC's grantmaking decisions. **The FCC ULS Advantage** The FCC Universal Licensing System publishes active and historical license records for all US amateur radio license holders, including name and mailing address. This makes WS1 fundamentally different from any survey that relies on opt-in, convenience sampling, or online panels. It provides a probability sample: every licensed operator in the country is in the frame, and every sampled operator has a known, non-zero probability of being selected. Findings from a probability sample can be generalized to the full population in ways that convenience samples cannot. The practical implication: WS1 can answer "what fraction of all licensed hams are essentially inactive?" with a defensible, population-level answer. A study based on club rosters, online forums, or self-selected participation cannot answer that question, because the people least likely to be active are the most likely to be absent from those channels. The ULS frame includes them all. The survey invitation model is mail-to-web: a physical letter is sent to the ULS mailing address, directing the recipient to a personalized online survey link. This reaches licensees who are not reachable by email, who have not opted into any list, and who have no digital presence in the amateur radio community. Address quality for active licenses is generally current; the FCC requires a valid address as a licensing condition. For Population 2 records (former licensees with expired status), additional data hygiene steps are required: NCOA processing and deduplication across active and expired files. The Charter phase specifies which vintage of expired records to include. **Qualitative Phase First** Before any closed-ended survey question is written, a qualitative phase maps the landscape. The goal is not to produce findings; it is to produce a better survey. WS1 qualitative work uses focus groups (two to three, organized by life stage and engagement level) and individual depth interviews (eight to twelve, targeting very new licensees, long-lapsed operators, and highly active operators across different modes). A qualitative findings report is produced before instrument design begins. **Survey Design** The WS1 main survey is organized around four measurement batteries: *Current engagement state.* Activity level, modes of participation, recent on-air activity, club affiliation, and involvement in specific activities. This produces the community snapshot. *Lifecycle stage and trajectory.* When was the license obtained? What has the arc of engagement looked like? These retrospective questions, combined with the FCC license cohort structure embedded in the sample, allow trajectory pattern distributions to be estimated from cross-sectional data. Appendix G describes the analysis approach in detail. *Adoption and retention drivers.* What drew the respondent into amateur radio? What has sustained or reduced engagement? What role did clubs, mentors, technology, and life changes play? *Perception and messaging.* How does the respondent describe amateur radio and digital communications to people outside the community? **Trajectory Analysis: What It Can and Cannot Deliver** The survey design supports engagement trajectory reconstruction, but the commitment is to try, not to guarantee a result. Retrospective recall is imperfect. The analysis will surface trajectory patterns where the data is clear and note where it is not. The community snapshot is the guaranteed primary output; trajectory analysis is a complementary layer built on top. Appendix G covers the analysis approach; Appendix F covers cognitive testing and pilot protocol. **Cognitive Testing and Pilot Fielding** Before the main survey launches, the instrument undergoes cognitive testing: sessions in which representative respondents think aloud while completing the draft survey, surfacing questions that are misunderstood or confusing. A small-scale pilot then tests the full logistical chain at scale before full fielding begins. Exhibit 7.1 summarizes WS1 as a single, vertically sequenced workstream from Charter through decision outputs. Arrows indicate where WS1 outputs feed downstream decisions and adjacent workstreams.

WS1 Workflow — Current and Former Licensees

1. Charter and Research Design Lock

Finalize scope, sample design assumptions, quality thresholds, and gate criteria for WS1 execution.

2. Qualitative Discovery and Instrument Build

Run focus groups/IDIs, translate findings into measurement batteries, and draft the survey instrument.

Output to WS2 setup: segment and message signals used in Stage 2 alignment.

3. Cognitive Testing and Pilot Gate

Validate comprehension and operational flow, then approve full field launch through a formal gate decision.

4. Full Survey Execution (ULS Mail-to-Web)

Field invitations and reminders, monitor response quality, and manage bias checks during data collection.

5. Analysis, Synthesis, and WS1 Reporting

Produce the community snapshot, trajectory insights, and investment lever guidance for licensed populations.

Output to integration: Decision Kit inputs, deliverable schedule, and board briefing materials.

Exhibit 7.1. WS1 vertical methodology flow, including quality gates and labeled output handoffs to adjacent workstreams and decision integration.

**WS1 Deliverables** - Qualitative Findings Report: themes, vocabulary, and experience range from the qual phase - WS1 Survey Report: full quantitative findings with cross-tabulations, segment profiles, and trajectory analysis - WS1 Decision Kit Component: community snapshot, trajectory analysis summary, and investment lever findings for the licensed community - Open data package (WS1): dataset, codebook, survey instrument (see Section 11 for open data terms) - Repeatable question bank and cohort design documentation (Core tier and above) --- ## 8. Workstream 2: The Digital Communications Community & Adjacent Populations WS2 is designed to inform ARDC's investment decisions outside the licensed-operator core: where adjacent and technical communities are already active, where participation is latent but reachable, and where ARDC funding can increase meaningful experimentation and contribution. The workstream is deliberately segmented. It does not assume one audience, one channel, or one outcome. WS2 is materially different from WS1 in one foundational way: there is no single master list equivalent to the FCC ULS that covers the full adjacent/DC universe. That means WS2 must be built through a structured access strategy across multiple sources -- partner channels, screened panels, and targeted outreach -- with explicit quality checks and a clearly stated confidence posture by segment. WS2 success is not defined as conversion to licensed amateur radio alone. Conversion is one useful metric. The larger success frame is movement along ARDC's learn-experiment-do continuum: increased awareness in the right technical audiences, increased hands-on experimentation and project participation, and stronger intent to continue engagement in AR/DC pathways over time. A participant who goes deeper into SDR, mesh, open-source radio software, or related DC work is a mission-aligned success whether or not that path ends in a ham license. To keep this workstream decision-useful and cost-effective, priority is given to technical-adjacent and technical-practitioner segments (Populations 3 and 4). The general population benchmark (Population 5) is treated as an optional directional module for baseline calibration, not as the center of WS2. WS2 outputs are reported with explicit inference limits and triangulated across segments. The result is an actionable map of where ARDC can invest for practical ecosystem growth without overclaiming representativeness where the sampling frame does not support it. Exhibit 8.1 shows WS2 as a standalone vertical workflow, highlighting where access-path design and inference management differ from WS1.

WS2 Workflow — Adjacent, DC, and Optional Benchmark Populations

1. Segment Prioritization and Access Strategy

Define priority segments and feasible access paths across partner channels, screened panels, and targeted outreach.

2. Source Qualification and Inference Posture

Evaluate coverage, bias risk, and expected confidence by segment; set explicit limits on representativeness claims.

Output to governance: approved WS2 source strategy and confidence assumptions.

3. WS2 Qualitative Input and Instrument Design

Use segment-specific qualitative signal to shape batteries, language, and pathway-focused measures.

4. Multi-Channel Fielding and Quality Monitoring

Execute by segment with channel-level performance tracking, bias checks, and optional benchmark module activation.

5. Cross-Segment Synthesis and WS2 Reporting

Deliver opportunity mapping, message/pathway guidance, and decision recommendations with explicit inference boundaries.

Output to integration: WS2 Decision Kit inputs, partner strategy implications, and continuity metric candidates.

Exhibit 8.1. WS2 vertical workflow, emphasizing fragmented population access, source qualification, and confidence-bounded synthesis distinct from WS1's single-frame design.

**WS2 Deliverables** - WS2 Qualitative Findings Report (adjacent and practitioner segment landscape) - DC Community Report (non-licensed practitioner findings and recommendations) - WS2 Survey Report (cross-segment synthesis with explicit inference limits) - WS2 Decision Kit Component (investment levers, partner pathways, and message guidance) --- ## 9. Expert Interviews The expert interview program is a structured listening track that runs throughout the project. It is not a workstream with its own survey. It does not produce a statistically generalizable set of findings. What it produces is something surveys cannot: direct access to the knowledge, perspective, and judgment of people who have been inside the ecosystem long enough to understand what the data will not fully explain. **Two Functions, Two Timings** Interviews conducted during Phase 0 and the early qualitative phases serve a design function. These conversations help the research team calibrate: Are the research questions aimed at the right targets? Are there aspects of the amateur radio or digital communications world that an outside researcher would miss? Is the population structure missing anyone important? Expert input at this stage, before instrument design is locked, produces better surveys. Interviews conducted after WS1 and WS2 survey findings are available serve an interpretation function. Survey data shows what. Expert conversations help explain why, and help distinguish between explanations that seem plausible on paper and explanations that resonate with people who know the community from the inside. **Who Gets Interviewed** The preliminary expert interview list is built in the Charter phase. The pool typically includes: At the Core scope: VE coordinators and exam session organizers, active club leaders with visibility into membership patterns, ARRL national staff and section leadership with perspective on licensee engagement and retention, and ARDC grantees with relevant project experience. At Plus and Max scope: voices from adjacent communities (STEM educators, maker community organizers, emergency communications program leaders), international ham radio organization representatives (IARU, national societies in key markets), and technical leaders from the open-source digital communications world. The list is not fixed. Interviews surface additional names; promising leads are followed where schedules and relevance allow. **Interview Format and Deliverables** Interviews are 30 to 45 minutes, conducted remotely, and semi-structured: a prepared set of topics and questions, but run as a conversation rather than a script. The goal is to surface non-obvious perspectives. Interviewees are not asked to ratify findings; they are asked to challenge, contextualize, and explain. Each interview produces a concise individual summary: key themes, notable quotes (with permission), and learnings relevant to the research program. These summaries are the primary deliverable, a running record of practitioner perspective that accumulates throughout the project. The Expert Interview Synthesis, delivered after the main survey phases, draws patterns across conversations and highlights recurring themes, notable disagreements, and insights that enrich the survey findings. It is a readable document organized around key learnings, not a heavy analytical product. **Integration into the Decision Kit** Expert interview themes are woven into the relevant sections of the Decision Kit, not appended separately. Where a survey finding is reinforced by expert consensus, that is noted. Where experts offer explanations that the survey does not reach, those are included as interpretive context. The synthesis is also published as a standalone document for readers who want the practitioner perspective without the full quantitative report. --- ## 10. Governance & Decision Gates Research projects that go wrong typically go wrong slowly: a scope question left unresolved becomes an assumption, the assumption becomes a design choice, the design choice is never reviewed, and by the time the problem is visible it is expensive to fix. Decision gates are the mechanism that prevents this. They create defined moments when the project pauses, the work completed is reviewed, and the plan for what comes next is confirmed before resources are committed to it. **Two Levels of Gates** Not all decisions require the same level of organizational attention. The governance structure uses two gate types: *Major gates* occur at the boundaries between the project's primary phases: end of the Charter phase, end of each workstream's survey phase, and before the integrated Decision Kit is finalized. Major gates require participation from the same level of the organization that signed the Charter -- project sponsors and relevant board members. The review covers the phase deliverable, the proposed design for the next phase, and any material scope or cost questions. Approval at a major gate is affirmative, not passive: the project does not proceed until sign-off is given. *Minor gates* occur within phases at the boundaries between sub-phases: qualitative complete and instrument design approved; pilot results reviewed and full fielding authorized; WS2 Charter session complete and WS2 design brief signed. Minor gates are reviewed by ARDC project contacts without requiring board-level participation. They keep the project moving without unnecessary overhead while maintaining a clear record of what was approved at each step. *Exhibit 10.1: Major/Minor Gate Structure -- phases, gate types, participation requirements, and sign-off flow (to be developed in production)* **What Happens at a Gate Review** Every gate review involves three items: the deliverable from the phase just completed, the proposed plan for the phase about to begin, and any open questions that require a decision before work proceeds. Nothing is presented as a fait accompli. If the design proposed for the next phase raises concerns, those concerns are addressed before fielding begins, not after. **Escalation Path** Questions that arise mid-phase -- a finding that changes the scope, a technical issue that affects cost, a new priority raised by ARDC -- are escalated to project contacts without waiting for the next gate. The project contacts decide whether the question requires a full gate review or can be resolved with a documented email exchange. Material changes always trigger a formal gate; minor adjustments do not. **Time Commitment** Board and sponsor participation in major gates requires approximately three to four structured sessions over the life of the project (Charter approval, WS1 survey review, WS2 survey review, Decision Kit review). Each session is planned for 60 to 90 minutes. Project contacts participate more frequently -- minor gate reviews and regular check-ins -- typically one brief session every two to three weeks during active phases. --- ## 11. Methodology Integrity: Validity & Bias Controls The value of this research depends on whether it can be trusted. Trust, in research, is not a feeling -- it is a property of design. This section describes the methodological standards and practices that make the findings from this project defensible to a skeptical audience, and the research values that are agreed to at the Charter phase and honored throughout. **Probability Sampling: Why It Matters** Most survey research about the amateur radio community is based on convenience samples: people who happen to see an online survey, members of a particular club or forum, respondents from a self-selected panel. These studies produce data, but they cannot tell you how representative the respondents are. The people most likely to respond to an online survey about amateur radio are the people most interested in amateur radio, which means the disengaged, the lapsed, and the curious outsiders are systematically underrepresented. WS1 uses a probability sample drawn from the FCC Universal Licensing System. Every licensed operator in the US is in the frame. The selection is random within defined strata. Response rates are measured and monitored. Non-response bias is assessed through late-responder analysis and, where feasible, comparison to available FCC administrative data. This is what allows WS1 findings to be stated as population-level estimates rather than sample observations. WS2 does not have the same probability sampling advantage -- there is no federal database of STEM educators, DC practitioners, or emergency preparedness volunteers. WS2 findings are reported with appropriate acknowledgment of their sampling limitations, and claims about representativeness are scoped accordingly. **Non-Response Bias Management** No survey achieves a 100% response rate. The question is whether the people who respond are systematically different from the people who do not. WS1 manages non-response through multi-touch invitation design (initial invitation, two reminder mailings), late-responder analysis (comparing early and late respondents as a proxy for non-response patterns), and post-stratification weighting where appropriate. Appendices B and C cover the sampling accuracy and bias management protocols in detail. **Instrument Design Standards** Questions are written to neutral-wording standards: no leading language, no loaded terms, no response options that suggest a preferred answer. Scales are balanced. The instrument is reviewed for question order effects before cognitive testing. Cognitive testing and pilot fielding (Appendix F) validate the instrument before full-scale deployment. **Research Values: What ARDC and the Research Team Agree To** These commitments are made explicit in the Charter phase and govern every downstream decision: *Bias management.* The design prioritizes finding the truth over confirming a hypothesis. No sponsor organization or advocacy position drives question wording or analysis. *Statistical quality.* Findings are reported with appropriate precision. Differences that are not statistically significant are not reported as meaningful. Margin of error and confidence interval conventions are documented; Appendix B covers the sampling accuracy framework in detail. *Privacy and data security.* Survey responses are collected with informed consent. PII is separated from survey responses at the point of data ingestion and is not included in the analytical dataset. Data storage and access controls meet current professional standards. *Data ownership.* Survey data collected through this project is owned by ARDC. The research team retains no independent rights to use the data for other purposes without ARDC's explicit permission. *Open data.* After project close, ARDC publishes the analytical dataset, the data codebook, and the survey instruments without restriction. Third parties can replicate, extend, or challenge the analysis. Proprietary analytical methods developed by the research team are not automatically included in the open release; their treatment is agreed in the Charter. The goal of open data is reproducibility and community trust, not complete disclosure of every input. **Appendices as Technical Foundation** The appendices to this proposal are not supplementary reading -- they are the technical documentation for the methodological commitments made here: - Appendix A: Questions and Measurement Map - Appendix B: Sampling Accuracy Framework - Appendix C: Bias Management Protocol - Appendix D: Retention Modules - Appendix E: Draft Survey Instruments - Appendix F: Cognitive Testing and Pilot Protocol - Appendix G: Analysis Plan - Appendix H: Reporting Templates - Appendix I: Fielding Options and Partnering Approach - Appendix J: Decision Registry (and Change Log) --- ## 12. Deliverables: What ARDC Receives ARDC does not wait until the end of the project to receive findings. Each completed phase produces a deliverable. By the time the final integrated Decision Kit is delivered, most of the findings within it will already have been seen in earlier interim reports. **Deliverable Inventory** *Phase 0:* Charter Document -- signed record of all scope, values, participation, and design decisions; the governing reference for the project. *WS1 Qualitative Phase:* WS1 Qualitative Findings Report -- themes, vocabulary, experience range, and attitudinal landscape from focus groups and IDIs; input to instrument design. *WS1 Quantitative Phase:* WS1 Survey Report -- full quantitative findings with cross-tabulations, segment profiles, trajectory analysis, and annotated tables. WS1 Decision Kit Component -- the licensed community section of the final Decision Kit, delivered as a standalone for Core-tier engagements. *Expert Interviews (running throughout):* Individual Interview Summaries -- delivered on a rolling basis as interviews are completed. Expert Interview Synthesis -- produced after the main survey phases; patterns, recurring themes, and practitioner-level insight that enriches the survey findings. *WS2 Qualitative Phase:* WS2 Qualitative Findings Report -- themes and landscape for adjacent and DC populations; optional directional baseline module for general population if included. *WS2 Quantitative Phase:* DC Community Report -- findings and recommendations for the DC practitioner population, delivered first within WS2. WS2 Survey Report -- full findings across Populations 3, 4, and 5. *Integrated Final Deliverables:* Decision Kit -- the integrated synthesis of all workstream and expert interview findings (see below). Open Data Package -- analytical dataset, codebook, survey instruments, and documentation; published by ARDC after project close. Repeatable Question Bank and Cohort Design Documentation (Core tier and above). **The Decision Kit** The Decision Kit is not a research report formatted as a slide deck. It is organized around the decisions ARDC needs to make: - Community snapshot: who the licensed community is, how engagement is distributed, segment sizing and characteristics - Trajectory analysis: the engagement journey distribution, what drives each trajectory, where the intervention-relevant moments are - Adjacent community profiles: who is most reachable from outside the licensed community, what motivates each group, what paths into the community exist - Messaging guidance by audience: what framing resonates with each population, what vocabulary works, what to avoid -- organized by audience type so it is directly usable for communications planning - Investment lever analysis: where ARDC investment has the most leverage, organized by decision type (programs, grants, community infrastructure, communications) with supporting evidence - Cross-workstream synthesis: the integrated picture across all populations and the expert interviews -- patterns, tensions, and strategic implications The Decision Kit is delivered as both a written report and a structured document optimized for use in meetings and planning sessions. It is designed to be consulted repeatedly, not read once and filed. **Open Data Package** After project close, ARDC publishes the analytical dataset, data codebook, and survey instruments without restriction. Anyone can replicate, extend, or challenge the analysis. Open data is a signal of quality, not just a transparency commitment: research that can be checked is research that can be trusted. Proprietary analytical methods are handled per the Charter terms. **Tier-Differentiated Deliverable Map** Core delivers: Charter document, WS1 qualitative and quantitative reports, WS1 Decision Kit component, expert interview summaries, open data package (WS1), repeatable question bank. Plus adds: WS2 qualitative and quantitative reports, DC Community Report, full Expert Interview Synthesis, integrated Decision Kit (all populations), open data package (WS2). Max adds: Interactive data dashboard, formal longitudinal design documentation, Wave 2 plan and cost estimate. *Exhibit 12.1: Deliverable Schedule -- all outputs mapped across program phases and timeline (to be developed in production)* --- ## 13. Building a Data-Driven Decision Culture This project is designed to produce a body of findings. But its more durable output is organizational: a set of capabilities that change how ARDC makes decisions long after the final report has been delivered. **The Three Capabilities** Section 2 named three organizational capabilities this engagement is designed to build. This section explains what those capabilities look like in practice. *Evidence-based grant evaluation.* ARDC currently evaluates grant proposals against criteria that are largely internal -- staff judgment, precedent, and mission alignment. This research adds a third layer: demonstrated gaps in the ecosystem. A proposal that addresses a gap the research identifies as large and underserved is a different kind of proposal than one that addresses a segment the research shows is already well-resourced. The research does not make decisions for ARDC. It makes the basis for decisions transparent and defensible. The baseline established in this project also makes impact measurement possible for the first time. If ARDC funds a set of onboarding programs over the next three years, the question "did it work?" has an answer only if there is a before-picture to compare against. Wave 2 of the research provides the after-picture. Without this baseline, there is no way to distinguish programs that moved the needle from programs that spent money without effect. *Proactive grantmaking.* The research identifies specific gaps: populations that are reachable but not being reached, lifecycle stages where disengagement is concentrated, communities with strong interest and no entry point, technology barriers that are consistently present when engagement falters. These gaps are actionable. ARDC can design grant initiatives around them, approach potential grantees with a specific brief, and fund in areas where the evidence shows the need -- rather than evaluating only what arrives in the inbox. This is a meaningful shift in how a grant-making organization operates. Most foundations are primarily reactive: they set criteria, publish a program, and evaluate what comes in. The research makes a proactive posture possible -- not by replacing the inbound grant process, but by supplementing it with a strategic investment agenda grounded in data. *Authoritative public presence.* ARDC is not currently the organization that outside audiences turn to for authoritative information about the state of the amateur radio and digital communications community. This research changes that. When ARDC publishes the findings -- the community snapshot, the trajectory analysis, the segment profiles, the messaging guidance -- it becomes the organization that produced the definitive data on this community. That positioning has compounding value: it builds trust with community organizations, signals seriousness to potential grantees and partners, and supports ARDC's credibility as a decision-making body with the board and external stakeholders. **The Path Forward** The research program as described in this proposal is a first step in a longer trajectory. The community snapshot and trajectory analysis produced by Core or Plus are genuinely valuable on their own. But their greatest value is as the foundation for what comes next: Wave 2, which measures whether the picture changed; cohort tracking, which follows real people through real engagement journeys; and the kind of cumulative knowledge that only comes from sustained, rigorous attention to the same community over time. ARDC has the opportunity to build that. No other organization in this space has made the commitment to systematic, longitudinal, population-level research about the amateur radio and digital communications community. Over time, this also positions ARDC as a trusted beacon of credible information on amateur radio participation trends, not only a source of funding. This proposal is how that commitment begins. --- ## 14. Scope & Program Options > The research program described in this proposal is designed to be executed in tiers. ARDC can choose the scope that fits its current appetite and budget, with each tier designed to stand on its own — and to connect cleanly to the next tier if ARDC chooses to extend. The tier structure reflects a deliberate logic: each level answers the most important questions first, in the order that matters most. The question "why is it broken up this way?" has a straightforward answer for each boundary. --- **Core — Know Your Community (Core Program)** *Who it covers:* Licensed and previously-licensed amateur radio operators (Populations 1 and 2). WS1 only. *What it produces:* The foundational picture of the licensed amateur radio community — who is in it, how engaged they are, what drew them in, and what has sustained or ended that engagement. This includes the community snapshot, the engagement trajectory analysis (to the extent the data supports it), and the WS1 component of the Decision Kit. The Core program also delivers a repeatable question bank and cohort structure designed for future use, so that even a one-time study plants the seeds of longitudinal tracking. *Why this is the right starting point:* The licensed community is the most clearly defined, most directly reachable, and most immediately relevant population for ARDC's grantmaking decisions. The FCC ULS gives this workstream a probability sample — a structural advantage no other approach can match. If ARDC understands only one thing from this entire engagement, it should be the shape of the licensed community it is trying to serve. Everything else builds on this foundation. *Expert interviews:* A targeted set of expert interviews is included in the Core program, scoped to voices most relevant to understanding the licensed community — VE coordinators, club leaders, licensing exam educators, and ARRL contacts with visibility into engagement and retention. *Deliverables:* Charter document; Qualitative Findings Report; WS1 Survey Report; WS1 Decision Kit component (community snapshot, trajectory analysis, investment lever findings for the licensed community); dataset and codebook (open); repeatable question bank and cohort design document. --- **Plus — See the Whole Ecosystem (Technical Depth Expansion)** *Who it adds:* Adjacent and affinity populations, digital communications practitioners, and the optional general population benchmark module (Populations 3, 4, and 5). WS2 added to Core. *What it produces:* Everything in Core, plus a complete picture of the communities surrounding licensed amateur radio — who is adjacent, what they know, and what would move them. Priority is on technical-adjacent and practitioner segments where ARDC can act directly. An optional directional general population benchmark can be included for calibration where useful. A fuller expert interview program is included, extending to voices from the adjacent communities and the digital communications world. The integrated cross-workstream analysis in the Decision Kit is what makes Plus meaningfully different from two separate Core studies: the findings are synthesized together to show how the pieces of the ecosystem relate. *Why Plus makes sense after Core:* Once ARDC understands its licensed community, the natural next question is: "Who is outside it, and how do we reach them?" Plus answers that — and it answers it in the context of what WS1 already revealed, which is why WS2 is staggered rather than parallel. *Deliverables:* Everything in Core, plus WS2 Survey Report; expanded Expert Interview Synthesis; integrated Decision Kit (full cross-workstream analysis, segment profiles across all five populations, messaging guidance by audience, investment lever analysis spanning licensed and non-licensed communities); full dataset and codebook (open). --- **Max — Build the Machine (Longitudinal Extension)** *What it adds to Plus:* Longitudinal infrastructure, interactive data dashboard, and Wave 2 planning. *What it produces:* Everything in Plus, plus the capability to do this research again — better, faster, and with the ability to track change over time. The Max tier is not primarily about more research now; it is about making this research the foundation of an ongoing, self-improving intelligence function for ARDC. Specifically, Max adds: *Longitudinal design.* The repeatable question bank built in Core is formalized and documented as a fielding-ready instrument for future waves. Cohort identifiers are maintained for participants who consent to re-contact. The design is structured so that Wave 2 (12 to 18 months later) can directly compare to Wave 1 and measure whether the community picture is changing. *Interactive data dashboard.* A web-accessible tool allowing ARDC staff to explore findings by segment, geography, license class, and other dimensions without requiring custom analysis requests for each question. *Wave 2 planning.* A formal plan and cost estimate for a follow-on research wave, ready to execute when ARDC chooses to proceed. *Why Max matters:* The Core and Plus programs answer "what is the situation now?" Max answers "are we making it better?" — and answers it with evidence. This is the research infrastructure that underlies ARDC's third organizational capability (described in Section 2): becoming the authoritative, data-grounded knowledge source on the state of the amateur radio and digital communications community, with the ability to update that knowledge as the community evolves. *Deliverables:* Everything in Plus, plus formal longitudinal design documentation; interactive dashboard; Wave 2 plan and cost estimate. --- **The True Longitudinal Vision: Where This Is Heading** The Max tier creates the infrastructure for ongoing research. But it is worth naming explicitly where that path leads over a longer horizon. True longitudinal panel research — following the same individuals over years, tracking their actual engagement trajectory from newly-licensed through renewal, lapse, or re-engagement — is the research design that would most powerfully answer the questions this proposal raises. It requires recruiting a willing panel of participants, re-contacting them at defined intervals (12 months, 24 months, at renewal time), and tracking real behavior rather than retrospective recall. It is the difference between asking someone "how did you get here?" and watching them travel. This proposal does not budget for a multi-year panel study. What it does — deliberately — is design the initial research program to make true longitudinal work achievable as a follow-on. The cohort identifiers, the repeatable question bank, the consent structure that allows re-contact, and the Wave 2 plan are all designed with that destination in mind. An ARDC that executes Core today, extends to Plus in the next planning cycle, and launches a true cohort panel in the third cycle will have built, step by step, one of the most sophisticated ongoing research programs in the amateur radio and digital communications world. That is the trajectory. This proposal is the first step. --- **Optional Add-Ons** The following are available as additions to any tier: *International population extension.* Non-US populations added to WS1 or WS2 through international licensing authority lists or extended online panels. Scope and cost determined during Charter phase. *Board presentation deck.* Executive-level slide presentation of key findings, designed for ARDC board use. Adds production work beyond the standard written report. *Additional geographic or demographic sub-sampling.* Oversampling of specific regions, license classes, or demographic groups to support analysis at finer resolution than the standard sample supports. --- **Summary** | | Core | Plus | Max | |---|---|---|---| | Licensed community (WS1) | Yes | Yes | Yes | | Adjacent communities and DC practitioners (WS2) | No | Yes | Yes | | General population benchmark (directional, optional) | No | Optional | Optional | | Expert interviews (targeted) | Yes | Yes (expanded) | Yes (expanded) | | Integrated cross-workstream analysis | No | Yes | Yes | | Interactive data dashboard | No | No | Yes | | Longitudinal design documentation | Seed | Seed | Full | | Wave 2 plan | No | No | Yes | | Open data (dataset, codebook, instruments) | Yes | Yes | Yes | --- ## 15. About Jim Idelson Jim Idelson (K1IR) brings a combination to this project that is difficult to replicate: long-form market research leadership experience and deep lived experience inside the amateur radio community. From the research side, Jim has led complex, mixed-method market research programs that combine qualitative discovery, survey instrument design, pilot testing, broad fielding, and decision-focused synthesis. His approach is not to deliver a report and leave. It is to build a decision system: clear objectives, explicit assumptions, method discipline, interim decision gates, and deliverables that map directly to the choices sponsors actually need to make. From the community side, Jim is not an external observer writing about amateur radio from a distance. He is an active participant in the AR/DC world as an operator and technologist, with long-standing relationships across organizations, practitioners, and project leaders. That perspective matters in practical ways: better question design, better interpretation of findings, better judgment about what is signal versus noise, and better ability to engage expert interviewees who might not open up to a generalist researcher. This proposal was developed from first principles around ARDC's goals, then refined using prior proposal patterns and external specialist input where appropriate (for example, field operations and survey execution options). AI-assisted drafting tools were used as writing and synthesis support, while project framing, method choices, and final judgment remained with Jim. For an engineering-oriented audience, this is the correct transparency posture: use tools where they increase speed and clarity, but keep accountability for research design and conclusions with the responsible lead. The practical value for ARDC is straightforward: one accountable lead who understands both the method and the domain, can run a disciplined research process, and can translate results into actionable funding decisions without losing context. --- ## 16. Timeline & Investment This section presents execution ranges and pricing logic designed for board-level decision-making. Contracted scope and pricing are expected to be finalized before project start. Charter then refines implementation detail -- especially sequence, pacing, and milestone timing -- within the agreed contractual scope. ### Timeline Framework The program is staged to deliver useful outputs early, not only at close. - **Phase 0 (Charter and design lock):** 3 to 5 weeks - **WS1 qualitative + instrument + pilot:** 6 to 10 weeks - **WS1 quantitative fielding + analysis:** 8 to 12 weeks - **WS2 design and fielding (Plus/Max):** 8 to 14 weeks, partially overlapping with late WS1 tasks - **Integrated synthesis and Decision Kit:** 4 to 6 weeks Typical elapsed duration by tier: - **Core:** approximately 5 to 7 months - **Plus:** approximately 7 to 9 months - **Max:** approximately 9 to 12 months (including longitudinal and dashboard design work) ### Investment Logic Investment is governed by a small number of cost drivers, not by arbitrary package labels: - target sample sizes and required sub-group cell sizes - number of invitation waves and reminder cycles (WS1 mail-to-web) - panel sourcing requirements and screening complexity (WS2) - incentive model and expected response-rate support level - instrument complexity, branching depth, and pilot iteration count - analysis depth, cross-tabulation breadth, and synthesis scope - optional deliverables (dashboard, Wave 2 plan depth, international extension) ### Board Calibration and Contracting Path ARDC can use this proposal to select an initial scope and authorize a contract before work begins. The practical sequence is: - choose a tier (Core, Plus, or Max) and any approved add-ons - finalize contracted pricing and payment terms for that scope - begin Charter to lock detailed execution design and adjust timeline sequencing as needed This keeps governance and commercial terms clean: ARDC enters with clear cost commitments, and the research team retains flexibility to optimize pace and sequencing based on Charter decisions and field realities. ### Tiered Investment Posture - **Core (Core Program):** WS1-centric evidence base and decision kit component, with targeted expert interviews - **Plus (Technical Depth Expansion):** adds WS2 coverage and integrated cross-workstream Decision Kit - **Max (Longitudinal Extension):** adds repeat-wave infrastructure, formal longitudinal design package, and dashboard capability Detailed budget values are finalized in contracting prior to kickoff, then tied to the agreed tier and add-on scope. Charter confirms the implementation plan and timing assumptions behind that contracted scope so execution can begin with both commercial clarity and methodological discipline. --- *This document is being developed on the `dev` branch. See [At-a-Glance](At-a-Glance.md) for a high-level summary as it develops.*