Contextual
Intelligence in
Adaptive Wellness
I. The Problem
The dominant question in the digital wellbeing literature is: do nudges work? That question is too blunt. It treats nudges as a monolithic treatment, treats students as a monolithic population, and treats engagement as the outcome that matters. The result is research that can tell you average effects exist — usually small ones — while telling you almost nothing about why they exist, for whom, or under what conditions.
Journey’s nine-month university pilot produced a directional signal strong enough to suggest something more precise is happening. Students did not respond uniformly. Timing appeared to matter more than content. Personalization functioned as a prerequisite, not a feature. Awareness increased before behavior changed. And the system’s scarcity constraint — only three nudges per week — forced every delivery decision to be a selection decision, not a scheduling decision.
That pattern raises a different question: what makes a nudge land? Not whether nudges work on average, but what specific combination of timing, context, personalization, and state-awareness turns a text message into a moment of genuine reflective interruption.
What most research asks
- Did the intervention group improve more than control?
- Is there a main effect of “nudging” on wellbeing?
- Do students who receive messages report less screen time?
- Is digital wellness messaging effective?
What this research asks
- Why did some nudges land and others get ignored?
- Does timing explain more variance than content?
- Which students benefit and which disengage — and why?
- Does the adaptive loop create lasting change or only short-term reaction?
II. The Central Argument
Anind Dey defined context as “any information that can be used to characterize the situation of an entity.” In computing, context-awareness means using that information to provide relevant services. Journey extends that definition into behavioral intervention: context-awareness means using the student’s actual situation — their schedule, cognitive load, energy state, time of day, and likely emotional condition — to determine not just when to send a nudge, but what kind of nudge to send.
This is a stronger claim than “personalization helps.” It is a claim about mechanism. The hypothesis is that timing and context do more work than content in determining whether a nudge produces reflective interruption or gets dismissed. A perfectly worded nudge delivered at the wrong moment is noise. A simple question delivered at the moment of actual decision is an intervention.
Scarcity converts the timing engine from a scheduler into a selection engine. It forces the system to look for moments when an intervention is both possible (the student has a real break) and needed (the student’s likely state would benefit from a particular kind of support). The aim is not productivity. It is helping the student manage life holistically — across academics, stress, health, and connection.
III. The Context Engine
Journey’s context engine draws on Dey’s four primary context categories — identity, location, activity, and time — and extends them with relational context and inferred internal state. The system does not send nudges and then hope they are relevant. It evaluates the student’s likely situation and selects the nudge type accordingly.
The Timing Logic
The system does not distribute nudges evenly across the week. It identifies specific windows where an intervention is both available (the student has a real break) and needed (the student’s likely state would benefit from a particular kind of support).
Before a class begins, a brief orienting prompt helps a student arrive mentally, not just physically. The nudge sets focus before the demand hits.
After class, the student either consolidates or dissipates. A reflection prompt catches learning while it is still warm.
After two or three back-to-back classes, the right nudge is not academic. It is physiological: stretch, walk, water, breathe. The system should know the difference.
A three-hour gap is an opportunity. Depending on the student’s goals and state, it could be a focus window or a connection window.
IV. The Tuesday Problem
Consider a student with three consecutive classes on Tuesday, ending around 1:30 PM, followed by a gap before a 4 PM class. A generic system might send a focus reminder Monday morning because Monday feels like the “start” of the week. A contextually intelligent system recognizes that Tuesday afternoon — after cognitive depletion, before another demand — is the highest-leverage moment.
The nudge that lands there should not say “remember to study.” It should recognize that after three back-to-back classes, the student is likely drained. The most helpful intervention may be a recovery prompt: a walk, a stretch, a glass of water, a few minutes outside. That is not a concession. That is the system working correctly. It is selecting the intervention that matches the student’s likely nervous-system state, not their generic goal list.
Because the system only sends three nudges per week, it has to be highly selective. It should look for the moments when an intervention is both possible and needed: when a student has a real break, when stress is likely peaking, when energy is dipping, or when a small action could meaningfully improve focus, recovery, connection, or wellbeing. The aim is not just productivity, but helping the student manage life holistically.
V. The Adaptive Wellness Model
Journey is not a messaging tool that happens to be personalized. It is an adaptive feedback system. The student enters an initial state. The system assesses that state. Nudges are delivered based on context. The student’s responses — and non-responses — feed back into the system. Over time, the system should learn and the student should grow. That adaptive loop is the novel research story.
That adaptive loop maps directly onto Anjala’s framing of the research as feedback-control analysis. The initial survey establishes the baseline state. The context engine produces the intervention. The response data and follow-up surveys measure the output. The novel claim is that the system’s intelligence lies not in any single nudge, but in the loop itself — the ability to start from a student’s declared state, intervene at contextually selected moments, observe the response, and adjust.
VI. The Illustrative Case
Before Journey was tested at scale, sixteen coaching transcripts from a single beta user produced 123 nudges and revealed three patterns that became foundational to the system’s design. These patterns illustrate, at the individual level, the mechanisms the larger study is designed to test at the population level.
The highest-performing nudges were not expert advice. They were the user’s own words — metaphors, self-descriptions, moments of clarity — reflected back during moments of decision. This suggests that what makes a nudge land is not its cleverness but its recognition: the student sees themselves in it.
Certain exchanges produced visible shifts: changes in wording, energy, or orientation. These inflection points occurred when the coaching surfaced a tension the person had not yet articulated. The system learned to create the conditions for insight rather than to deliver insight directly.
A recurring distinction emerged between “should” language (obligation, external pressure) and “pull” language (attraction, intrinsic motivation). When nudges were reframed from the obligation register to the agency register, engagement changed. This became a design rule: orient toward positive agency, not deficiency correction.
These patterns matter for the larger research because they suggest testable mechanisms. Mirror material tests whether self-referential content outperforms generic content. Inflection points test whether productive instability outperforms resolution. Obligation vs. agency language tests whether framing moderates engagement. Each pattern derived from n=1 becomes a hypothesis testable at n=74 and beyond.
VII. Research Questions
These questions are organized hierarchically. The first is the central question. The remaining four collectively build the explanatory model: not “prove the product works,” but “build a serious model of what is happening, for whom, why, and over what time horizon.”
This is the longitudinal question. It requires before-and-after measurement using comparable instruments. It is the question Anjala most clearly wants answered: does the initial state actually change, or does the intervention produce momentary compliance that fades when nudges stop?
This tests the core claim of the contextual intelligence model: that when you intervene matters as much or more than what you say. It requires comparing timing-matched vs. randomly timed nudges, or modeling timing features as predictors of response.
This tests whether the intervention works differently for different people, and whether those differences are predictable from intake data. It also addresses the critical sub-question: why do some students disengage, and what does that reveal about the intervention’s limits?
This is the question that prevents the paper from becoming a black-box evaluation. It tests whether observed effects can be attributed to specific ingredients.
This connects the awareness-first model to outcomes universities care about. It also tests the theoretical claim that awareness is upstream of behavior: if you increase a student’s capacity to notice and choose, downstream improvements follow.
VIII. Study Design
- What
- Map the incoming student landscape before any intervention. Analyze initial survey data to identify correlations among goals, pain points, wellbeing, digital behavior, and demographic segments.
- Protocol
- Correlational analysis across intake variables. Cluster analysis to identify natural student segments. Path diagram showing baseline relationships among variables.
- Measurement
- Self-selected goals, stated challenges, wellbeing (ACIP / Healthy Minds Index), adapted QoL instrument, digital awareness items, demographic variables.
- Output
- A baseline model: what the world looks like before Journey does anything. No causal claims. The status quo, clearly mapped.
- What
- After a semester of adaptive nudging, test what changes, for whom, and through what mechanisms.
- Protocol
- Comparable follow-up survey administered before finals. Engagement segmentation from SMS data. Before-after comparison with paired change scores. Mechanism modeling: timing-match, personalization, and agency as mediators. Context features as predictors of individual nudge response.
- Measurement
- All Study 1 instruments repeated, plus: engagement profile (read rate, action rate, latency), timing-match score (system-logged), perceived personalization, perceived agency, self-compassion (abbreviated SCS), advisor notes (structured form).
- Output
- The core research contribution: does the adaptive loop produce durable change? For whom? Through what mechanisms?
The Correlation vs. Causation Problem
Anjala’s most persistent concern is the directionality question: are students with better wellbeing simply more likely to respond to nudges, or do nudges actually improve wellbeing? This is not a minor methodological footnote. It is the central threat to the study’s credibility. A paper that accidentally implies causation from correlational data will fail with skeptical readers.
The design addresses this at multiple levels. The before-and-after structure allows person-level change scores. Engagement segmentation allows trajectory comparison across groups. Mechanism variables test whether specific features predict change. Future iterations will incorporate micro-randomized components enabling causal claims about specific features.
The paper should state this clearly. The strength is in the explanatory model, not the causal claim. That is what makes it publishable: honest about limits, rich in mechanism.
IX. Variables & Measurement
| Variable | Type | Instrument / Source | When |
|---|---|---|---|
| Wellbeing (ACIP) | Outcome | Healthy Minds Index (17 items) | Baseline + End |
| Quality of Life | Outcome | Adapted college-student QoL (Krishen instrument, 5-pt, 5–7 items) | Baseline + End |
| Digital Awareness | Outcome | Custom: noticing compulsive use, interrupting patterns, values alignment | Baseline + End + Pulse |
| Doom Scrolling | Outcome | Self-report frequency + optional iOS Screen Time donation | Baseline + End |
| Engagement Profile | Moderator | SMS data: read rate, action rate, latency, reply patterns | Continuous (logged) |
| Timing Match | Mediator | System-generated: was nudge in high-leverage window? | Per-nudge (logged) |
| Perceived Personalization | Mediator | Survey: “The nudges understood my situation” | End + Pulse |
| Perceived Agency | Mediator | Survey: “I feel more able to direct my digital behavior” | Baseline + End |
| Schedule Density | Context | Derived: consecutive classes, seated min, gap length | Semester-level |
| Self-Compassion | Mediator | Self-Compassion Scale (abbreviated) | Baseline + End |
| Goals / Pain Points | Baseline | Journey initial assessment | Intake |
| Advisor Notes | Qualitative | Structured follow-up form (coded + open-ended) | Per check-in |
X. Engagement Segmentation
One of Anjala’s clearest analytical instincts is that the research gets much stronger when it stops treating students as one undifferentiated group. Different students respond to nudges differently for intelligible reasons. The study should identify those segments and test whether outcomes differ across them.
High read rate, high action rate, consistent engagement. The system works as designed. Were they already doing well, or did the intervention move them?
Low initial engagement, increasing over time. The most interesting mechanism story: what changed? Did trust develop gradually? Did circumstances shift?
Enrolled but disengaged. Not hostile. Notification fatigue? Anti-AI sentiment? Wrong tone? Wrong timing? This group is as important as the responders.
Flagged by initial assessment or advisor observation. Does the intervention help those who need it most, or primarily serve the already-capable?
The avoider segment is not a data loss. It is a research finding. Understanding why some students do not engage — and whether their reasons are addressable — prevents the study from reading as one-sided or promotional. It also generates the most interesting questions for intervention refinement.
XI. Hypotheses
Falsifiable if: timing-match scores do not predict engagement or outcomes.
Falsifiable if: behavior change occurs without awareness change, or awareness change does not predict subsequent behavior change.
Falsifiable if: segments do not differ on outcomes, or baseline variables do not predict segment membership.
Falsifiable if: improvements are concentrated in early weeks and decay, or post-intervention follow-up shows full regression to baseline.
XII. Study Protocol
Phase 1: Baseline Analysis (Now)
Analyze existing initial survey data. Map correlations among goals, pain points, wellbeing indicators, and student characteristics. Identify natural clusters. Build the baseline model. Publish as Study 1.
Phase 2: End-of-Semester Survey (Before Exams)
Administer follow-up survey using comparable items: adapted QoL instrument, Healthy Minds Index, digital awareness items, perceived personalization, perceived agency. Deploy before students mentally check out.
Phase 3: Engagement Segmentation
Classify students into responder / late-responder / avoider / high-risk segments using SMS engagement data. Compare baseline characteristics across segments.
Phase 4: Longitudinal Analysis
Compare baseline vs. follow-up on all outcome variables. Test whether changes differ by segment. Model timing, personalization, and context-match as predictors. Core of Study 2.
Phase 5: Qualitative Integration
Code advisor notes using structured categories. Integrate as supporting evidence for the mechanism story. Create follow-up form for future cohorts.
- What
- Establish causal effects of specific nudge features by randomizing components (type, timing, tone, question vs. directive) within-person across occasions.
- Proximal Outcomes
- Time-to-open, session length, reply rate, self-reported urge/goal conflict, next-day digital behavior.
- Moderators
- Schedule density, time of day, baseline engagement profile, prior response pattern.
- Why It Matters
- The MRT converts the research from descriptive to genuinely causal at the feature level. It tests whether timing-matched outperforms random, question outperforms directive, mirror-material outperforms generic, recovery outperforms productivity — all within-person.
XIII. Preliminary Evidence
The UNH pilot was not designed as a formal study. But its descriptive findings establish the feasibility and directional signal that justify the protocol above.
The system was not optimized for engagement, yet retention was high. It did not tell students to stop scrolling, yet scrolling decreased. It did not give answers, yet awareness increased. It did not maximize session length, yet students continued responding. That pattern — relevance-driven engagement without addictive design — is the signal that justifies testing the mechanism formally.
XIV. Theoretical Framework
The paper integrates five theoretical streams. None is new individually. The contribution is in showing how they interact within a single adaptive system and testing that interaction empirically.
Context-Aware Computing (Dey, 2001)
Any information that characterizes the situation of an entity — and its use to provide relevant services. Journey extends this from computing into behavioral intervention.
ACIP Framework (Davidson et al., 2020)
Awareness, Connection, Insight, and Purpose as trainable dimensions of wellbeing. The Healthy Minds Index operationalizes ACIP in 17 items.
Just-in-Time Adaptive Interventions (Nahum-Shani et al., 2018)
Delivering the right support at the right time by adapting to changing context and state. Journey’s context engine implements JITAI logic through schedule-awareness and scarcity-constrained selection.
Motivational Interviewing (Miller & Rollnick, 2013)
Resolving ambivalence by evoking the person’s own reasons for change. Journey’s nudge architecture draws on MI: question-first, reflective, permission-giving.
Digital Harm Reduction
An awareness-first frame replacing abstinence logic with agency-building. The target is not less screen time but more intentional technology use.
XV. The Contribution
Most nudge studies treat timing as a delivery parameter. This paper treats it as the primary mechanism of action and tests that claim with schedule-derived features.
The three-nudges-per-week constraint introduces a novel design logic: intervention value comes not from frequency but from precision.
The paper does not treat the system as a black box. It maps the adaptive loop explicitly and tests whether the loop itself produces change.
Shows the baseline clearly, measures change, models mechanisms, accounts for segmentation, and states what the design can and cannot prove. Scientific honesty as contribution.
A paper that says “nudges helped students” is common. A paper that says “here is the baseline model, here is the adaptive system, here is the observed change, here are the plausible mechanisms, here is who it works for and who it does not, and here are the causal limits” — that is a contribution. The discussion should maintain a healthy scientific tension: optimistic enough to show promise, sober enough to acknowledge what the design cannot prove.
The deepest question the discussion should address is whether adaptive, context-aware wellness interventions represent a genuinely different category of digital intervention — one built around awareness and agency rather than restriction and compliance — and whether that difference matters for how emerging adults develop the capacity to direct their own cognitive and emotional lives.
Journey Research Group · Working Paper · April 2026
Correspondence: development@valerin-group.com