Enter the Autism Experience
Immersive walk-through that approximates auditory and visual overload from a first-person perspective.
Seeing the world through the lens of others.
A research prototype that takes everyday video of a candidate environment, surfaces the sensory triggers it contains, and renders a sensory-modulated version of the same clip — so families and practitioners can make informed choices about autism-friendly spaces before a site visit.
A restaurant. An AirBnB. The space looks fine in photos — but for an autistic person, the same room may be saturated with sensory triggers that turn a casual visit into a draining experience.
How can we generate sensory-modulated videos from everyday video footage to help make informed decisions about autism-friendly environments?
Prior efforts invite neurotypical viewers to step into a VR simulation. The thesis is sound — the best way to learn something is to experience it — but every viewer has to gear up.
Immersive walk-through that approximates auditory and visual overload from a first-person perspective.
Generative bursts of colour and motion staged in everyday spaces — malls, classrooms, public concourses.
We wanted to use everyday video — recorded with a phone — as the substrate, so a parent or therapist can run an assessment in two minutes.
Real-world video footage of any place of your choice — restaurant, AirBnB, classroom, lobby.
A modified video that identifies common stressors and visually modulates the sensory effects of the original, plus a structured JSON report.
Cognitive Amplification. We augment the viewer's ability to understand the perception of others — surfacing what is already in the scene rather than taking anything away.
Parmar KR, Porter CS, Dickinson CM, Pelham J, Baimbridge P, Gowen E. Visual Sensory Experiences From the Viewpoint of Autistic Adults. Front Psychol. 2021 Jun 8;12:633037. doi: 10.3389/fpsyg.2021.633037
Orchestrated by
analyzers/main_processor.py; HTTP surface in
routers/analyze.py; entry point main.py.
Per-job artefacts are written to results/<job_id>/.
| Light flicker | Frame-brightness variance — proxy for fluorescent / PWM flicker |
|---|---|
| Brightness | Mean luminance, scored 0–10 |
| Colour saturation | HSV saturation distribution + dominant-colour palette |
| Visual clutter | Edge density & pattern repetition |
| Motion | Inter-frame difference magnitude |
| Background level | RMS → estimated dB SPL, mean & peak |
|---|---|
| Sudden events | Energy-spike onsets with timestamps |
| Low-frequency hum | Spectral mass below ~120 Hz (HVAC / mechanical) |
| High-frequency content | Spectral mass above ~4 kHz (beeps / alarms) |
| Speech intelligibility | SNR-based estimate |
{job_id}_video.mp4 — demuxed video stream{job_id}_audio.wav — extracted audio{job_id}_video_processed.mp4 — overlays / annotations applied{job_id}_audio_processed.wav — re-rendered audio{job_id}_final.mp4 — muxed deliverable, shown as the right-hand player on the Results tab{job_id}_report.json — structured report (visual / audio / timeline / concerns / recommendations)Each modality (visual, audio) yields a 0–10 sub-score, combined into an overall sensory score on the same scale. Lower is calmer; higher indicates more potential triggers. The report also exposes a per-second sensory_timeline so peaks within a clip can be inspected.
Source uploads live under uploads/, processed artefacts under
results/{job_id}/. The
Results
tab pairs every input with its processed output and surfaces the
parameters extracted from the report for visual comparison.