WEB ATELIER (UDIT) · Learning by doing, with theory, practice and shared reflection

Final Presentation: Demo Day & Monograph

Draft

URL: https://ruvebal.github.io/web-atelier-udit/lessons/en/react/final-presentation/

📋 Table of Contents

“Code that cannot be explained does not yet exist for the team.”


🎯 Sprint Goal

By the end of this sprint: you can demonstrate a working application, explain your architecture decisions, justify trade-offs, and submit a monograph that documents process, constraints, failures, and learning (including AI usage).


✅ Deliverables (Required)

  • Live application
    • Deployed URL (production build)
    • Demo credentials (if auth exists) or a public demo mode
    • Accessibility baseline (keyboard navigation + visible focus + semantic structure)
  • Source code
    • GitHub repo with clean commit history (no secrets committed)
    • README with setup + scripts + architecture summary
  • Monograph (PDF or Markdown exported to PDF)
    • Short, precise, evidence-based (screenshots, diagrams, metrics)
  • Demo Day Slide Deck (optional but recommended)

🧭 Demo Day Format (Suggested)

1) The Problem (60–90s)

  • Who is the user?
  • What pain do you solve?
  • Why does it matter?

2) The Product (3–5 min)

  • Guided walkthrough of the “happy path”
  • One “edge case” flow (error state, empty state, loading state)

3) The Architecture (3–5 min)

  • Component boundaries (why these components?)
  • State model (UI / server / URL / shared)
  • Data flow (how does state change propagate?)
  • Integration points (API, auth, deployment)

4) The Critique (2–3 min)

  • What failed, what you changed, why
  • What you would do next (roadmap)

🧾 Monograph Template (Minimal, High Signal)

1) Abstract (5–8 lines)

What you built + what you learned.

2) Context & Constraints

  • Time constraints
  • Team constraints
  • Technical constraints
  • Risk spikes (what you tested early)

3) Architecture Decisions (ADR-style)

For 3–6 key decisions, write:

  • Decision: What you chose
  • Alternatives: What you considered
  • Trade-offs: What you gained and what you lost
  • Evidence: Links to commits/issues, screenshots, metrics

4) Accessibility & UX

  • Keyboard-only demo notes
  • Color/contrast checks
  • Reduced motion handling (if relevant)
  • UX patterns you adopted (and why)

5) Performance & Reliability

  • Performance budget or at least 1 measurement (Lighthouse / Web Vitals snapshot)
  • Error handling strategy (loading/error/empty states)
  • Any monitoring/logging notes (even basic)

6) AI Usage Disclosure (Mandatory)

Write exactly what AI did and did not do:

  • Prompts you used (representative samples)
  • What you accepted vs rejected
  • Failures: where AI misled you
  • Verification: how you validated outputs (tests, docs, manual checks)

7) Reflection (Atelier)

  • What did you learn about your own thinking?
  • What did you learn about your team?
  • What did you learn about tools (and their limits)?

🎓 Assessment Rubric (Aligned with Course)

Component Weight What “Excellent” Looks Like
Working application 40% Stable, coherent UX, production-ready baseline
Live demo 15% Clear narrative, handles edge case, no hand-waving
Technical presentation 15% Decisions + trade-offs + evidence, not feature listing
Monograph 20% Concise, structured, reflective, with artifacts
Reflection quality 10% Honest self-critique + concrete next steps

💭 Reflection Questions

💭 Which decision was “expensive” to change? What would you do earlier next time?

💭 What is one bug you shipped that taught you more than any tutorial?

💭 What did AI make easier—and what did it make riskier?


🧘 Koan

“When the demo breaks, do not blame the projector. Study the code you did not understand.”