Commit graph

5 commits

Author SHA1 Message Date
af6062c1f9 Add social cost proxies to impact tracking hooks
Extend pre-compact-snapshot.sh to extract 5 new per-conversation
metrics from the transcript: automation ratio (deskilling proxy),
model ID (monoculture tracking), test pass/fail counts (code quality
proxy), file churn (edits per unique file), and public push detection
(data pollution risk flag). Update show-impact.sh to display them.

New plan: quantify-social-costs.md — roadmap for moving non-environmental
cost categories from qualitative to proxy-measurable.

Tasks 19-24 done. Task 25 (methodology update) pending.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-16 15:05:53 +00:00
735ac1cc4b Add anticipated criticisms plan
Identifies 8 likely criticisms, prioritizes which must be addressed
before launch. AI authorship transparency is the highest priority.
2026-03-16 10:32:49 +00:00
f882b30030 Add pre-launch plans: competitive landscape, audience, impact measurement
- Competitive landscape: maps existing tools (CodeCarbon, EcoLogits, etc.)
  and research, identifies our unique positioning (breadth beyond carbon)
- Audience analysis: identifies 5 segments, recommends targeting ethics/
  governance professionals and developers first
- Project impact measurement: defines success thresholds and metrics to
  determine whether the project itself is net-positive
2026-03-16 10:21:00 +00:00
b0afef0de3 Update plans and tasks to reflect completed publication
Forgejo instance is live at llm-impact.org with landing page.
H1 and H3 are done, H2 (external sharing) remains.
2026-03-16 10:04:32 +00:00
0543a43816 Initial commit: AI conversation impact methodology and toolkit
CC0-licensed methodology for estimating the environmental and social
costs of AI conversations (20+ categories), plus a reusable toolkit
for automated impact tracking in Claude Code sessions.
2026-03-16 09:46:49 +00:00