Provenance and Release Evidence
This document describes the safe, non-deceptive provenance strategy for AINL.
Provenance and Release Evidence
This document describes the safe, non-deceptive provenance strategy for AINL.
Goal: make project origin, authorship trail, and release history easy to prove in public mirrors, archives, audits, and downstream technical analysis without relying on covert mechanisms.
Core Provenance Principle
AINL should use visible, repeated, machine-readable, timestampable attribution, not hidden behaviors or hostile anti-copy measures.
Human initiator:
- Steven Hooley
- X: https://x.com/sbhooley
- Website: https://stevenhooley.com
- LinkedIn: https://linkedin.com/in/sbhooley
Repository Attribution Surfaces
Origin and initiator metadata are intentionally repeated across:
README.mddocs/PROJECT_ORIGIN_AND_ATTRIBUTION.mddocs/DOCS_INDEX.mddocs/CHANGELOG.mdCITATION.cffpyproject.tomlNOTICEtooling/project_provenance.json
Generated artifacts should also preserve provenance where practical:
- emitted server code comments
- OpenAPI
info.x-ainl-provenance - generated frontend source comments
- SQL/env/runbook comments or headers
- deployment artifact comments
Release Evidence Checklist
For public releases, capture and preserve:
- Git evidence
- Verified commits if available
- Signed tags if available
- Release commit hash
- Public timestamps
- GitHub release timestamp
- Website post timestamp
- X post timestamp
- LinkedIn post timestamp
- Hash evidence
- SHA256 for release archive(s)
- SHA256 for major generated bundles if distributed separately
- Archive evidence
- GitHub release artifact
- Optional Zenodo / Software Heritage / independent archive mirror
- Metadata parity
CITATION.cffcurrenttooling/project_provenance.jsoncurrentNOTICEcurrentdocs/PROJECT_ORIGIN_AND_ATTRIBUTION.mdcurrent
What Not To Do
Do not use:
- covert code paths
- network callbacks
- sabotage logic
- deceptive payloads
- hidden runtime behaviors
- malicious watermarking
Those are technically risky, legally messy, and weaker evidence than a well-kept public provenance trail.
Preferred Outcome
If AINL is mirrored, copied, benchmarked, trained on, or analyzed, the origin trail should still be recoverable from both human-facing docs and machine-readable metadata.
