Hong Minhee's rebuttal to calls for blocking AI crawlers argues that withdrawing code entrenches corporate control; instead, he proposes 'training copyleft' licenses requiring open model weights to reclaim generative AI as a commons.
The collision between large language models and free/open source software has ignited one of the most consequential debates in modern software ethics. When a recent blog post advocated for F/OSS developers to block AI crawlers, withdraw from GitHub, and shun proprietary LLM users, developer Hong Minhee responded not with agreement but with a radically different framework rooted in historical materialism and licensing evolution. His central thesis: the free software movement must reclaim generative AI through strategic licensing innovation rather than retreat into digital isolationism. This position synthesizes decades of F/OSS philosophy with urgent technological reality, proposing that LLMs could become the next frontier of communal knowledge—if the community evolves its tools.
Minhee acknowledges the validity of the anger driving withdrawal advocates. AI corporations exhibit profound disrespect by exploiting permissive licenses, ignoring opt-out requests, and treating public code as disposable training fodder. Current copyright law remains woefully inadequate, unable to address how statistical pattern extraction differs from traditional code reuse. Yet where critics see an insurmountable threat, Minhee identifies a recurring historical pattern. From GPLv2 closing the binary-distribution loophole in 1991 to AGPL countering SaaS exploitation in 2007, F/OSS licensing has always adapted to new forms of value extraction. Each evolution followed material shifts in technology: hardware locks, cloud services, and now neural networks. The training-data loophole—where companies privatize models built on communal code—isn’t a death knell but the latest dialectical challenge demanding a materialist response.
Withdrawal strategies falter, Minhee argues, on practical and philosophical grounds. Technically, major AI firms have already scraped vast code repositories; blocking crawlers now primarily harms open-source LLM projects like Llama or Mistral by limiting their training data. Socially, attempts to shun developers using tools like GitHub Copilot risk fracturing communities through unenforceable purity tests. More fundamentally, retreat contradicts the F/OSS ethos of freedom through reciprocity. The genius of copyleft licenses lies not in restricting access but in ensuring improvements flow back to the commons. Denial cedes the ideological battlefield, allowing corporations to define AI’s ownership norms while F/OSS becomes a gated enclave.
Minhee’s counterproposal—a hypothetical GPLv4 or Training GPL (TGPL)—explicitly permits LLM training while imposing revolutionary obligations: models trained on copylefted code must release their weights under compatible licenses, document training data provenance, extend obligations to fine-tuned derivatives, and treat API access as distribution. This framework mirrors past adaptations: just as AGPL redefined network use as equivalent to distribution, TGPL would redefine model weights as derivative works requiring source-code-equivalent disclosure. Technical enforcement challenges (e.g., proving code inclusion in training sets) mirror early GPL skepticism, but community vigilance, dataset transparency, and legal pressure could establish workable norms. Crucially, mixed training scenarios would adopt established solutions for license compatibility, akin to handling GPL and proprietary code linkages.
The implications extend beyond legal mechanics. Minhee positions this as a battle for the soul of AI’s future: will models become proprietary monopolies or communal infrastructure? Citing Salvatore Sanfilippo’s acceptance of LLMs as inevitable productivity tools, he agrees adaptation is necessary but insists ownership remains non-negotiable. If millions of developers contributed to the code commons, the resulting models should belong to humanity, not shareholders. Failure to act risks entrenching a feudal AI landscape where open models, starved of quality data and legal protections, become permanently disadvantaged. Conversely, seizing this historical moment could democratize AI as profoundly as GPL democratized operating systems.
Critics might argue that enforcement remains impractical or that corporations will circumvent new licenses. Yet Minhee’s materialist analysis suggests such obstacles are temporary. Every licensing leap faced similar doubts before becoming foundational—from Linux’s dominance to AGPL’s adoption by MongoDB and Elastic. The deeper risk isn’t imperfection but inaction; without proactive licensing, five years hence the training loophole could be cemented by corporate-friendly jurisprudence. As with all F/OSS struggles, success hinges on collective action: projects adopting training copyleft, foundations lobbying for legal recognition, and developers rejecting fatalism.
Ultimately, Minhee reframes the conflict not as technology versus ethics but as two visions of community survival. Withdrawal offers catharsis but surrenders agency; reclamation through licensing continues F/OSS’s core mission of converting collective labor into collective benefit. In this light, LLMs become not predators but potential allies—if their power is harnessed by the same principles that built the digital commons. The path forward demands not less open source, but more: expanding copyleft to ensure that when code passes through neural networks, it emerges as freedom.

Comments
Please log in or register to join the discussion