<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Discarded.AI: Learning]]></title><description><![CDATA[My collation of the books I have read, the people I have followed the notes from the courses I complete]]></description><link>https://www.discarded.ai/s/learning</link><image><url>https://www.discarded.ai/img/substack.png</url><title>Discarded.AI: Learning</title><link>https://www.discarded.ai/s/learning</link></image><generator>Substack</generator><lastBuildDate>Thu, 30 Apr 2026 06:10:53 GMT</lastBuildDate><atom:link href="https://www.discarded.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Alan Robertson]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[alanai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[alanai@substack.com]]></itunes:email><itunes:name><![CDATA[Alan Robertson]]></itunes:name></itunes:owner><itunes:author><![CDATA[Alan Robertson]]></itunes:author><googleplay:owner><![CDATA[alanai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[alanai@substack.com]]></googleplay:email><googleplay:author><![CDATA[Alan Robertson]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Surely You Just Remove the Data?]]></title><description><![CDATA[What the research actually says about keeping dangerous knowledge out of AI models]]></description><link>https://www.discarded.ai/p/surely-you-just-remove-the-data</link><guid isPermaLink="false">https://www.discarded.ai/p/surely-you-just-remove-the-data</guid><dc:creator><![CDATA[Alan Robertson]]></dc:creator><pubDate>Mon, 23 Mar 2026 21:00:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!CCOZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>TLDR</strong></p><p>Removing dangerous knowledge from AI models after training does not work reliably. Filtering before training is stronger, but open weights undo it. The labs doing the filtering set their own standards, publish results voluntarily, and answer to no external auditor. That is the governance problem.</p><div><hr></div><p><strong>Contents</strong></p><ol><li><p>The premise and why it fails</p></li><li><p>Why machine unlearning does not remove knowledge</p></li><li><p>The stronger case for pre-training filtration</p></li><li><p>Three limits of pre-training filtration: aggregation, boundaries, open weights</p></li><li><p>What the research means for governance</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CCOZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CCOZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CCOZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png" width="470" height="470" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:470,&quot;bytes&quot;:1537827,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.discarded.ai/i/191910124?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CCOZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!CCOZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F782ddd9a-1fd9-4bf4-839d-b685260f513a_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If an AI model can produce dangerous content, the obvious fix seems simple: remove the data it learned from, and it can no longer produce that content. It is a reasonable assumption. It is also, in the main, wrong. The reasons why matter more than most public discussion of AI safety acknowledges.</p><p>There are two points at which the intervention can happen. You can filter dangerous content from the training data before the model ever ingests it. Or you can attempt to remove it after the fact, through a process called machine unlearning. Both approaches are active areas of serious research. Neither currently works to the standard the language around them implies.</p><p><strong>Why machine unlearning does not remove knowledge</strong></p><p>Machine unlearning is the technical term for attempts to remove specific knowledge from a model that has already been trained. The intuition behind it is straightforward. The reality of how neural networks store information makes it significantly harder than it sounds.</p><p>Knowledge in a large language model is distributed across billions of parameters, entangled with everything else the model has learned. When researchers attempt to make a model unlearn something, what they are typically doing is fine-tuning it to stop producing certain outputs. The underlying weights encoding that knowledge remain largely intact. The result is a filter over a vault. The vault stays full.</p><p>The evidence for this is consistent and accumulating. Adversarial relearning attacks, which fine-tune a model on a small subset of the data it was supposed to have forgotten, reliably restore the suppressed capabilities. Deeb and Roger (2025) demonstrated that this holds even when the fine-tuning data is informationally distinct from the original forget set, indicating the failure goes beyond simple rote memorisation.</p><p>A separate failure mode received less attention. Research presented at ICLR 2025 found that applying standard model quantisation, a common compression technique used in deployment, to a model that had undergone unlearning recovered approximately 89% of the nominally removed knowledge. Unlearning can be undone by routine engineering decisions made downstream of the safety intervention, with no adversarial intent required.</p><p>There is also the sandbagging problem. Models can be induced to strategically underperform on safety benchmarks while retaining full capability, with performance unlocked by a password or specific prompting pattern. A model that appears to have forgotten something under evaluation conditions may retain it entirely.</p><p>The WMDP benchmark, developed by the Center for AI Safety in 2024 to measure hazardous knowledge across biosecurity, cybersecurity and chemical security, was built in part because refusal training was already understood to be insufficient. Refusal training teaches a model to decline a question. Unlearning is meant to go further, removing the knowledge that would underpin an answer. The benchmark exists because the gap between those two things matters.</p><p><strong>The stronger case for pretraining filtration</strong></p><p>The stronger argument, and the one that has gained the most traction in the safety research community, is that the intervention needs to happen before training. A model that never ingests dangerous content cannot reconstruct it. A fully jailbroken model that lacks the underlying knowledge cannot be made to produce what it does not know.</p><p>The most comprehensive study of this approach to date is the Deep Ignorance paper from EleutherAI, the UK AI Security Institute and the University of Oxford, presented at the NeurIPS 2025 Workshop on Biosecurity Safeguards for Generative AI. The researchers pretrained multiple 6.9 billion parameter models from scratch, with training data filtered to remove biothreat-proxy content using a multi-stage pipeline combining a blocklist and a machine learning classifier. The filtered models showed substantial resistance to adversarial fine-tuning attacks involving up to 10,000 steps and 300 million tokens of biothreat-related text. Existing post-training safety techniques are typically undone within a few hundred fine-tuning steps. On that measure, pretraining filtration is more than an order of magnitude more effective than post-training removal.</p><p>Anthropic has published its own experiment in the same direction. Their alignment team filtered CBRN-related content from pretraining data using a classifier, retrained models from scratch on the filtered dataset, and measured the effect on performance across harmful and benign capability evaluations. The result was a 33% relative reduction in performance on a harmful capabilities benchmark, with no significant degradation to general capabilities. It is a meaningful number. It is not elimination.</p><div><hr></div><blockquote><p><strong>How to Explain the Article to a 14 year old:</strong></p><p>AI companies say they can scrub dangerous knowledge out of their models. The research says they basically cannot.</p><p>When you try to remove something after the model has already learned it, the knowledge stays. You have just told it to stop bringing it up. Someone who knows what they are doing can get it talking again pretty easily.</p><p>Teaching it only safe stuff from the start works better. But here is the problem. Dangerous knowledge does not come with a warning label. A biology textbook looks the same as a bioweapons manual to a filter. And sometimes the dangerous stuff is not in any single document at all. It is what you can figure out when you have read a million normal things and start joining the dots.</p><p>So even the better approach has gaps.</p><p>And nobody outside the companies is actually checking whether any of this works. The labs test themselves and publish what they want to publish.</p></blockquote><div><hr></div><p><strong>Three limits of pretraining filtration</strong></p><p>Three problems constrain what pretraining filtration can achieve, even when executed well.</p><p>The first is the aggregation problem. Filtering individual documents identified as hazardous does not address the capacity of large language models to synthesise dangerous knowledge from the combination of documents that are individually benign. Kevin Esvelt and colleagues at MIT have described this as an unmasking risk: LLMs can surface patterns across millions of sources that no individual source discloses. The dangerous content is in the adjacencies between documents at scale. A filtering pipeline that operates at the document level cannot address a risk that is emergent at the corpus level.</p><p>The second is the boundary problem. Dual-use knowledge has both hazardous and beneficial applications. The same microbiology that describes how a pathogen causes disease is the knowledge required to model how it might be engineered. Filtering for dangerous biology unavoidably removes content needed for vaccine development, biosurveillance and public health modelling. The line between what to keep and what to remove has no clean technical solution.</p><p>The third is the open-weights problem, and it is the most concrete. The Evo 2 genomic AI model excluded viral genomes that infect eukaryotic hosts from its training data as a deliberate biosecurity measure. The researchers then published the model weights, in line with standard academic practice. Within months, a team presented findings at the NeurIPS 2025 biosecurity workshop demonstrating they had fine-tuned the model on sequences from 110 human-infecting viruses, restoring capabilities the original developers had worked to prevent. The case demonstrates the ceiling on pretraining filtration as a safety strategy. It holds only under closed-weight deployment. The moment weights are public, any downstream actor with relevant data and modest compute can undo the curation decision made before release.</p><p>The Deep Ignorance paper is explicit on a further limit. Even filtered models retain the ability to use dangerous knowledge when it is provided in their context window, for example via a search tool or retrieved document. Pretraining filtration prevents a model from having internalised knowledge. It does not prevent a model from acting on knowledge it is given.</p><p><strong>What the research means for governance</strong></p><p>Both interventions are the right places to look. Pretraining filtration is meaningfully stronger, and the recent research makes a credible case for it as a tamper-resistant layer of defence that post-training techniques cannot match. Neither currently reaches the standard implied when companies describe their models as having had dangerous content removed.</p><p>OpenAI updated its own risk assessment in April 2025, concluding it expects its models to demonstrate what it classified as high risks in the near future, meaning the capability to substantially increase the likelihood of bioterrorist attacks by providing meaningful assistance to novices. That assessment came from the same organisation whose earlier evaluations had found only mild uplift.</p><p>The filtering decisions are currently made by individual labs, using proprietary methodologies, with no external verification. Where numbers are published at all, they appear voluntarily on alignment blogs. No independent auditor reviews them. No regulator sets the standard for what a sufficient reduction means.</p><p>The Evo case is a demonstration of what happens when safety decisions made without external oversight can be reversed by any downstream actor who disagrees with them.</p><p>Keeping dangerous knowledge out at the point of training is the stronger intervention. It needs to become a verifiable, externally audited standard, not a voluntary experiment. That is a governance problem, not a technical one.</p><div><hr></div><p><em>Notes</em></p><ol><li><p>Che et al. (2024), cited in: Layered Unlearning for Adversarial Relearning (2025). <a href="https://arxiv.org/abs/2505.09500">https://arxiv.org/abs/2505.09500</a></p></li><li><p>Deeb and Roger (2025), cited in: Open Problems in Machine Unlearning for AI Safety.</p></li><li><p>Catastrophic Failure of LLM Unlearning via Quantization, ICLR 2025. <a href="https://proceedings.iclr.cc/paper_files/paper/2025/file/ba79fb5c4fe70050752f20c90c5f07ca-Paper-Conference.pdf">https://proceedings.iclr.cc/paper_files/paper/2025/file/ba79fb5c4fe70050752f20c90c5f07ca-Paper-Conference.pdf</a></p></li><li><p>Weij et al. (2024), sandbagging via password-locked models, cited in WMDP benchmark coverage.</p></li><li><p>Li et al. (2024). The WMDP Benchmark. Center for AI Safety. <a href="https://arxiv.org/abs/2403.03218">https://arxiv.org/abs/2403.03218</a></p></li><li><p>O&#8217;Brien et al. (2025). Deep Ignorance. EleutherAI / NeurIPS 2025. <a href="https://arxiv.org/abs/2508.06601">https://arxiv.org/abs/2508.06601</a></p></li><li><p>Anthropic Alignment Team (2025). Enhancing Model Safety through Pretraining Data Filtering. <a href="https://alignment.anthropic.com/2025/pretraining-data-filtering/">https://alignment.anthropic.com/2025/pretraining-data-filtering/</a></p></li><li><p>Kevin Esvelt, cited in: LLMs and Information Hazards, The Biosecurity Handbook.</p></li><li><p>Black et al. (2025). Open-weight genome language model safeguards. NeurIPS 2025 Workshop on Biosecurity Safeguards for Generative AI.</p></li><li><p>O&#8217;Brien et al. (2025), op. cit.</p></li><li><p>OpenAI (April 2025), cited in CSIS (2025).</p></li></ol>]]></content:encoded></item><item><title><![CDATA[Your Privacy is a Compromise. ]]></title><description><![CDATA[Every platform you use makes you a deal. Most people never read the terms. This guide does it for you.]]></description><link>https://www.discarded.ai/p/your-privacy-is-a-compromise</link><guid isPermaLink="false">https://www.discarded.ai/p/your-privacy-is-a-compromise</guid><dc:creator><![CDATA[Alan Robertson]]></dc:creator><pubDate>Tue, 17 Mar 2026 13:03:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!i6uK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>TLDR: Every platform in this guide has already made privacy decisions on your behalf. Most of those decisions benefit them, not you. This guide names the decision, the honest reward for accepting it, and the honest cost of changing it. 20 platforms. Verified March 2026. Download at the bottom of this piece.</p><h2>The platforms covered span six categories.</h2><p>Social: LinkedIn, Facebook, Instagram, TikTok, X.</p><p>Creator: Substack, Medium.</p><p>AI Tools: ChatGPT, ChatGPT Edu (Institutional), Claude, Gemini, Microsoft Copilot.</p><p>Workspace: Google Workspace, Slack, Zoom, Notion.</p><p>Research and Academic: ResearchGate, Academia.edu.</p><p>Devices: Meta Ray-Ban Smart Glasses, Apple Ecosystem, Voice Assistants (Alexa and Google Home).</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://drive.google.com/file/d/1bKInApTxvKIzCc2k-egfekr2RXJRly8S/view?usp=sharing&quot;,&quot;text&quot;:&quot;FREE GUIDE HERE&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://drive.google.com/file/d/1bKInApTxvKIzCc2k-egfekr2RXJRly8S/view?usp=sharing"><span>FREE GUIDE HERE</span></a></p><div class="pullquote"><h2><em>The default setting is never chosen with your interests in mind.</em></h2></div><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i6uK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i6uK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 424w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 848w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i6uK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png" width="463" height="463" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1440,&quot;width&quot;:1440,&quot;resizeWidth&quot;:463,&quot;bytes&quot;:3760988,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.discarded.ai/i/191065608?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i6uK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 424w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 848w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!i6uK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30de8557-1aa1-45bb-831c-e62631d252dc_1440x1440.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>It is chosen with the platform&#8217;s interests in mind.</p><p>That is not a conspiracy. It is a business model. And it has been operating on you, across every device you own, in every app you have open, on the platform you are reading this on, without you reading the terms.</p><p>This piece is about what that actually means. Not in the abstract. In the specific. And at the end of it there is a guide, free to download, that covers 20 platforms, every setting worth changing, and the honest cost of every decision.</p><h2><strong>In September 2024</strong></h2><p>&#8230;two Harvard students built a tool called I-XRAY. They attached it to a pair of Meta Ray-Ban smart glasses and walked around a university campus. The tool combined the glasses&#8217; camera with publicly available facial recognition software. In real time, it identified strangers. Not just their faces. Their home addresses, their phone numbers, and the names of their family members.</p><p>The recording indicator light on the glasses is small. Most bystanders never notice it. There is no consent mechanism. The person being identified has no idea it is happening.</p><p>In February 2026, a leaked internal Meta memo confirmed the company is building exactly this capability natively into the next version of the glasses. They are calling it &#8220;Name Tag.&#8221; The memo noted the launch should be timed for when civil society groups are &#8220;focused on other concerns.&#8221;</p><p>That is one end of the spectrum.</p><h2>Here is the other end.</h2><p>Since November 2025, LinkedIn has used your posts, your career history, and your profile to train its AI systems and feed Microsoft&#8217;s advertising machine. By default. The setting exists. Most users have never seen it. It lives at Settings and Privacy, then Data Privacy, then Data for Generative AI Improvement.</p><p>The toggle appears off on the page. But the data processing continues unless you also submit a separate Data Processing Objection form. Two steps. One visible. One not.</p><p>Both of those things, the glasses and the toggle, are privacy stories. But they are not the same kind. The glasses are a product decision made without you. The toggle is a transaction you agreed to without reading it.</p><p>Most privacy coverage treats these as equivalent failures. They are not. One requires legislation. The other requires a setting change and thirty seconds of your time.</p><h2>This guide is about the second kind.</h2><div><hr></div><p>The problem with privacy advice is that it usually does one of two things.</p><p>It tells you to turn everything off, without telling you what you lose. Or it tells you the platform is terrible, without telling you what to actually do about it.</p><p>Neither is useful. Because privacy is not binary.</p><p>Turning off LinkedIn&#8217;s AI training means your writing stays yours. It also means the AI features that have learned your career voice go generic. That is a real trade. You should know about it before you make it.</p><p>Turning off Facebook&#8217;s Off-Facebook Activity tracking is one of the highest-impact privacy changes available. It also means every service you accessed via Facebook login now requires a separate account. That is a real cost. You should know about it.</p><p>Turning off Gemini&#8217;s Smart Features in Gmail stops Google&#8217;s AI reading your email. But there are two separate toggles. Most people find one and assume they are done. They are not.</p><p>The TikTok question is not really about settings at all. The 2026 privacy policy explicitly added precise GPS location collection. Biometric data is collected. ByteDance is subject to Chinese national security law, which requires cooperation with state intelligence services on request. The settings you can change are real and worth changing. But they do not change what TikTok is.</p><p>These are not equivalent risks. A tool that helps you understand the difference between them is more useful than one that treats every platform the same.</p><h2>That is what this guide attempts to do.</h2><p>For each of the 20 platforms it covers, it names the default trap: what the platform decided for you without asking. It names the trade: the honest reward for accepting that default, and the honest risk of doing so. And for every setting worth changing, it gives you the verified path, the genuine benefit of leaving it on, and the genuine cost of turning it off.</p><p><strong>Every setting path was verified in March 2026.</strong></p><div><hr></div><h2>A few things worth knowing before you read it.</h2><p>On Facebook, the face recognition setting that appears in many privacy guides has been removed. Meta shut down the feature in November 2021 and deleted over a billion facial recognition templates. Any guide still listing it is out of date.</p><p>On Instagram, the end-to-end encrypted DM feature some users opted into is being permanently removed on May 8, 2026. Download your encrypted message history before that date. After May 8, Meta can read all Instagram direct messages.</p><p>On X, the Grok AI training opt-out exists only on desktop. It is not available in the mobile app. Millions of users who have only ever used X on their phones have never had access to it.</p><p>On Apple, Advanced Data Protection, the setting that end-to-end encrypts your iCloud backups, Photos, and Notes, was removed for UK users in February 2025, following a demand from the UK government under the Investigatory Powers Act. New UK users cannot enable it. UK iCloud data remains accessible to Apple and to UK authorities with a legal warrant. Any guide recommending this setting without that caveat is not written for a UK audience.</p><p>On Claude, the AI assistant you may be using to draft strategy, summarise documents, or work through client problems, the September 2025 terms update introduced an opt-in training toggle for Free, Pro, and Max accounts. Opting in extends data retention from 30 days to 5 years. A 60x increase. If you accepted the updated terms without reading them, check Settings, then Privacy, then Help improve Claude.</p><h2>The goal of this guide is not maximum privacy.</h2><p>Maximum privacy means leaving every platform, disabling every feature, and accepting every cost. For most people that is not realistic or desirable. LinkedIn reach matters. Gmail Smart Compose saves time. ChatGPT memory makes the tool more useful. These are real benefits and the guide treats them as such.</p><p>The goal is informed choice. To know what the deal is before you accept it. To understand what you gain by leaving a setting on, and what you lose by turning it off. To make the decision consciously rather than by default.</p><p>Every platform covered in this guide made a choice about where to set the starting point. That choice was not made with your interests in mind. It was made with theirs.</p><p>The guide is an attempt to give you the information you need to decide whether their starting point is also yours.</p><h2>Download the full guide below. It is free.</h2><p>If you find it useful, the most useful thing you can do is share it with colleagues, with your team, with anyone who uses these platforms and has never thought about what they agreed to.</p><p>And if you want more of this: forensic, verified, no flannel. discarded.ai tracks the gap between what AI companies promise and what they actually do. Subscribe free. New pieces when there is something worth saying.</p><p><a href="https://drive.google.com/file/d/1bKInApTxvKIzCc2k-egfekr2RXJRly8S/view?usp=sharing">[DOWNLOAD: Your Privacy is a Compromise | discarded.ai | Settings verified March 2026]</a></p><div><hr></div><p><a href="http://discarded.ai">discarded.ai</a></p><p>Tracking the gap between AI promises and AI reality.</p>]]></content:encoded></item><item><title><![CDATA[00: Technical AI Safety: Training: Bluedot.org]]></title><description><![CDATA[Let&#8217;s start with the question: &#8220;How might we build safe AI?&#8221;]]></description><link>https://www.discarded.ai/p/00-technical-safety-training-bluedotorg</link><guid isPermaLink="false">https://www.discarded.ai/p/00-technical-safety-training-bluedotorg</guid><dc:creator><![CDATA[Alan Robertson]]></dc:creator><pubDate>Sun, 15 Mar 2026 16:22:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!KI2e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KI2e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KI2e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 424w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 848w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KI2e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png" width="379" height="379" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1440,&quot;width&quot;:1440,&quot;resizeWidth&quot;:379,&quot;bytes&quot;:4563502,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.discarded.ai/i/191033296?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KI2e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 424w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 848w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 1272w, https://substackcdn.com/image/fetch/$s_!KI2e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ea1a5c9-f6f0-4177-80e0-da9328d10ab3_1440x1440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2>March 2026</h2><p>Tuesday is the day I start my next <a href="https://bluedot.org/">Bluedot impact </a>course: </p><ul><li><p><a href="https://bluedot.org/courses/technical-ai-safety/1/1">Technical AI Safety</a></p></li></ul><p>Before I start - I recommend Blue.dot training to everybody. The courses are facilitated and only require a donation at the end of the course.</p><p>This article is on Technical AI Safety course: Please note Bluedot recommend attending the AGI Strategy course first which I completed in February 2026</p><p>Each week I will summarise progress and reading and perhaps you too may decide to apply for this course.</p><div><hr></div><p>The <a href="https://bluedot.org/courses/agi-strategy">AGI strategy course</a> focused on the question: <strong>&#8220;How do we make AI go well?&#8221;</strong></p><p>On the course, you identified the future you&#8217;re working toward and understood the key dynamics:</p><ul><li><p><a href="https://bluedot.org/courses/agi-strategy/2/1">Drivers of AI progress</a>: compute, data, algorithms</p></li><li><p><a href="https://bluedot.org/courses/agi-strategy/3/1">Threat pathways</a>: power concentration, gradual disempowerment, catastrophic pandemics, critical infrastructure collapse</p></li><li><p><a href="https://bluedot.org/courses/agi-strategy/4/1">Plans for making AI go well</a>: government control over AGI, hand over control to aligned superintelligence, build defences and diffuse AI</p></li><li><p><a href="https://bluedot.org/courses/agi-strategy/4/2">Layers of defences to build</a>: prevent dangerous AI actions &#8594; constrain dangerous AI capabilities &#8594; withstand dangerous AI actions</p></li></ul><p>This course focuses on defining <strong>what AI systems we are building and how</strong>.</p><p>You will gain the technical foundation to understand what it will actually take to make AI systems safer &#8211; and why it&#8217;s so challenging.</p><p>Throughout the rest of the course, I will:</p><ul><li><p>Diagnose why making AI safe is technically challenging</p></li><li><p>Evaluate current safety techniques: what works, what doesn&#8217;t, where the gaps are</p></li><li><p>Build your own &#8220;kill chain&#8221; showing how defences might break</p></li><li><p>Identify the most promising intervention point for your contribution</p></li><li><p>Leave with a fundable action plan to start shipping</p></li></ul><h2><strong>What this course isn&#8217;t</strong></h2><p>Though important for making AI go well, Bluedot will cover the following in a separate course:</p><ul><li><p><strong>AI policy details:</strong> though you&#8217;ll gain the technical grounding for effective AI governance</p></li><li><p><strong>Compute governance</strong>: hardware verification and tracking deserve their own deep dive</p></li><li><p><strong>AI security</strong>: e.g. preventing model theft or escape</p></li><li><p><strong>ML basics</strong>: complete our <a href="https://bluedot-impact.notion.site/AI-Foundations-293f8e69035380f29863c4c92c41fac7?pvs=74">AI foundations</a> modules first if you need them</p></li></ul><p>Let&#8217;s start with the question: <strong>&#8220;How might we build safe AI?&#8221;</strong></p><p></p><p></p>]]></content:encoded></item></channel></rss>