<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Tony Robinson</title>
    <description>The latest articles on Forem by Tony Robinson (@tonserrobo).</description>
    <link>https://forem.com/tonserrobo</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3853115%2Fe34c1351-3cd2-446b-94af-dbad4aff80d9.jpeg</url>
      <title>Forem: Tony Robinson</title>
      <link>https://forem.com/tonserrobo</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/tonserrobo"/>
    <language>en</language>
    <item>
      <title>Building a Misinformation Resilience Playbook</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:08:04 +0000</pubDate>
      <link>https://forem.com/techethics/building-a-misinformation-resilience-playbook-34dp</link>
      <guid>https://forem.com/techethics/building-a-misinformation-resilience-playbook-34dp</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Misinformation moves faster than most organisations can react. A false health claim can circulate to millions before a fact-checker publishes a correction. A manipulated image can reshape a political narrative before its provenance is questioned. A coordinated disinformation campaign can exploit platform algorithms to achieve the reach of a major news outlet without any of the editorial accountability. In this environment, defensive instincts are necessary but insufficient. Organisations that want to protect their communities, their reputations, and the integrity of the information ecosystems they operate in need a proactive, operationalised playbook that tightens signals, empowers people, and measures impact rather than headlines.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Matters Now¶
&lt;/h2&gt;

&lt;p&gt;Three converging forces have made misinformation resilience an operational priority rather than a communications afterthought. The first is rapid amplification. Recommender systems across major platforms are engineered to reward novelty and emotional intensity, which means that sensational or outrage-provoking claims, regardless of their factual grounding, consistently outperform measured, evidence-based content in reach and engagement. The algorithms do not distinguish between virality driven by genuine public interest and virality driven by manipulation.&lt;/p&gt;

&lt;p&gt;The second force is trust fragility. Research consistently shows that once trust in an institution, platform, or information source is eroded, subsequent corrections are discounted rather than accepted. Audiences who have been exposed to repeated misinformation develop a generalised scepticism that makes accurate information harder to communicate even when it is available. This means that the cost of allowing misinformation to circulate unchecked compounds over time in ways that are far more damaging than any single false claim.&lt;/p&gt;

&lt;p&gt;The third force is regulatory convergence. Codes of practice on disinformation, AI transparency requirements, and emerging content provenance standards are converging across jurisdictions. Organisations that lack demonstrable misinformation resilience practices will increasingly find themselves on the wrong side of regulatory expectations, procurement requirements, and public accountability demands.&lt;/p&gt;

&lt;h2&gt;
  
  
  Early Warning Signals¶
&lt;/h2&gt;

&lt;p&gt;Effective detection begins with narrative heat maps that track the velocity and geographic spread of emerging claims in near real-time. Volume alone is a poor indicator; what matters is the combination of volume, source credibility, and coordination patterns. A claim that spreads rapidly across loosely connected authentic accounts signals different risks than the same claim amplified by a network of newly created or previously dormant accounts. Layering source credibility scores onto volume tracking separates organic concern from manufactured consensus.&lt;/p&gt;

&lt;p&gt;Asset provenance is becoming a critical detection capability. As generative AI makes synthetic media increasingly convincing, the ability to verify the origin and integrity of images, video, and audio is no longer a nice-to-have. For high-risk domains, including elections, public health, and active conflict, cryptographic provenance standards such as C2PA (Coalition for Content Provenance and Authenticity) provide a technical foundation for flagging assets that lack verifiable origin. Organisations should require provenance metadata for high-risk media and treat its absence as a signal worthy of scrutiny.&lt;/p&gt;

&lt;p&gt;Audience vulnerability mapping adds a crucial dimension that purely content-focused detection misses. Not all audiences are equally susceptible to a given false claim. Segmenting by topic literacy, prior exposure to related misinformation, and trust in relevant institutions allows organisations to identify which communities are most at risk of persuasion and to target interventions where they will have the greatest impact rather than deploying blanket responses that may be ignored by those who need them most.&lt;/p&gt;

&lt;h2&gt;
  
  
  Response Playbook¶
&lt;/h2&gt;

&lt;p&gt;Response should be calibrated to the severity and coordination level of the threat. Tiered interventions avoid the twin failures of under-reaction, which allows harmful content to spread unchecked, and over-reaction, which generates accusations of censorship and erodes credibility. For low-risk claims that are misleading but not coordinated, fact labels and contextual annotations are proportionate. For coordinated disinformation campaigns that meet defined harm thresholds, stronger measures such as algorithmic demotion, distribution limits, and in extreme cases removal are warranted. The key is that the criteria for each tier are defined in advance, documented, and applied consistently.&lt;/p&gt;

&lt;p&gt;Context overlays represent a more constructive alternative to pure takedowns. Pairing disputed claims with concise, sourced counter-narratives and links to primary data gives users the information they need to evaluate the claim themselves rather than simply removing it and inviting accusations of suppression. This approach respects user agency while materially reducing the persuasive power of false claims.&lt;/p&gt;

&lt;p&gt;Messenger strategy is often more important than message content. Corrections delivered through corporate communications channels are frequently dismissed by the audiences most at risk. Routing accurate information through trusted community figures, local organisations, and culturally relevant media channels dramatically increases its uptake. This requires building relationships with community partners before a crisis occurs, not scrambling to identify them during one.&lt;/p&gt;

&lt;p&gt;Resilient user experience design embeds friction at the points where misinformation spreads most efficiently. Share flows that surface source quality indicators, publication dates, and semantic similarity warnings before a user reposts content create moments of reflection that reduce thoughtless amplification without preventing deliberate sharing. These design interventions are small in isolation but cumulative in effect.&lt;/p&gt;

&lt;h2&gt;
  
  
  Measurement and Governance¶
&lt;/h2&gt;

&lt;p&gt;What gets measured gets managed, and misinformation resilience is no exception. Tracking exposure and engagement separately, distinguishing between impressions, dwell time, and shares, prevents organisations from underestimating silent spread. A false claim that is widely viewed but rarely shared may be doing more damage than one that generates visible engagement, because passive exposure shapes beliefs without triggering the social signals that detection systems typically monitor.&lt;/p&gt;

&lt;p&gt;Time-to-mitigation service-level agreements create operational discipline. Defining target intervals for detection to label, detection to de-amplification, and detection to removal, and then measuring performance against those targets, transforms misinformation response from an ad hoc activity into a managed capability with clear accountability.&lt;/p&gt;

&lt;p&gt;Red-team drills, conducted quarterly with synthetic campaigns designed to test detection pipelines and moderation policies under realistic conditions, reveal gaps that routine monitoring misses. These exercises should simulate the full range of adversarial tactics, from coordinated inauthentic behaviour to generative AI content to cross-platform amplification, and should result in documented findings and remediation plans.&lt;/p&gt;

&lt;p&gt;Transparency notes published after major interventions serve both accountability and legitimacy functions. Describing what was detected, what actions were taken, what worked, and what gaps remain demonstrates good faith and reduces the suspicion that moderation decisions are arbitrary or politically motivated.&lt;/p&gt;

&lt;h2&gt;
  
  
  Building Public Literacy¶
&lt;/h2&gt;

&lt;p&gt;Technology-side interventions address the supply of misinformation but do little about the demand. Building public resilience requires investing in the critical thinking skills that enable people to evaluate information independently. Embedding lateral reading prompts directly into platform experiences, such as suggestions to check other sources, perform reverse image searches, or verify publication dates, meets users at the moment they encounter dubious content rather than relying on them to seek out media literacy resources on their own.&lt;/p&gt;

&lt;p&gt;Partnerships with schools, newsrooms, libraries, and civil society organisations extend the reach of literacy efforts beyond what any single platform or organisation can achieve. Micro-curricula designed for reuse across contexts, from classroom lessons to newsroom training to community workshops, create a multiplier effect that scales literacy investment.&lt;/p&gt;

&lt;p&gt;Funding independent research on algorithmic amplification and releasing privacy-safe APIs for external auditors builds the evidence base that the entire field depends on. Organisations that hoard data about how their systems interact with misinformation are ultimately undermining their own credibility, because without external validation, their claims about the effectiveness of their interventions remain unverifiable.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Leaders Can Deliver This Quarter¶
&lt;/h2&gt;

&lt;p&gt;For organisations ready to move from intention to action, four concrete steps can be taken within a single quarter. First, stand up a cross-functional misinformation pod that brings together engineering, policy, communications, and legal expertise in a single team with clear ownership and a direct reporting line to leadership. Siloed responses to misinformation are slow responses.&lt;/p&gt;

&lt;p&gt;Second, ship provenance checks for images and video on all high-risk topics. This does not require solving the entire synthetic media problem; it requires implementing existing standards for the content categories where the stakes are highest.&lt;/p&gt;

&lt;p&gt;Third, pilot narrative heat maps in two priority regions with weekly executive reviews. Starting small allows the team to refine detection thresholds and response protocols before scaling, while executive visibility ensures that findings translate into action.&lt;/p&gt;

&lt;p&gt;Fourth, publish the first transparency note describing interventions taken, gaps identified, and next steps planned. This establishes the cadence and the expectation that future notes will follow, creating an accountability rhythm that builds institutional discipline over time.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;The goal is not perfect truth arbitration. No organisation can eliminate misinformation entirely, and claims to the contrary invite justified scepticism. The goal is credible, timely reductions in harm and a visible commitment to integrity that earns and sustains public trust. Teams that operationalise these steps can respond faster, communicate more clearly, and rebuild the public confidence that misinformation erodes. The playbook is not a destination; it is a discipline, and the organisations that practise it will be the ones that communities, regulators, and partners trust most.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/building-a-misinformation-resilience-playbook" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>misinformation</category>
      <category>trustandsafety</category>
      <category>medialiteracy</category>
    </item>
    <item>
      <title>Digital Twins for Population Modeling: Ethics, Signal Quality, and Public Good</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:07:47 +0000</pubDate>
      <link>https://forem.com/techethics/digital-twins-for-population-modeling-ethics-signal-quality-and-public-good-1gk6</link>
      <guid>https://forem.com/techethics/digital-twins-for-population-modeling-ethics-signal-quality-and-public-good-1gk6</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Digital twins of populations promise better planning for health, climate, and mobility. They also invite new risks: biased inputs, opaque assumptions, and governance gaps. This outline helps teams design for fidelity and legitimacy from day one.&lt;/p&gt;

&lt;h2&gt;
  
  
  What a population twin is (and is not)¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;A &lt;strong&gt;decision sandbox&lt;/strong&gt; that mirrors behaviours, constraints, and feedback loops across demographics.&lt;/li&gt;
&lt;li&gt;Not an oracle: outputs are probabilistic and highly sensitive to data quality and model design.&lt;/li&gt;
&lt;li&gt;Most valuable when paired with real-world sensing and community validation.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Core design questions¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Purpose clarity:&lt;/strong&gt; what policy or operational decisions will the twin inform, and who is accountable for them?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Boundary setting:&lt;/strong&gt; which variables are in-scope (health, mobility, energy) and which are out-of-scope to prevent mission creep?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Granularity:&lt;/strong&gt; choose spatial and temporal resolution that balances utility with re-identification risk.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data ethics and privacy¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Default to &lt;strong&gt;data minimisation&lt;/strong&gt; and use synthetic data where possible to test model behaviour.&lt;/li&gt;
&lt;li&gt;Apply &lt;strong&gt;differential privacy&lt;/strong&gt; or cohort-level aggregation for sensitive attributes.&lt;/li&gt;
&lt;li&gt;Maintain &lt;strong&gt;data lineage&lt;/strong&gt; logs so every forecast can be traced back to source datasets and preprocessing steps.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Model integrity¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Stress-test for &lt;strong&gt;representation gaps&lt;/strong&gt; (rural vs. urban, age groups, low-connectivity regions).&lt;/li&gt;
&lt;li&gt;Use &lt;strong&gt;causal structures&lt;/strong&gt; where possible to avoid spurious correlations driving policy.&lt;/li&gt;
&lt;li&gt;Require &lt;strong&gt;scenario audits&lt;/strong&gt; before deployment: best-case, base-case, worst-case outcomes with distributional impacts.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Participatory governance¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Establish a &lt;strong&gt;civic oversight panel&lt;/strong&gt; with community groups, domain experts, and policy leads.&lt;/li&gt;
&lt;li&gt;Publish &lt;strong&gt;assumption cards&lt;/strong&gt; that explain data sources, known gaps, and model limits in plain language.&lt;/li&gt;
&lt;li&gt;Provide &lt;strong&gt;challenge mechanisms&lt;/strong&gt; so affected communities can contest outputs and propose corrections.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Operational playbook¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Start with a &lt;strong&gt;narrow pilot&lt;/strong&gt; (e.g., vaccination logistics in one region) before expanding nationally.&lt;/li&gt;
&lt;li&gt;Set &lt;strong&gt;refresh cadences&lt;/strong&gt; for both data and model parameters; stale twins erode trust quickly.&lt;/li&gt;
&lt;li&gt;Integrate &lt;strong&gt;early warning alerts&lt;/strong&gt; when forecasts diverge from observed ground truth beyond agreed thresholds.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Population digital twins can unlock smarter, fairer planning only if they are transparent, auditable, and co-owned. Designing for accountability early prevents backlash later and keeps the focus on public value.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/digital-twins-for-population-modeling-ethics-signal-quality-and-public-good" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>digitaltwins</category>
      <category>populationhealth</category>
      <category>datagovernance</category>
      <category>aiethics</category>
    </item>
    <item>
      <title>Designing Digital Public Squares: Dialog Tools That Earn Trust</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:07:30 +0000</pubDate>
      <link>https://forem.com/techethics/designing-digital-public-squares-dialog-tools-that-earn-trust-5f3</link>
      <guid>https://forem.com/techethics/designing-digital-public-squares-dialog-tools-that-earn-trust-5f3</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Democratic governance depends on the ability of people to deliberate together about the decisions that affect their lives. For most of history, this deliberation happened in physical spaces: town halls, public meetings, community centres, and legislative chambers. The digital era has promised to expand participation far beyond the constraints of geography and schedule, but the platforms that dominate online discourse were never designed for deliberation. They were designed for engagement, and the difference between the two is the difference between a conversation that builds understanding and a feed that amplifies outrage.&lt;/p&gt;

&lt;p&gt;Public institutions and civic technologists now face a design challenge that is as much about democratic theory as it is about software engineering: how to build digital spaces where people can debate, disagree, converge on shared priorities, and see their input genuinely influence outcomes. This article outlines the principles, features, and implementation practices that distinguish platforms capable of earning and sustaining public trust.&lt;/p&gt;

&lt;h2&gt;
  
  
  Product Principles¶
&lt;/h2&gt;

&lt;p&gt;The first principle is safety by design. Public deliberation on contentious issues, from housing policy to immigration to climate adaptation, will always attract bad-faith actors, coordinated disruption, and heated exchanges that can escalate into harassment. Systems that treat moderation as an afterthought, bolted on once problems emerge, will always be reactive and overwhelmed. Effective platforms build friction into the architecture itself: rate limiting that prevents flooding, civility nudges that prompt users to reconsider hostile phrasing before posting, and clear escalation paths that route sensitive or threatening content to trained human moderators rather than relying solely on automated filters.&lt;/p&gt;

&lt;p&gt;The second principle is deliberation over virality. The ranking signals that drive commercial social media, novelty, engagement, and emotional intensity, are precisely the wrong signals for public deliberation. Platforms designed for democratic dialogue must instead prioritise argument quality, diversity of viewpoints, and evidence. This means surfacing contributions that introduce new perspectives or cite verifiable sources, rather than contributions that generate the most reactions. It means designing share flows that encourage reflection rather than reflexive amplification.&lt;/p&gt;

&lt;p&gt;The third principle is plain-language accessibility. Policy discussions are frequently conducted in jargon that excludes the majority of the people affected by those policies. Digital public squares must actively demystify this language, providing contextual glossaries, plain-language summaries of complex proposals, and clear explanations of how citizen input will be used. If people cannot understand what they are being asked to comment on, participation becomes performative rather than substantive.&lt;/p&gt;

&lt;h2&gt;
  
  
  Core Features to Prioritise¶
&lt;/h2&gt;

&lt;p&gt;Structured prompts are the foundation of productive deliberation. Rather than presenting an open text box and hoping for coherent contributions, effective platforms frame discussions around specific questions, present the key trade-offs at stake, and provide background materials that ground the conversation in evidence. Taiwan’s vTaiwan platform and Barcelona’s Decidim have both demonstrated that structured formats consistently produce higher-quality input and broader participation than unstructured forums.&lt;/p&gt;

&lt;p&gt;Argument maps provide a visual layer that clusters claims, evidence, and counterpoints, reducing the repetition that plagues traditional comment threads and helping participants see where consensus exists and where genuine disagreement remains. These maps also serve a transparency function: they make the structure of a debate legible to newcomers and to the decision-makers who will act on its results.&lt;/p&gt;

&lt;p&gt;Verification tiers allow platforms to balance openness with accountability. Lightweight identity checks may be sufficient for general discussion, while stronger verification is appropriate for phases where input directly influences binding decisions. Anonymous participation modes, essential for protecting vulnerable voices in sensitive consultations, can coexist with verified modes through careful design that preserves both safety and legitimacy.&lt;/p&gt;

&lt;p&gt;An evidence locker, a shared repository of sources with credibility signals, versioning, and citation tracking, raises the quality of discourse by making it easy to ground claims in verifiable information and difficult to sustain assertions that have already been refuted. When evidence is accessible, shared, and transparent, the cost of misinformation rises and the quality of deliberation improves.&lt;/p&gt;

&lt;h2&gt;
  
  
  Participation Equity¶
&lt;/h2&gt;

&lt;p&gt;The most carefully designed platform is useless if it is only accessible to the digitally fluent, the well-connected, and the already-engaged. Participation equity requires deliberate effort across multiple dimensions. Translation is a minimum requirement, not just of interfaces but of summaries, prompts, and outcomes. Text-only formats exclude people with low literacy or visual impairments; audio channels, SMS integration, and video summaries expand the tent significantly.&lt;/p&gt;

&lt;p&gt;Community moderators who reflect the demographics of participants bring cultural competence that no automated system can replicate. Training these moderators in trauma-informed facilitation is essential for consultations that touch sensitive topics, from refugee resettlement to transitional justice to community policing. Accessibility defaults, including captioning, screen reader support, keyboard navigation, and high-contrast modes, must be built into the platform from the start rather than retrofitted after disability advocacy groups file complaints.&lt;/p&gt;

&lt;p&gt;Recruitment also matters. Platforms that rely solely on self-selection will consistently over-represent the motivated, the opinionated, and the digitally comfortable. Sortition-based panels, partnerships with community organisations, and targeted outreach to underrepresented groups all help ensure that the voices in the digital square reflect the community it serves.&lt;/p&gt;

&lt;h2&gt;
  
  
  Making Dialog Count¶
&lt;/h2&gt;

&lt;p&gt;Participation without consequence breeds cynicism. The single most important feature of any civic deliberation platform is visible impact: participants must be able to see how their input influenced the decision it was meant to inform. Decision hooks that trace the path from citizen contribution to policy draft, budget allocation, or programme design transform participation from an exercise in venting into a genuine act of governance.&lt;/p&gt;

&lt;p&gt;Response service-level agreements create accountability on the institutional side. Acknowledging submissions within days, publishing synthesis reports within weeks, and closing the loop after decisions are made tells participants that their time was valued and their contributions were heard. Platforms that collect input and then go silent erode trust faster than platforms that never asked in the first place.&lt;/p&gt;

&lt;p&gt;Civic impact metrics replace vanity metrics. Rather than measuring success by sign-ups, page views, or comment counts, effective platforms track representation balance across demographic groups, completion rates for structured deliberations, evidence quality in contributions, and the proportion of policy outcomes that demonstrably incorporate citizen input. These metrics keep the platform honest about whether it is achieving its purpose or merely generating activity.&lt;/p&gt;

&lt;h2&gt;
  
  
  Implementation Playbook¶
&lt;/h2&gt;

&lt;p&gt;Launching a civic deliberation platform at full scale without testing is a recipe for failure. Pilot cohorts, drawn from trusted partners like libraries, youth councils, neighbourhood forums, and civil society organisations, allow teams to refine facilitation practices, identify technical issues, and build a body of evidence for what works before the stakes are high. These partners also become advocates who can credibly promote the platform to wider audiences.&lt;/p&gt;

&lt;p&gt;Time-boxed deliberations with clear start and end dates, rotating moderators, and published agendas prevent the fatigue and drift that afflict open-ended forums. When participants know that a consultation has a defined scope and timeline, they are more likely to engage seriously and less likely to disengage from frustration.&lt;/p&gt;

&lt;p&gt;Open data APIs that allow journalists, researchers, and civic watchdog organisations to audit contributions and outcomes add a layer of external accountability that keeps the platform and the institutions using it honest. Transparency at the infrastructure level, not just the interface level, is what distinguishes genuine civic technology from consultation theatre.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Digital public squares must be safe, comprehensible, and consequential. The platforms that earn trust are those that prove, through visible impact and transparent process, that citizen input genuinely changes outcomes. When participation deepens because people can see it matters, the democratic promise of digital deliberation begins to be fulfilled.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/designing-digital-public-squares-dialog-tools-that-earn-trust" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>civictech</category>
      <category>publicdialogue</category>
      <category>democracy</category>
      <category>participation</category>
    </item>
    <item>
      <title>AI, ML, and Fundamental Rights: Privacy, Equality, Fairness</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:07:13 +0000</pubDate>
      <link>https://forem.com/techethics/ai-ml-and-fundamental-rights-privacy-equality-fairness-1b28</link>
      <guid>https://forem.com/techethics/ai-ml-and-fundamental-rights-privacy-equality-fairness-1b28</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Artificial intelligence now shapes access to credit, housing, employment, healthcare, and justice. In each of these domains, the decisions an algorithm makes can determine whether a person receives a service or is denied it, whether they are surveilled or left in peace, and whether they are treated as an individual or reduced to a statistical profile. Because these decisions touch the rights that democratic societies have spent decades codifying into law, the intersection of AI and fundamental rights is no longer a theoretical concern. It is an operational reality that demands practical attention from everyone who builds, deploys, or regulates these systems.&lt;/p&gt;

&lt;p&gt;This article maps the most significant points of contact between AI and core human rights, traces the mechanisms through which harm occurs, and outlines the design and governance practices that can keep systems aligned with the values they are supposed to serve.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where Rights Intersect with AI¶
&lt;/h2&gt;

&lt;p&gt;The most immediate tension lies in privacy. Modern AI systems thrive on data, and the scale of collection required to train and operate them routinely exceeds what individuals knowingly consent to. Facial recognition cameras in public spaces, behavioural inference engines that predict purchasing intent from browsing patterns, and data fusion techniques that combine innocuous datasets to reveal sensitive attributes all expand the surveillance surface far beyond what traditional data protection frameworks were designed to address. The result is that people lose meaningful control over information about themselves, often without realising it has happened.&lt;/p&gt;

&lt;p&gt;Equality and non-discrimination present a second, equally urgent challenge. Bias can enter an AI system at any point in its lifecycle: through training data that reflects historical patterns of exclusion, through proxy variables that correlate with protected characteristics without naming them, or through deployment contexts that impose disproportionate burdens on particular groups. A hiring algorithm trained on a decade of successful applicants at a company that historically favoured men will learn to replicate that preference. A credit scoring model that uses postcode as a feature will encode decades of housing segregation into its risk assessments. These are not edge cases; they are structural patterns that require deliberate effort to identify and correct.&lt;/p&gt;

&lt;p&gt;Due process and explainability form a third critical axis. When a decision that materially affects someone’s life is made or heavily influenced by an opaque algorithm, the ability to understand, challenge, and appeal that decision is undermined. Procedural fairness requires legible reasoning, and many machine learning models resist legible explanation by design. This is not merely an inconvenience; in domains like criminal justice, immigration, and welfare eligibility, it represents a direct erosion of rights that legal systems have recognised for centuries.&lt;/p&gt;

&lt;p&gt;Finally, there is the question of autonomy and dignity. Behavioural manipulation through hyper-personalised content, dark patterns that exploit cognitive biases, and recommendation systems engineered to maximise engagement at the expense of informed choice all erode the capacity for genuine consent and meaningful decision-making that underpins human agency.&lt;/p&gt;

&lt;h2&gt;
  
  
  System Lifecycle Checkpoints¶
&lt;/h2&gt;

&lt;p&gt;Rights-aware AI development begins long before any model is trained. At the problem framing stage, teams must validate that the objective they are optimising for does not encode exclusionary assumptions. A recidivism prediction model optimised purely for accuracy, for example, may achieve that accuracy by learning correlations that reproduce structural disadvantage. Defining unacceptable use cases up front, and documenting the reasons for those boundaries, creates accountability before the technical work begins.&lt;/p&gt;

&lt;p&gt;Data sourcing is the next critical checkpoint. Provenance, consent basis, and known gaps must all be documented. Representative sampling and balance checks help ensure that the populations the system will serve are adequately reflected in the data it learns from. Where data poverty exists for marginalised groups, this must be flagged as a limitation rather than papered over with synthetic augmentation that may introduce its own distortions.&lt;/p&gt;

&lt;p&gt;During model development, bias testing, drift analysis, and robustness evaluations should be standard practice. Interpretable performance slices for protected characteristics, where lawful and appropriate, allow teams to identify disparate impact before deployment rather than discovering it through complaints. This is also the stage where trade-offs between fairness metrics must be confronted honestly, since optimising for one definition of fairness often comes at the expense of another.&lt;/p&gt;

&lt;p&gt;Deployment and monitoring close the loop. Tracking disparate impact over time, logging decisions to enable redress, and establishing clear criteria for sunsetting models that fail fairness or privacy thresholds are all essential. The assumption that a model validated at launch will remain fair indefinitely is one of the most dangerous misconceptions in the field. Populations shift, contexts change, and feedback loops can amplify small initial biases into significant structural harms.&lt;/p&gt;

&lt;h2&gt;
  
  
  Remedies and Controls¶
&lt;/h2&gt;

&lt;p&gt;Privacy-by-design principles provide the foundation for data protection: minimisation, differential privacy, federated learning, and strict retention schedules reduce the attack surface and limit the scope for misuse. These are not aspirational goals but well-understood engineering practices with mature tooling available across major machine learning frameworks.&lt;/p&gt;

&lt;p&gt;Fairness-by-design requires a broader toolkit. Pre-processing techniques that rebalance training data, in-processing constraints that penalise discriminatory outcomes during training, and post-processing adjustments that calibrate outputs across demographic groups all have roles to play. Counterfactual testing, which asks whether the model’s output would change if a protected characteristic were different, provides a particularly intuitive and legally defensible form of bias detection. Impact assessments tied to concrete, pre-specified thresholds transform fairness from a vague aspiration into a measurable requirement.&lt;/p&gt;

&lt;p&gt;Rights to notice and contestation must be built into the user-facing layer. Clear explanations of how a decision was reached, accessible appeal channels, and human override for high-stakes contexts are not optional extras. They are legal requirements under frameworks like the GDPR and the EU AI Act, and they are practical necessities for maintaining the trust that any system operating at scale depends on.&lt;/p&gt;

&lt;p&gt;Governance structures tie these technical measures together. Assigning accountable owners for each system, maintaining model cards and data cards that document design choices and known limitations, and conducting regular audits against both policy and law create the institutional scaffolding that prevents good intentions from eroding under commercial pressure or operational convenience.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Rights-safe AI is a continuous practice, not a compliance checkbox. The systems that earn and maintain public trust are those whose builders treat privacy, equality, fairness, and due process as design constraints from the outset rather than afterthoughts to be addressed when regulators come calling. The cost of building these protections in is real, but it is consistently lower than the cost of repairing the damage when they are absent.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/ai-ml-and-fundamental-rights-privacy-equality-fairness" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>humanrights</category>
      <category>aigovernance</category>
      <category>fairness</category>
      <category>privacy</category>
    </item>
    <item>
      <title>When AI Harms the Vulnerable: Lessons from Refugee, Justice, and Humanitarian Contexts</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:06:57 +0000</pubDate>
      <link>https://forem.com/techethics/when-ai-harms-the-vulnerable-lessons-from-refugee-justice-and-humanitarian-contexts-2ja9</link>
      <guid>https://forem.com/techethics/when-ai-harms-the-vulnerable-lessons-from-refugee-justice-and-humanitarian-contexts-2ja9</guid>
      <description>&lt;h1&gt;
  
  
  When AI Harms the Vulnerable: Lessons from Refugee, Justice, and Humanitarian Contexts¶
&lt;/h1&gt;

&lt;h2&gt;
  
  
  Introduction¶
&lt;/h2&gt;

&lt;p&gt;Artificial intelligence systems are rarely tested most rigorously in comfortable conditions. They are tested at borders in the middle of the night, in bail hearings where a wrong prediction can mean months in pre-trial detention, and in disaster zones where connectivity is intermittent and data is incomplete. It is in these environments - high-stakes, resource-constrained, and populated by people who have the least power to push back - that the weaknesses of AI systems are exposed first and felt most acutely.&lt;/p&gt;

&lt;p&gt;At &lt;a href="https://techethics.co.uk/" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt;, our mission is to develop technology that strengthens democratic resilience and social cohesion - which means engaging seriously with the cases where technology has done the opposite. This article examines documented and representative failures of AI deployment across three domains: refugee and asylum processing, criminal justice, and humanitarian response. In each domain, we trace the specific mechanisms through which harm occurred, identify the design and governance failures that enabled it, and set out the practical safeguards that responsible practitioners should build in before deployment - not after.&lt;/p&gt;

&lt;p&gt;The goal is not to condemn AI as unsuitable for these contexts. On the contrary, thoughtfully designed systems can dramatically improve efficiency, consistency, and outcomes. But the path to those benefits runs directly through an honest reckoning with what has already gone wrong.&lt;/p&gt;




&lt;h2&gt;
  
  
  Part One: Refugee and Asylum Processing¶
&lt;/h2&gt;

&lt;h3&gt;
  
  
  The Promise and the Problem¶
&lt;/h3&gt;

&lt;p&gt;Asylum systems are chronically under-resourced. Caseloads in many jurisdictions have grown faster than adjudicator capacity for years, creating backlogs that stretch to years and impose profound uncertainty on people fleeing persecution. AI tools that can triage cases, flag inconsistencies, or prioritise urgent claims hold genuine appeal.&lt;/p&gt;

&lt;p&gt;The problem is that asylum adjudication is among the most context-sensitive, linguistically complex, and existentially consequential decisions a government makes. A missed nuance in a survivor’s testimony - a culturally specific expression of distress, an indirect reference to sexual violence common in certain communities, a tribal term with no clean translation - can convert a legitimate fear of persecution into a score that suggests fabrication. The model does not know what it does not know, and neither does the caseworker who trusts it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Risk Scoring Without Transparency or Appeal¶
&lt;/h3&gt;

&lt;p&gt;Several national immigration authorities have piloted automated risk scoring tools to assist in processing asylum claims. These systems typically ingest application data, cross-reference it against watchlists and inconsistency flags, and produce a score or recommendation that influences how quickly a case is heard and whether the claimant is detained pending decision.&lt;/p&gt;

&lt;p&gt;In practice, these scores have been used to de-prioritise or deny services on the basis of correlations that are opaque to applicants, their legal representatives, and, in many cases, the caseworkers themselves. When a claimant’s risk score is elevated because their route of travel matched a pattern associated with smuggling networks - without any mechanism to account for the fact that asylum seekers frequently have no legal means of travel - there is no channel through which that error can surface and be corrected. The applicant is simply in a slower lane, or in detention, without knowing why.&lt;/p&gt;

&lt;p&gt;This opacity is not merely an ethical problem - it is increasingly a legal one. As our guide to &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;UK and EU AI regulation&lt;/a&gt; sets out, Article 22 of the GDPR establishes a general prohibition on decisions “based solely on automated processing” that produce legal or similarly significant effects on individuals, with mandatory requirements for human intervention, transparency, and the right to contest. Asylum and immigration decisions squarely meet this threshold. Authorities deploying scoring tools without genuine contestation mechanisms are not merely falling short of best practice - they are operating in likely violation of data protection law on both sides of the Channel.&lt;/p&gt;

&lt;p&gt;The harm is also structural. When AI systems create outcomes without legible reasons, identifying systematic bias - say, that claims from a particular country of origin are scored less favourably - requires external audit rather than routine process. Errors compound silently.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Language Model Misclassification of Asylum Narratives¶
&lt;/h3&gt;

&lt;p&gt;A separate class of failures involves natural language processing tools applied to the written or transcribed accounts that form the core of many asylum claims. These tools have been deployed to assess credibility, detect inconsistency across multiple interviews, and in some pilots, produce summaries for adjudicators. The training data for such models is almost always dominated by formal, literate, Western-European-language text.&lt;/p&gt;

&lt;p&gt;Asylum seekers frequently communicate in ways that create systematic disadvantages in these systems. Oral narrative traditions from parts of sub-Saharan Africa, the Middle East, and Southeast Asia do not organise chronology the way that Western legal testimony expects. Trauma affects memory and coherence in ways that are well-documented in clinical and legal scholarship but poorly represented in model training data. Dialects and regional idioms - particularly those of minority communities who may face persecution precisely because of their linguistic identity - are underrepresented or absent. A model trained on standard Modern Standard Arabic will not handle Levantine dialect in the same way; a model trained on Northern Somali variants may mishandle Southern ones.&lt;/p&gt;

&lt;p&gt;When these systems flag narrative features as indicators of unreliability, they do so based on patterns that reflect the structure of the training corpus more than the truth-value of the claimant’s account. The result is that individuals most at risk - members of minority communities, people with significant trauma histories, those without formal education - receive systematically worse outputs than those whose communication style happens to match the training distribution. This is precisely the kind of proxy discrimination that the &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;EU AI Act’s risk classification&lt;/a&gt; targets as a high-risk application requiring conformity assessment before deployment.&lt;/p&gt;

&lt;h3&gt;
  
  
  Safeguards: What Responsible Deployment Looks Like¶
&lt;/h3&gt;

&lt;p&gt;The failures above share a common structure: consequential decisions are made or heavily influenced by AI systems that lack transparency, mechanisms for challenge, and were never evaluated against the populations they would actually serve. The fixes are demanding, but they are not mysterious.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Human-in-the-loop adjudication&lt;/strong&gt; must be a genuine requirement, not a formality. This means AI outputs are advisory inputs to a human decision-maker who is trained to understand the tool’s limitations and is not incentivised simply to accept machine recommendations. Targets that reward rapid case resolution create structural pressure to defer to automation; governance structures must counteract this. Our &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;AI Ethics &amp;amp; Governance Reviews&lt;/a&gt; are specifically designed to help organisations assess whether their human oversight mechanisms are genuine or nominal - a distinction that matters both legally and ethically.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Culturally aware evaluation datasets&lt;/strong&gt; should be built in partnership with linguistic communities and civil society organisations that work directly with asylum seekers. Evaluation of NLP tools must include performance disaggregated by language, dialect, nationality, and trauma history before deployment, and must be repeated at intervals after.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mandatory explanation and contestation channels&lt;/strong&gt; mean that any AI-influenced outcome must be accompanied by a legible account of the factors that contributed to it, expressed in terms that the applicant and their representative can engage with. There must be a formal process for contesting that account, accessible without requiring legal expertise or resources the applicant does not have.&lt;/p&gt;




&lt;h2&gt;
  
  
  Part Two: Criminal Justice¶
&lt;/h2&gt;

&lt;h3&gt;
  
  
  The Legitimacy Problem¶
&lt;/h3&gt;

&lt;p&gt;Criminal justice is the domain where AI failures carry the clearest coercive power. A miscalibrated recidivism score does not merely inconvenience someone - it can keep them in prison or impose conditions that destabilise their life and employment. A predictive policing algorithm that over-predicts crime in a neighbourhood does not just misallocate patrol resources - it generates the arrests that then train future models to over-predict again.&lt;/p&gt;

&lt;p&gt;The use of algorithmic tools across the justice system has expanded significantly over the past two decades, driven by a combination of genuine efficiency pressures, a cultural faith in quantification as objectivity, and commercial interests from vendors who are rarely accountable for the downstream effects of their products. The academic and journalistic record of failures is now substantial.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Predictive Policing and the Feedback Loop¶
&lt;/h3&gt;

&lt;p&gt;Predictive policing systems use historical crime data - typically arrest records and incident reports - to generate risk scores for geographic areas or, in some implementations, individuals. The output is used to direct patrol resources: more officers to high-score areas, potentially more stops and searches, and more resulting arrests.&lt;/p&gt;

&lt;p&gt;The structural problem is that arrest data does not measure crime - it measures policing. Communities that have historically been over-policed generate more arrest records, which feed into models that predict more risk, which justify more policing. The system is not learning about crime; it is learning about itself. This feedback loop has been documented in deployments in Chicago, Los Angeles, New Orleans, and a number of European cities, and it produces outcomes that are both empirically invalid and deeply discriminatory.&lt;/p&gt;

&lt;p&gt;The legal landscape has responded. The EU AI Act now explicitly prohibits AI-based individual predictive policing based solely on profiling - one of the banned practices enforceable &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;since February 2025&lt;/a&gt;. In the UK, the landmark &lt;em&gt;R (Bridges) v Chief Constable of South Wales Police&lt;/em&gt; [2020] EWCA Civ 1058 established that public authorities must proactively investigate potential algorithmic bias &lt;strong&gt;before deployment&lt;/strong&gt;, not retrospectively - a precedent with implications well beyond facial recognition. Despite these developments, existing systems in many jurisdictions continue operating without the retrospective review these rulings demand.&lt;/p&gt;

&lt;p&gt;Beyond the statistical problem is the question of legitimacy. When a resident of a high-score neighbourhood is stopped or searched, the basis for that interaction is, in part, an algorithm that learned to flag their neighbourhood from data that reflected prior discriminatory policing. The accountability chain that should connect police action to articulable suspicion has been replaced by an opaque output from a commercial system, the details of which are often protected as proprietary information.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Recidivism Scoring in Bail and Sentencing¶
&lt;/h3&gt;

&lt;p&gt;Tools like COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) and its equivalents have been used across the United States and in several other jurisdictions to inform decisions about pre-trial detention, bail conditions, and in some cases sentencing. The ProPublica investigation of COMPAS in 2016, examining records from Broward County, Florida, found that the system’s false positive rate for Black defendants was nearly double that for white defendants: Black defendants who did not reoffend were predicted to be higher risk at significantly greater rates than white defendants in the same situation.&lt;/p&gt;

&lt;p&gt;The features these models use are rarely disclosed in detail, but research has shown that proxies for race - neighbourhood, education, family history of incarceration - are frequently included. Since these proxies are correlated with race through decades of discriminatory policy in housing, education, and law enforcement, the model encodes historical structural disadvantage and applies it to individuals as if it were a predictive signal about their future behaviour.&lt;/p&gt;

&lt;p&gt;The judicial oversight problem compounds this. Judges using these scores often receive a number or category - “medium risk” - without the feature contributions that drove it, without the error rates for the demographic group the defendant belongs to, and without training adequate to critically assess what they are being given. Studies have found that judges who defer to algorithmic recommendations rather than exercising independent judgment produce worse outcomes and perpetuate bias more reliably than those who treat scores as one input among many. This is the “rubber-stamping” problem that GDPR Article 22 guidance specifically identifies as failing to constitute meaningful human oversight - an issue discussed in detail in our &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;AI regulation overview&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Safeguards: What Responsible Deployment Looks Like¶
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Exclude protected proxies and their correlates.&lt;/strong&gt; Any feature substantially correlated with a protected characteristic requires explicit justification and should be subject to adversarial review. The argument that a proxy is “predictive” does not justify its use if it is predictive partly because it encodes structural disadvantage. Our &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;algorithmic bias audits&lt;/a&gt; provide exactly this kind of adversarial scrutiny, working through feature sets to identify indirect discrimination pathways before systems go live.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conduct mandatory disparate impact audits before and after deployment.&lt;/strong&gt; Performance metrics - accuracy, false positive rates, false negative rates - must be disaggregated by demographic group, and a decision to deploy must require that disparities fall within pre-specified tolerances. The &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;UK’s Algorithmic Transparency Recording Standard&lt;/a&gt;, now mandatory for central government and recommended for all public bodies, provides a useful baseline framework for documentation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Publish model cards and evaluation documentation.&lt;/strong&gt; Vendors who sell AI tools to justice system actors should be required to publish detailed model cards specifying training data, feature sets, known limitations, evaluation methodology, and demographic performance breakdowns. Proprietary protection of these details is legally incompatible with their use in consequential decisions. Defendants must have the right to access and challenge the algorithmic basis of decisions made about them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Require genuine judicial oversight.&lt;/strong&gt; AI tools should have no automatic authority in the justice system. Scores should be advisory, accompanied by confidence intervals and error rates for the relevant demographic group, and the decision-maker who relies on them should be able to articulate why the score was or was not weighted in their reasoning.&lt;/p&gt;




&lt;h2&gt;
  
  
  Part Three: Humanitarian Response¶
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Data Poverty and the Invisible¶
&lt;/h3&gt;

&lt;p&gt;Humanitarian response - the logistical challenge of delivering aid, shelter, medical care, and protection to people in the acute phase of disaster or conflict - has long attracted interest from the data science community. The problems are real and significant: resources are scarce, needs are vast and geographically dispersed, and the window for effective intervention is often narrow. AI tools for needs assessment, supply chain optimisation, and beneficiary registration hold genuine promise.&lt;/p&gt;

&lt;p&gt;The critical challenge is that the populations most in need of humanitarian assistance are often the least represented in the data that AI systems train and operate on. Communities that are geographically remote, linguistically marginalised, or deliberately displaced are not generating digital footprints at the rate of connected urban populations. When resource allocation algorithms are trained on connectivity patterns, mobile data, or administrative records, they systematically underestimate need in precisely the places where need is greatest.&lt;/p&gt;

&lt;p&gt;This problem sits at the heart of why TechEthics’ &lt;a href="https://techethics.co.uk/solutions/atlas" rel="noopener noreferrer"&gt;Atlas conflict mapping platform&lt;/a&gt; is designed around geospatial and community-level data sources rather than digital footprint proxies - recognising that the communities who need early warning most urgently are those whose signals are hardest to detect through conventional data pipelines.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Resource Allocation and the Unconnected¶
&lt;/h3&gt;

&lt;p&gt;During several large-scale displacement crises, AI-assisted needs assessment tools were piloted by humanitarian organisations. These tools aggregated mobile phone data, satellite imagery, and social media signals to map population movement and estimate resource requirements. The outputs informed decisions about where to position food distribution points, medical teams, and shelter materials.&lt;/p&gt;

&lt;p&gt;Post-distribution monitoring consistently showed that the populations furthest from distribution points - those who walked multiple days to access aid, those in areas with no mobile coverage, those from communities that did not use formal mobile networks - had been systematically under-counted by the algorithmic assessment. Resources were directed toward populations the model could see, not necessarily toward those with the greatest need. The model had operationalised connectivity as a proxy for presence and visibility as a proxy for need.&lt;/p&gt;

&lt;p&gt;This is not a technical edge case - it is a predictable consequence of training data that reflects the digital divide. Addressing it requires &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;conflict-sensitive technology design&lt;/a&gt; that begins with explicit interrogation of whose data is absent from the training set, and what that absence means for the outputs. It also requires institutional willingness to prioritise the unconnected, which is not always present when speed and scale are the dominant operational pressures.&lt;/p&gt;

&lt;h3&gt;
  
  
  Case: Facial Recognition at Checkpoints and the Cost of False Positives¶
&lt;/h3&gt;

&lt;p&gt;Facial recognition has been piloted in humanitarian settings for beneficiary verification - preventing duplicate registration and ensuring aid reaches intended recipients. The efficiency rationale is genuine: registration fraud is a real problem that diverts resources from intended beneficiaries.&lt;/p&gt;

&lt;p&gt;The asymmetry of error costs has not been adequately reckoned with. In commercial facial recognition deployments, a false positive - incorrectly matching an individual to a record - is typically a minor inconvenience. In a humanitarian checkpoint context, a false positive match to a watchlist or a deduplication error can mean denial of food or medical care, detention for questioning, or, in conflict contexts, exposure to armed actors. The tolerance for false positives appropriate in a smartphone unlock system is several orders of magnitude higher than what is appropriate at a food distribution point in an active conflict zone.&lt;/p&gt;

&lt;p&gt;Systems trained predominantly on faces from North American and European datasets perform measurably worse on darker-skinned faces, as documented extensively in research from the MIT Media Lab and NIST. The populations served by humanitarian operations - predominantly from Africa, the Middle East, South Asia, and Southeast Asia - are exactly the populations for whom commercial facial recognition systems have the highest error rates. The EU AI Act’s prohibition on real-time biometric identification in public spaces except under strictly limited conditions, &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;now enforceable&lt;/a&gt;, is relevant even in humanitarian contexts: the humanitarian label does not exempt a system from the obligation to perform reliably on the people it is processing.&lt;/p&gt;

&lt;h3&gt;
  
  
  Safeguards: What Responsible Deployment Looks Like¶
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Data minimisation&lt;/strong&gt; should be the default stance in humanitarian AI. The question is not “what data could we collect and use?” but “what is the minimum data required to achieve the specific operational purpose?” This matters both for privacy - humanitarian data in conflict zones can endanger lives if accessed by the wrong parties - and for model validity, since minimal and targeted data collection is more likely to be representative than comprehensive collection that systematically excludes the most vulnerable.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Offline-first design&lt;/strong&gt; ensures that systems degrade gracefully when connectivity is unavailable, rather than excluding entire populations when the system cannot reach them. Critical functions - registration, verification, resource allocation - must operate without continuous connectivity, with synchronisation as a supplement rather than a requirement. TechEthics’ &lt;a href="https://techethics.co.uk/bespoke-development-services" rel="noopener noreferrer"&gt;bespoke development approach&lt;/a&gt; explicitly prioritises solutions designed for complex, real-world environments where infrastructure cannot be assumed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fallback manual processes&lt;/strong&gt; must be designed, resourced, and tested before deployment, not improvised when the system fails. Humanitarian operations in crisis conditions will always encounter system failures; the contingency plan must be as robust as the primary system.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Independent humanitarian ethics review&lt;/strong&gt;, distinct from standard organisational ethics review, should assess AI deployments against specialised frameworks - including the humanitarian principles of humanity, neutrality, impartiality, and operational independence - before deployment and at defined intervals thereafter. Our &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;Conflict &amp;amp; PeaceTech Advisory&lt;/a&gt; service is designed precisely for this kind of contextual review, drawing on decades of experience in post-conflict and fragile-state settings.&lt;/p&gt;




&lt;h2&gt;
  
  
  Part Four: Cross-Cutting Protections¶
&lt;/h2&gt;

&lt;p&gt;The domain-specific failures above share deeper structural causes. Addressing them requires governance measures that cut across all AI deployments in high-stakes contexts involving vulnerable populations.&lt;/p&gt;

&lt;h3&gt;
  
  
  Harm Thresholds and Kill Switches¶
&lt;/h3&gt;

&lt;p&gt;Every consequential AI deployment should be accompanied by pre-specified harm thresholds: defined conditions under which the system is automatically suspended or rolled back pending investigation. These thresholds should be set during the design phase, not retrospectively. They should include demographic performance disparities beyond defined tolerances, error rates above specified limits, and any pattern of outcomes that diverges significantly from baseline expectations established during evaluation. The EU AI Act’s risk management requirements under Article 9 - which mandate continuous monitoring and iterative risk assessment throughout a system’s lifecycle - formalise this principle as a legal obligation for high-risk systems. Our &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;governance framework development&lt;/a&gt; service helps organisations design these mechanisms before deployment, rather than reaching for them in crisis.&lt;/p&gt;

&lt;h3&gt;
  
  
  Participatory Design with Affected Communities¶
&lt;/h3&gt;

&lt;p&gt;The most consistently valuable insight about how AI systems will fail in a given context comes from the people who live in that context. Participatory design - engaging affected communities as genuine co-designers, not as focus group subjects or checkbox consultees - surfaces failure modes that technical teams will not identify in the lab. For asylum systems, this means working with refugee community organisations and legal aid providers from the earliest stages of specification. For justice tools, it means involving public defenders, impacted community groups, and formerly incarcerated people. For humanitarian tech, it means systematic partnership with local NGOs and community leaders.&lt;/p&gt;

&lt;p&gt;This is not simply an ethical commitment. It is an epistemic one. Affected communities are the most reliable source of information about how a system’s assumptions will fail in the specific context where it is deployed. Excluding that knowledge produces worse systems. TechEthics’ &lt;a href="https://techethics.co.uk/bespoke-development-services" rel="noopener noreferrer"&gt;co-design approach&lt;/a&gt; is built on this principle - every platform we develop is designed with end users rather than for them, with accessibility and data protection built in from the ground up rather than retrofitted. Our &lt;a href="https://techethics.co.uk/solutions/dialogai" rel="noopener noreferrer"&gt;DialogAI platform&lt;/a&gt; extends this further, providing structured digital consultation infrastructure for facilitated dialogue and real-time consensus detection in exactly the kinds of divided or fragile settings where participatory design is most important and most challenging.&lt;/p&gt;

&lt;h3&gt;
  
  
  Monitoring, Logging, and Redress¶
&lt;/h3&gt;

&lt;p&gt;Consequential decisions made with AI assistance must be logged in sufficient detail to allow retrospective audit. This means preserving the model version, input data, the output, and the human decision that followed, associated with each case. Without this, identifying systematic errors is impossible, accountability is illusory, and learning from failures cannot happen. The UK’s Algorithmic Transparency Recording Standard and the EU AI Act’s technical documentation requirements under Article 11 both operationalise this principle - though compliance remains uneven in practice.&lt;/p&gt;

&lt;p&gt;Alongside logging, affected individuals and communities must have meaningful access to redress: the ability to appeal AI-influenced decisions, to access the reasoning behind them, and to receive compensation where harm is documented. The cost of building redress mechanisms is far lower than the cost of repairing institutional trust after a pattern of unexplained harm becomes public.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Role of Vendors and Procurement¶
&lt;/h3&gt;

&lt;p&gt;A significant share of AI tools used in these contexts are procured from commercial vendors rather than built in-house. Procurement processes have rarely been adequate to the risks. Vendor contracts should specify evaluation requirements, performance standards disaggregated by demographic group, disclosure obligations, and liability for documented harm. Proprietary protection should not extend to the information required for meaningful accountability. As our &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;AI regulation guide&lt;/a&gt; notes, the enforcement trend is clear: regulators across jurisdictions are actively penalising AI systems that violate fundamental rights, and the accumulated penalties from cases like Clearview AI make robust procurement governance a business necessity, not just an ethical aspiration.&lt;/p&gt;

&lt;p&gt;Organisations that lack in-house capacity to conduct this level of vendor due diligence can commission &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;ethical impact assessments&lt;/a&gt; from independent specialists. This is consistently more effective than relying on vendor-provided documentation alone.&lt;/p&gt;




&lt;h2&gt;
  
  
  Conclusion¶
&lt;/h2&gt;

&lt;p&gt;The argument for AI in refugee processing, criminal justice, and humanitarian response is not wrong - it is incomplete. These systems face genuine capacity and consistency problems, and thoughtfully designed AI can help address them. But the same power asymmetries that make these populations vulnerable to other forms of institutional harm make them vulnerable to algorithmic harm as well. They have less capacity to identify when a system has erred, less leverage to demand correction, and less ability to absorb the consequences of mistakes.&lt;/p&gt;

&lt;p&gt;The safeguards described here - human oversight with genuine authority, culturally aware evaluation, transparency and contestation, participatory design, harm thresholds, and meaningful redress - are practical requirements that distinguish responsible from irresponsible deployment. They cost more at the design phase. They cost less than repairing harm after the fact, and incomparably less than the damage done to individuals and to the institutional trust that humanitarian and justice systems depend on.&lt;/p&gt;

&lt;p&gt;AI deployed in contact with vulnerable populations must earn the right to scale by demonstrating that it can be governed. That demonstration begins not with a successful pilot, but with a design process that treats the people who will be most affected not as data points, but as the primary stakeholders whose interests the system must serve. At TechEthics, that principle sits at the centre of everything we build - from &lt;a href="https://techethics.co.uk/solutions/veritas" rel="noopener noreferrer"&gt;Veritas&lt;/a&gt; and &lt;a href="https://techethics.co.uk/solutions/metis" rel="noopener noreferrer"&gt;Metis&lt;/a&gt; to the advisory work we do with governments, NGOs, and civil society organisations navigating these challenges.&lt;/p&gt;

&lt;p&gt;If you are deploying AI in high-stakes contexts and want to ensure your governance frameworks are fit for purpose, &lt;a href="https://techethics.co.uk/contact-us" rel="noopener noreferrer"&gt;get in touch with our team&lt;/a&gt;. We offer &lt;a href="https://techethics.co.uk/consultancy-services" rel="noopener noreferrer"&gt;AI ethics reviews, algorithmic bias audits, conflict-sensitive design consultancy, and bespoke development&lt;/a&gt; - all grounded in the same commitment to accountability and dignity that this article argues for.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Related reading: &lt;a href="https://techethics.co.uk/insights/uk-and-eu-ai-regulation-what-organisations-need-to-know-in-2025" rel="noopener noreferrer"&gt;UK and EU AI Regulation: What Organisations Need to Know in 2025&lt;/a&gt; · &lt;a href="https://techethics.co.uk/insights/misinformation-and-fake-news-a-guide-to-critical-information-literacy" rel="noopener noreferrer"&gt;Misinformation and ‘Fake News’: A Guide to Critical Information Literacy&lt;/a&gt; · &lt;a href="https://techethics.co.uk/insights/the-hidden-architects-of-division-how-social-medias-recommendation-engines-shape-our-reality" rel="noopener noreferrer"&gt;The Hidden Architects of Division: How Social Media’s Recommendation Engines Shape Our Reality&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/when-ai-harms-the-vulnerable-lessons-from-refugee-justice-and-humanitarian-contexts" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>humanitariantech</category>
      <category>justice</category>
      <category>riskmanagement</category>
      <category>casestudies</category>
    </item>
    <item>
      <title>Responsible AI Guidelines: Principles, Frameworks, and Emerging Global Standards</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:06:40 +0000</pubDate>
      <link>https://forem.com/techethics/responsible-ai-guidelines-principles-frameworks-and-emerging-global-standards-4o62</link>
      <guid>https://forem.com/techethics/responsible-ai-guidelines-principles-frameworks-and-emerging-global-standards-4o62</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Responsible AI is moving from slideware to enforceable standards. This outline surveys leading principles, governance patterns, and policy moves teams should track.&lt;/p&gt;

&lt;h2&gt;
  
  
  Core principles (converging themes)¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Lawfulness and rights:&lt;/strong&gt; privacy, non-discrimination, and due process by default.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Safety and robustness:&lt;/strong&gt; resilience to misuse, attacks, and drift; transparent incident handling.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transparency and explainability:&lt;/strong&gt; appropriate disclosure, traceability, and user-understandable explanations.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Accountability:&lt;/strong&gt; clear ownership, auditability, and effective remedy mechanisms.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Governance frameworks in practice¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Model and data cards&lt;/strong&gt;: artefacts that document purpose, limits, and evaluation results.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Risk tiers and gates&lt;/strong&gt;: stricter reviews for high-risk use (health, finance, employment, public sector).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Human oversight patterns&lt;/strong&gt;: approval workflows, escalation paths, and kill-switch criteria.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Vendor management&lt;/strong&gt;: contractual controls, assurance evidence, and third-party risk assessments.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Emerging standards and regulation¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;EU AI Act&lt;/strong&gt;: risk-based obligations, prohibited uses, documentation, and post-market monitoring.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;NIST AI RMF&lt;/strong&gt; and &lt;strong&gt;ISO/IEC 42001&lt;/strong&gt;: operational guidance for managing AI risk and governance.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data protection laws&lt;/strong&gt; (GDPR, adequacy regimes): lawful bases, DPIAs, and automated decision safeguards.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sector codes&lt;/strong&gt;: financial model risk guidelines, healthcare safety cases, and platform content policies.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Implementation playbook¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Start with a &lt;strong&gt;policy baseline&lt;/strong&gt;: what uses are in/out of scope; who signs off.&lt;/li&gt;
&lt;li&gt;Build a &lt;strong&gt;controls library&lt;/strong&gt; mapped to risks (privacy, fairness, robustness, security, transparency).&lt;/li&gt;
&lt;li&gt;Stand up &lt;strong&gt;assurance loops&lt;/strong&gt;: pre-deployment review, post-deployment monitoring, and incident retros.&lt;/li&gt;
&lt;li&gt;Publish &lt;strong&gt;transparency notes&lt;/strong&gt; for users and regulators; update as models evolve.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Responsible AI is a moving target, but directionally clear: risk-tiered controls, documented accountability, and demonstrable safety. Teams that align early reduce regulatory friction and earn user trust.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/responsible-ai-guidelines-principles-frameworks-and-emerging-global-standards" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>aigovernance</category>
      <category>policy</category>
      <category>standards</category>
      <category>compliance</category>
    </item>
    <item>
      <title>Balancing AI Innovation with Human Rights: Knowing When to Stop or Slow Down</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:06:23 +0000</pubDate>
      <link>https://forem.com/techethics/balancing-ai-innovation-with-human-rights-knowing-when-to-stop-or-slow-down-5bcg</link>
      <guid>https://forem.com/techethics/balancing-ai-innovation-with-human-rights-knowing-when-to-stop-or-slow-down-5bcg</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;The default posture in technology development is forward motion. Ship the feature, scale the product, iterate later. In most domains, this instinct serves both companies and users well. But artificial intelligence is not most domains. When an AI system determines who receives welfare benefits, who is flagged at a border checkpoint, or who is released on bail, the consequences of getting it wrong are not bugs to be patched in the next sprint. They are harms to real people, often those with the least capacity to push back or seek redress.&lt;/p&gt;

&lt;p&gt;Knowing when to pause, limit, or refuse deployment is not a failure of ambition. It is a discipline that separates responsible innovation from recklessness. This article offers a practical framework for making those decisions, grounded in the kinds of trade-offs that practitioners actually face.&lt;/p&gt;

&lt;h2&gt;
  
  
  Situations to Pause or Refuse¶
&lt;/h2&gt;

&lt;p&gt;Certain deployment contexts carry risks that no amount of technical optimisation can adequately mitigate without fundamental changes to how the system is governed. The clearest cases involve structural power imbalances: welfare eligibility decisions, immigration processing, criminal justice risk scoring, and employment screening all place AI systems in a position where errors or biases impose irreversible harms on people who have little or no ability to challenge the outcome. When the gap between the decision-maker and the person affected is wide, and the stakes are existential, the burden of proof for deployment should be correspondingly high.&lt;/p&gt;

&lt;p&gt;Low-consent environments present a related but distinct concern. Workplace surveillance systems, education proctoring tools, and public-space monitoring technologies operate in contexts where meaningful opt-out is effectively impossible. An employee cannot choose not to be monitored without choosing not to be employed. A student cannot decline proctoring software without declining to sit the exam. When consent is structurally coerced rather than freely given, the legitimacy of any data processing that follows is fundamentally compromised.&lt;/p&gt;

&lt;p&gt;Weak data foundations represent a third category of situations where pause is warranted. When the training data is sparse, unrepresentative, or heavily reliant on proxy variables, the system’s outputs are unlikely to be fair regardless of how sophisticated the model architecture is. Emotion recognition systems, which claim to infer internal states from facial expressions despite contested scientific validity, represent an extreme version of this problem. But the principle applies more broadly: if the data cannot support the claims being made about the system’s capabilities, deployment should wait until it can.&lt;/p&gt;

&lt;h2&gt;
  
  
  Oversight Patterns That Work¶
&lt;/h2&gt;

&lt;p&gt;When deployment proceeds, the question becomes what oversight structures can genuinely constrain the system’s behaviour and catch failures before they compound. Human-in-the-loop arrangements are the most commonly cited safeguard, but their effectiveness depends entirely on implementation. A human reviewer who processes hundreds of algorithmic recommendations per day and overrides fewer than one percent of them is not providing meaningful oversight; they are providing a compliance narrative. Genuine human-in-the-loop means the reviewer has the authority, the time, the training, and the institutional incentive to exercise independent judgment.&lt;/p&gt;

&lt;p&gt;Ethics and risk boards serve a valuable function when they have teeth: the authority to block or delay high-risk launches and the mandate to track mitigation commitments over time. The most effective boards are those that include external members with relevant domain expertise, operate with genuine independence from commercial pressures, and publish at least summary findings to maintain accountability. Boards that exist primarily to approve what has already been decided are worse than no board at all, because they create a false sense of security.&lt;/p&gt;

&lt;p&gt;Shadow mode trials, in which the AI system runs alongside human decision-makers without its outputs being acted on, provide a powerful way to evaluate real-world performance before the consequences become real. Comparing AI recommendations to human decisions across a meaningful sample reveals both the system’s strengths and its failure modes in the actual environment it will operate in, rather than in the sanitised conditions of a test dataset.&lt;/p&gt;

&lt;p&gt;Finally, every high-stakes deployment should have kill-switch criteria defined before launch: specific, measurable conditions under which the system is automatically suspended pending investigation. These might include error rates exceeding a defined threshold, demographic performance disparities beyond agreed tolerances, or rising complaint volumes from affected populations. Defining these criteria in advance, when judgment is not clouded by sunk costs and launch momentum, is essential to ensuring they are actually enforced.&lt;/p&gt;

&lt;h2&gt;
  
  
  Deciding to Proceed, Delay, or Stop¶
&lt;/h2&gt;

&lt;p&gt;The decision framework for any given deployment rests on three tests. The proportionality test asks whether the benefit the AI system delivers is commensurate with the rights risks it introduces and whether adequate mitigations exist. A system that modestly improves processing speed but introduces significant bias risks fails this test. A system that dramatically improves consistency in a domain plagued by arbitrary human variation may pass it, provided the mitigations are robust.&lt;/p&gt;

&lt;p&gt;The alternatives analysis asks whether simpler approaches, including rule-based systems, human processes, or existing workflows with targeted improvements, could achieve comparable outcomes with less risk. AI is not always the right tool, and the assumption that automation necessarily improves on human decision-making is often wrong, particularly in domains where context, nuance, and empathy are central to good outcomes.&lt;/p&gt;

&lt;p&gt;The evidence threshold asks whether the system has been validated to a standard appropriate to the stakes. This means demonstrated performance on the specific populations it will serve, robustness under realistic stress conditions, and clear remediation paths for the harms it might cause. A system validated on a convenience sample from a different jurisdiction or demographic context has not met this threshold, regardless of how impressive its headline accuracy figures appear.&lt;/p&gt;

&lt;h2&gt;
  
  
  Communicating Limits¶
&lt;/h2&gt;

&lt;p&gt;Transparency about what a system can and cannot do, where it is allowed to operate and where it is not, is both an ethical obligation and a practical strategy for building the trust that any deployment at scale requires. Publishing deployment boundaries, including explicit statements of prohibited uses and the reasoning behind them, signals that the organisation takes limits seriously rather than treating them as obstacles to be minimised.&lt;/p&gt;

&lt;p&gt;Providing genuine user recourse, through accessible appeal processes, meaningful human review of contested decisions, and fair compensation where harm occurs, creates the accountability structures that sustain public acceptance over time. Organisations that make these commitments before problems emerge are far better positioned than those that scramble to improvise them after a public incident.&lt;/p&gt;

&lt;p&gt;Sharing impact reviews, including honest assessments of what worked, what failed, and what was changed in response, builds institutional credibility in a way that marketing language never can. Restraint, communicated clearly, is a feature. It demonstrates that the organisation values the people its systems affect as much as the efficiencies those systems deliver.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Balancing innovation with rights is ultimately about discipline: clear stop rules that are enforced, oversight structures with genuine authority, and transparency about the trade-offs involved. The teams and organisations that practice this discipline protect not only the people their systems serve but the long-term legitimacy of the AI programmes they are building. In a field where public trust is fragile and hard-won, restraint is not a constraint on progress. It is a condition for it.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/balancing-ai-innovation-with-human-rights-knowing-when-to-stop-or-slow-down" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>humanrights</category>
      <category>riskmanagement</category>
      <category>aigovernance</category>
      <category>oversight</category>
    </item>
    <item>
      <title>Technology in Conflict Zones: Surveillance, Evidence, and Rights Risks</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:06:06 +0000</pubDate>
      <link>https://forem.com/techethics/technology-in-conflict-zones-surveillance-evidence-and-rights-risks-11i0</link>
      <guid>https://forem.com/techethics/technology-in-conflict-zones-surveillance-evidence-and-rights-risks-11i0</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;In conflict zones, technology amplifies both harm and accountability. This outline surfaces how surveillance, data collection, and digital evidence shape civilian risk and post-conflict justice.&lt;/p&gt;

&lt;h2&gt;
  
  
  Patterns of use and misuse¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Mass surveillance and tracking:&lt;/strong&gt; IMSI catchers, spyware, and commercial data brokers used to locate activists, journalists, and aid workers.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Biometric databases:&lt;/strong&gt; enrolment at checkpoints without consent; later repurposed for targeting or discrimination.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Open-source intelligence (OSINT):&lt;/strong&gt; citizen investigators geolocate atrocities; regimes scrape the same data to identify dissidents.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Connectivity control:&lt;/strong&gt; shutdowns and throttling to disrupt organising, coupled with targeted disinformation to fragment narratives.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Evidence and documentation¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Chain of custody:&lt;/strong&gt; secure capture, hashing, and metadata preservation to make digital evidence admissible.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Context over virality:&lt;/strong&gt; structured fact patterns (who/what/where/when) to avoid misattribution.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Safety by design:&lt;/strong&gt; redaction and face-blurring for civilian protection before publication.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Safeguards and governance¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data minimisation:&lt;/strong&gt; collect only what is necessary for protection or accountability; set deletion triggers.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Security baselines:&lt;/strong&gt; threat modeling, encrypted channels, and tamper-evident storage.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ethical release policies:&lt;/strong&gt; publish only when risk to individuals is mitigated and consent is obtained where feasible.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Tech in conflict is never neutral. Secure practices and clear governance are essential to reduce harm while preserving pathways to justice.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/technology-in-conflict-zones-surveillance-evidence-and-rights-risks" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>conflicttech</category>
      <category>humanrights</category>
      <category>surveillance</category>
      <category>accountability</category>
    </item>
    <item>
      <title>Humanitarian Tech Trade-offs: Privacy, Transparency, Automation, Autonomy</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:05:49 +0000</pubDate>
      <link>https://forem.com/techethics/humanitarian-tech-trade-offs-privacy-transparency-automation-autonomy-17he</link>
      <guid>https://forem.com/techethics/humanitarian-tech-trade-offs-privacy-transparency-automation-autonomy-17he</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Digital tools can speed aid and improve accountability, but they also create new risks for the communities they aim to help. This outline explores core ethical trade-offs and how to manage them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Privacy vs. transparency¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data minimisation&lt;/strong&gt; for beneficiaries vs. &lt;strong&gt;donor reporting&lt;/strong&gt; demands; use aggregated, delayed, or differential privacy where possible.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Consent fatigue:&lt;/strong&gt; build plain-language, layered notices; allow refusal without penalty.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Risk-adjusted disclosure:&lt;/strong&gt; share operational metrics without exposing individuals or vulnerable sites.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Automation vs. human judgment¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Triage algorithms&lt;/strong&gt; can prioritise cases but risk entrenching bias; keep humans in/on-the-loop for edge cases.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Escalation protocols:&lt;/strong&gt; clear thresholds for when humans override or halt automated decisions.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Explainability:&lt;/strong&gt; simple rationales for field staff and beneficiaries to contest or correct outputs.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Dependency vs. community autonomy¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Avoid lock-in to proprietary platforms; prefer &lt;strong&gt;portable data&lt;/strong&gt; and &lt;strong&gt;open standards&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;Co-design with local actors; include &lt;strong&gt;offline-first&lt;/strong&gt; modes to respect connectivity realities.&lt;/li&gt;
&lt;li&gt;Build &lt;strong&gt;handover plans&lt;/strong&gt; so communities can run or retire tools without external vendors.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Operational safeguards¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Security hygiene: least-privilege access, audit logs, breach playbooks.&lt;/li&gt;
&lt;li&gt;Data retention tied to mission timelines; delete after purpose is fulfilled.&lt;/li&gt;
&lt;li&gt;Independent ethics review and periodic community feedback loops.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Ethical humanitarian tech is about proportionate data use, meaningful human oversight, and respect for local agency. Balancing these tensions keeps digital interventions supportive rather than extractive.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/humanitarian-tech-trade-offs-privacy-transparency-automation-autonomy" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>humanitariantech</category>
      <category>ethics</category>
      <category>privacy</category>
      <category>governance</category>
    </item>
    <item>
      <title>Lessons from AI in Crisis Response and Post-Conflict Recovery</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:05:32 +0000</pubDate>
      <link>https://forem.com/techethics/lessons-from-ai-in-crisis-response-and-post-conflict-recovery-14f6</link>
      <guid>https://forem.com/techethics/lessons-from-ai-in-crisis-response-and-post-conflict-recovery-14f6</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;AI can speed aid and accountability, but field results depend on context, data quality, and governance. These cases highlight what worked, what failed, and how to improve.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI for crisis response¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Damage assessment from satellite/air imagery&lt;/strong&gt; accelerates resource allocation but can miss informal settlements; pair with local validation teams.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Demand forecasting for supplies&lt;/strong&gt; reduces stockouts yet struggles with fast-changing ground truth; keep human override and rapid re-training loops.&lt;/li&gt;
&lt;li&gt;Key safeguard: transparency about model confidence and clear escalation when predictions conflict with field reports.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Supply-chain tracing in reconstruction¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Ledger-based provenance&lt;/strong&gt; for construction materials can deter diversion but requires reliable on-ramps and tamper-resistant IDs.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Risk scoring vendors&lt;/strong&gt; helps spot corruption but may unfairly penalise small local firms; include appeals and manual review.&lt;/li&gt;
&lt;li&gt;Key safeguard: publish criteria, avoid black-box scoring, and rotate auditors to prevent capture.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data-based human rights monitoring¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Crowdsourced incident reporting&lt;/strong&gt; scales coverage but invites misinformation; use verification tiers and geolocation checks.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Automated pattern detection&lt;/strong&gt; can surface hotspots but risks false positives; blend OSINT with trusted local sources.&lt;/li&gt;
&lt;li&gt;Key safeguard: protect witnesses with redaction, consent gates, and secure storage; delay publication if safety risks persist.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Cross-cutting lessons¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Start in &lt;strong&gt;shadow mode&lt;/strong&gt; to calibrate against human assessments before acting on outputs.&lt;/li&gt;
&lt;li&gt;Invest in &lt;strong&gt;data quality pipelines&lt;/strong&gt; and feedback loops; retire models that drift beyond agreed thresholds.&lt;/li&gt;
&lt;li&gt;Maintain &lt;strong&gt;public transparency notes&lt;/strong&gt; summarising methods, limits, and mitigation steps.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Responsible AI in crises and recovery demands humility, validation, and continuous oversight. Treat models as decision aids - not decision makers - and build in contestability from the start.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/lessons-from-ai-in-crisis-response-and-post-conflict-recovery" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>crisisresponse</category>
      <category>supplychain</category>
      <category>humanrights</category>
      <category>casestudies</category>
    </item>
    <item>
      <title>EU vs. Global AI Standards: What Builders and Policymakers Need to Know</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:05:15 +0000</pubDate>
      <link>https://forem.com/techethics/eu-vs-global-ai-standards-what-builders-and-policymakers-need-to-know-346d</link>
      <guid>https://forem.com/techethics/eu-vs-global-ai-standards-what-builders-and-policymakers-need-to-know-346d</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;EU regulation sets a high bar, but global approaches vary. This outline compares frameworks and distils practices that travel well across jurisdictions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Regulatory contrasts¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;EU AI Act:&lt;/strong&gt; risk tiers, conformity assessment, post-market monitoring, prohibitions (e.g., certain biometric uses).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;US patchwork:&lt;/strong&gt; sectoral guidance, NIST AI RMF adoption, state privacy laws shaping automated decision notices.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Asia-Pacific:&lt;/strong&gt; differentiated strategies - sandboxing in Singapore, safety and security emphasis in China, rights-forward bills in Australia and India.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Standards bodies:&lt;/strong&gt; ISO/IEC 42001 (AI management systems), IEEE guidance, OECD AI principles as soft-law anchors.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Impact by stakeholder¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;NGOs:&lt;/strong&gt; documentation and DPIAs for grant compliance; clearer contestability for affected communities.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Industry:&lt;/strong&gt; design controls for high-risk systems, supply-chain assurance, and harmonised model documentation.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Governments:&lt;/strong&gt; procurement standards, vendor audits, and public-sector transparency to set market norms.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Best-practice recommendations¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Build a &lt;strong&gt;jurisdiction-agnostic controls stack&lt;/strong&gt;: data governance, model cards, human oversight, incident response.&lt;/li&gt;
&lt;li&gt;Use &lt;strong&gt;risk tiering&lt;/strong&gt; to prioritise assurance depth; map to EU high-risk categories even when not required.&lt;/li&gt;
&lt;li&gt;Maintain &lt;strong&gt;portability&lt;/strong&gt;: modular policies and technical logs that can be tailored to local law with minimal rework.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Convergence is forming around risk management, transparency, and accountability. Preparing for EU-level rigor positions teams to meet or exceed other regimes with minimal friction.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/eu-vs-global-ai-standards-what-builders-and-policymakers-need-to-know" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>airegulation</category>
      <category>standards</category>
      <category>policycomparison</category>
      <category>governance</category>
    </item>
    <item>
      <title>Calls for Standards and Ethical Governance: Balancing Innovation, Accountability, and Rights</title>
      <dc:creator>Tony Robinson</dc:creator>
      <pubDate>Wed, 08 Apr 2026 14:04:59 +0000</pubDate>
      <link>https://forem.com/techethics/calls-for-standards-and-ethical-governance-balancing-innovation-accountability-and-rights-335o</link>
      <guid>https://forem.com/techethics/calls-for-standards-and-ethical-governance-balancing-innovation-accountability-and-rights-335o</guid>
      <description>&lt;h1&gt;
  
  
  Introduction¶
&lt;/h1&gt;

&lt;p&gt;Innovation and rights do not have to be at odds - but they do require disciplined governance. This outline sketches a balanced approach to standards, accountability, and public interest.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why governance now¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Rapid deployment of frontier models without matching safety evidence.&lt;/li&gt;
&lt;li&gt;Rising regulatory momentum and public concern about discrimination, privacy, and misinformation.&lt;/li&gt;
&lt;li&gt;Need for predictable rules so responsible builders can ship with confidence.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Elements of effective governance¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Risk-based tiers&lt;/strong&gt; with proportional controls and independent review for high-stakes uses.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Transparent documentation&lt;/strong&gt;: model cards, data cards, and release notes with limitations and known risks.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Human oversight&lt;/strong&gt;: clear override authority, escalation paths, and kill-switch criteria.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Accountability and remedy&lt;/strong&gt;: incident reporting, audits, and accessible channels for contestation and redress.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Calls to action¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;For &lt;strong&gt;industry&lt;/strong&gt;: adopt open standards, publish evaluation summaries, and align incentives to safety metrics.&lt;/li&gt;
&lt;li&gt;For &lt;strong&gt;NGOs and civil society&lt;/strong&gt;: participate in standards development, push for community consultation, and monitor impacts on vulnerable groups.&lt;/li&gt;
&lt;li&gt;For &lt;strong&gt;governments&lt;/strong&gt;: set procurement baselines, fund public-good evaluations, and require post-market monitoring for high-risk AI.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Balancing innovation and rights¶
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Encourage &lt;strong&gt;sandboxing&lt;/strong&gt; with guardrails and transparency rather than blanket bans.&lt;/li&gt;
&lt;li&gt;Invest in &lt;strong&gt;evaluation infrastructure&lt;/strong&gt; (benchmarks, red-teaming) to close the gap between lab metrics and real-world risk.&lt;/li&gt;
&lt;li&gt;Promote &lt;strong&gt;interoperable standards&lt;/strong&gt; so compliance is cumulative, not fragmented across jurisdictions.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion¶
&lt;/h1&gt;

&lt;p&gt;Clear standards and accountable governance enable innovation that earns trust. Acting now builds a safer, more rights-respecting AI ecosystem.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This article was originally published on the &lt;a href="https://techethics.co.uk" rel="noopener noreferrer"&gt;TechEthics&lt;/a&gt; website. &lt;a href="https://techethics.co.uk/insights/calls-for-standards-and-ethical-governance-balancing-innovation-accountability-and-rights" rel="noopener noreferrer"&gt;Read the original here&lt;/a&gt;. You can also explore our &lt;a href="https://techethics.co.uk/veritas" rel="noopener noreferrer"&gt;disinformation detection and analysis tools, Veritas&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>aigovernance</category>
      <category>ethics</category>
      <category>standards</category>
      <category>humanrights</category>
    </item>
  </channel>
</rss>
