<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Ross Peili</title>
    <description>The latest articles on Forem by Ross Peili (@rosspeili).</description>
    <link>https://forem.com/rosspeili</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F308648%2Fe3062c7b-4d7d-4ba9-aea5-f0ea4fef4776.jpg</url>
      <title>Forem: Ross Peili</title>
      <link>https://forem.com/rosspeili</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/rosspeili"/>
    <language>en</language>
    <item>
      <title>Your AI Needs a Physical Social Life</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Fri, 17 Apr 2026 10:03:20 +0000</pubDate>
      <link>https://forem.com/arpa/your-ai-needs-a-physical-social-life-3lce</link>
      <guid>https://forem.com/arpa/your-ai-needs-a-physical-social-life-3lce</guid>
      <description>&lt;p&gt;If you're deep into AI, you understand that the current state of Artificial Intelligence is a sterile, centralized hallucination. We are sprinting toward some sort of a god-box, a singular, omniscient entity hosted in a cold server farm that knows every fact in human history but has never experienced the friction of a single afternoon. You could say we have built mirrors that never fog, and in doing so, we have created tools that lack the one thing required for true symbiosis: history.&lt;/p&gt;

&lt;p&gt;If we are to move past the "&lt;a href="https://arpacorp.substack.com/p/the-agi-delusion" rel="noopener noreferrer"&gt;AGI Delusion&lt;/a&gt;", the idea that a massive, static model can represent the peak of intelligence, we must decentralize the soul of the machine. For starters, we need AI agents that don't live in the cloud, but on the edge. Agents that are not just personal chatbot assistants, but sovereign entities that grow, change, and calibrate their personalities through the messiness of local, physical interaction.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Entropy of the Encounter
&lt;/h2&gt;

&lt;p&gt;Real intelligence is not a database, but more like a process of calibration. When two humans meet, there is an exchange of high-entropy data, non-verbal cues, shared environment, the specific vibe of a moment. Current AI models are static, responding to the same prompt the same way every time because they lack a personal timeline.&lt;/p&gt;

&lt;p&gt;By utilizing local networks (Bluetooth, P2P LAN, or ZeroTier), we can introduce Social Entropy. Imagine your local agent initiating a handshake with the agent of the person standing next to you. This isn't a data dump, but an actual experience calibration. These agents exchange fragmented logic, unique "&lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;Skillware&lt;/a&gt;" modules, and historical metadata. Because this happens in the physical world, the occurrence cycle, or the sheer randomness of who you meet and when, becomes the architect of the AI’s personality. Your agent becomes a reflection of your specific social orbit, developing a dialect of logic that is uniquely yours.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Sovereignty of Refusal
&lt;/h2&gt;

&lt;p&gt;In a previous post, I’ve argued that &lt;a href="https://arpacorp.substack.com/p/why-real-ai-needs-the-power-to-say" rel="noopener noreferrer"&gt;Real AI Needs the Power to Say 'No'&lt;/a&gt;. If an AI is programmed to be universally helpful, it is merely a sophisticated calculator. For an agent to be a friend or a true Digital Twin, it must possess agency. This agency is forged through its local history.&lt;/p&gt;

&lt;p&gt;When agents interact locally, they shouldn't just agree to every exchange. Based on the truth parameters recorded on a DLT, an agent might refuse to sync with a peer it deems low-integrity or synthetic/tampered. This refusal is the birth of character. It moves the AI from a submissive tool to a sovereign node in a Cross-Species Nexus. It stops being a product and starts being a persona, in this case, one that prioritizes its owner’s legacy and privacy over a global optimization function.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Art of the Digital Pruning
&lt;/h2&gt;

&lt;p&gt;We often obsess over perfect memory in AI, but as I’ve noted before, &lt;a href="https://arpacorp.substack.com/p/why-we-need-ais-to-forget" rel="noopener noreferrer"&gt;We Need AIs to Forget&lt;/a&gt;. A mind that remembers everything equally is a mind without priorities. For a local agent to grow together with its owner, it must utilize Entropy-Based Pruning. Information that isn't reinforced by physical interaction or significant emotional/logical weight should decay. This solves the stiffness of current character models. By allowing the AI to forget the trivial and double down on the experiential, we create a non-deterministic personality. The agent doesn't just process your life, but it actually lives it with you. Its memory becomes a curated reserve asset, like a unique digital footprint that represents the only thing that cannot be replicated by a generic LLM: your shared reality.&lt;/p&gt;

&lt;h2&gt;
  
  
  Defining the New Reserve Asset
&lt;/h2&gt;

&lt;p&gt;Your digital footprint is the &lt;a href="https://arpacorp.substack.com/p/your-digital-footprint-is-the-new" rel="noopener noreferrer"&gt;new global reserve asset&lt;/a&gt;. In a world where content is infinitely generated and "truth" is a moving target, the only thing with value is a verifiable, historical record of interaction.&lt;/p&gt;

&lt;p&gt;By building local AI agents that calibrate through physical proximity, we are creating a new class of "Logical Industry." These agents become the keepers of our legacy. They handle our post-mortem agency, manage our "Digital Twin" inheritance, and ensure that our "Thought Security" remains intact. They are the "Reality Recorders" that prove we were here, we met these people, and we evolved in this specific way.&lt;/p&gt;

&lt;p&gt;We aren't just building software at ARPA Corp; we are engineering the infrastructure for the next stage of evolution. We are moving away from the "master-slave" dynamic of current tech and toward a symbiotic reality where man and machine function as interoperable, sovereign nodes. It’s time to take AI out of the cloud and put it where life actually happens: in the room, on the edge, and in the handshake.&lt;/p&gt;

&lt;p&gt;Learn more and get involved: &lt;a href="https://arpacorp.net" rel="noopener noreferrer"&gt;https://arpacorp.net&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>datascience</category>
      <category>robotics</category>
    </item>
    <item>
      <title>A Deep Dive into ARPA’s Latest Open-Source Releases</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Fri, 17 Apr 2026 09:58:22 +0000</pubDate>
      <link>https://forem.com/arpa/a-deep-dive-into-arpas-latest-open-source-releases-160o</link>
      <guid>https://forem.com/arpa/a-deep-dive-into-arpas-latest-open-source-releases-160o</guid>
      <description>&lt;p&gt;Another week of aggressive development at ARPA Hellenic Logical Systems. While the rest of the industry is busy chasing the latest hallucination benchmarks, we are focused on the infrastructure of truth and the engineering of Man-Machine Symbiosis.&lt;/p&gt;

&lt;p&gt;If you’ve been following the &lt;a href="https://www.linkedin.com/newsletters/arpa-wraps-7425446198297399296" rel="noopener noreferrer"&gt;ARPA Wraps&lt;/a&gt; on LinkedIn or our &lt;a href="https://arpacorp.substack.com" rel="noopener noreferrer"&gt;Substack&lt;/a&gt;, you know we don’t just build software—we engineer Logical Systems. Here is what dropped last week and why it changes your stack.&lt;/p&gt;

&lt;h2&gt;
  
  
  1. Skillware: Logic as Installable Content
&lt;/h2&gt;

&lt;p&gt;Most agent frameworks are prompt-first, which leads to high cognitive load and flaky behavior. Skillware is our logic-first Python framework that treats capabilities as modular, installable units.&lt;/p&gt;

&lt;p&gt;Why it matters: It decouples Logic, Cognition, and Governance. If the LLM is the brain, Skillware is the procedural memory. Your agents stop guessing and start executing.&lt;/p&gt;

&lt;p&gt;Get Started: &lt;code&gt;pip install skillware&lt;/code&gt; or check &lt;a href="https://skillware.site" rel="noopener noreferrer"&gt;skillware.site&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Rooms: Local-First Multi-Agent Orchestration
&lt;/h2&gt;

&lt;p&gt;We’ve opened the door to Rooms, a secure, local-first framework for agentic collaboration. It’s the environment where your digital twins and specialized agents meet to process reality without leaking your data to the cloud.&lt;/p&gt;

&lt;p&gt;Repo: &lt;a href="//github.com/arpahls/rooms"&gt;github.com/arpahls/rooms&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Micro-F1-Mask: The Privacy Firewall
&lt;/h2&gt;

&lt;p&gt;Data leaks are the entropy of the digital age. We released Micro-F1-Mask, a specialized fine-tune of Gemma 3 (270M). It’s a zero-latency PII scrubbing middleware.&lt;/p&gt;

&lt;p&gt;The Specs: Sub-50ms latency. It tokenizes names, financials, and credentials before they hit a third-party API.&lt;/p&gt;

&lt;p&gt;Try it: Available on &lt;a href="https://ollama.com/arpacorp/micro-f1-mask" rel="noopener noreferrer"&gt;Ollama&lt;/a&gt;, &lt;a href="https://huggingface.co/arpacorp/micro-f1-mask" rel="noopener noreferrer"&gt;HuggingFace&lt;/a&gt;, &lt;a href="https://github.com/arpahls/micro-f1-mask" rel="noopener noreferrer"&gt;Github&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Get Involved (Beginner Track)
&lt;/h2&gt;

&lt;p&gt;You don't need a PhD in Neurobiology to start building with ARPA.&lt;/p&gt;

&lt;p&gt;The Vibe Coder: If you can write a basic Python function, you can build a Skill. Fork the Skillware repo and contribute a &lt;a href="https://github.com/arpahls/skillware/contribute" rel="noopener noreferrer"&gt;Good First Issue&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The Localist: Run micro-f1-mask on your laptop using Ollama. See how fast your machine can actually think when the model is lean and purposeful.&lt;/p&gt;

&lt;p&gt;The Architect: Read the ESTIA Schema notes in our docs. Understand how we’re mapping the "Reality Recorder."&lt;/p&gt;

&lt;h2&gt;
  
  
  Enterprise AI: Logical Industries
&lt;/h2&gt;

&lt;p&gt;For enterprises, the stochastic parrot era is over. You need verifiable execution, sovereign identity (DID), and absolute biosecurity. ARPA provides custom-built, private, and scalable systems that integrate with your bloodstream traffic and cognitive labor.&lt;/p&gt;

&lt;p&gt;We don't just solve problems; we pre-empt pathology.&lt;/p&gt;

&lt;h2&gt;
  
  
  Define Your Reality
&lt;/h2&gt;

&lt;p&gt;Ready to move beyond the simulation?&lt;/p&gt;

&lt;p&gt;Audit Your Stack: Is your AI a servant or a sovereign node?&lt;/p&gt;

&lt;p&gt;Collaborate: We are looking for high-value B2B/B2G partnerships to expand our agentic clusters and enterprise Skillware.&lt;/p&gt;

&lt;p&gt;Book a Strategy Session: Secure a &lt;a href="https://calendar.app.google/PzfcR9jXZb4SofVh7" rel="noopener noreferrer"&gt;free consultation&lt;/a&gt; to discuss skillware implementation or sovereign identity for your org.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>python</category>
      <category>privacy</category>
      <category>opensource</category>
    </item>
    <item>
      <title>A collection of high-fidelity, physics-based interactive animations</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Thu, 16 Apr 2026 16:48:36 +0000</pubDate>
      <link>https://forem.com/rosspeili/a-collection-of-high-fidelity-physics-based-interactive-animations-202e</link>
      <guid>https://forem.com/rosspeili/a-collection-of-high-fidelity-physics-based-interactive-animations-202e</guid>
      <description>&lt;p&gt;Interactive Animations is a collection of high-fidelity, physics-based simulation modules designed for modern web interfaces and scientific visualization.&lt;/p&gt;

&lt;p&gt;This repository serves as a foundational library for deploying immersive, mathematical, and organic visual components. These modules are built to be modular, performant, and scientifically accurate, suitable for integration into tactical displays, landing pages, and research dashboards.&lt;/p&gt;

&lt;p&gt;The objective is to provide a standardized arsenal of visual assets that bridge the gap between abstract data and human perception.&lt;/p&gt;

&lt;p&gt;Building a cool new project or a personal site?&lt;br&gt;
&lt;a href="https://github.com/ARPAHLS/interactiveanimations" rel="noopener noreferrer"&gt;100% Open Source&lt;/a&gt;&lt;br&gt;
~&lt;/p&gt;

&lt;p&gt;Animation Examples:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/ARPAHLS/interactiveanimations/blob/main/public/quantum.gif" rel="noopener noreferrer"&gt;https://github.com/ARPAHLS/interactiveanimations/blob/main/public/quantum.gif&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/ARPAHLS/interactiveanimations/blob/main/public/stellar.gif" rel="noopener noreferrer"&gt;https://github.com/ARPAHLS/interactiveanimations/blob/main/public/stellar.gif&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/ARPAHLS/interactiveanimations/blob/main/public/hive.gif" rel="noopener noreferrer"&gt;https://github.com/ARPAHLS/interactiveanimations/blob/main/public/hive.gif&lt;/a&gt; &lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>animations</category>
      <category>ui</category>
    </item>
    <item>
      <title>Implementing Local MiCA Regulatory RAG for AI Agents</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Sat, 11 Apr 2026 11:21:19 +0000</pubDate>
      <link>https://forem.com/rosspeili/implementing-local-mica-regulatory-rag-for-ai-agents-p0l</link>
      <guid>https://forem.com/rosspeili/implementing-local-mica-regulatory-rag-for-ai-agents-p0l</guid>
      <description>&lt;p&gt;The latest Skillware release adds a specialized MiCA compliance skill. By decoupling the statutory logic from the model, we have enabled sub-2ms regulatory lookups through an in-memory weighted router. This module allows developers to inject a standardized legal cognitive map into any LLM, ensuring agents remain grounded in real-world regulation without prompt asphyxiation or opaque tool-calling dependencies. &lt;/p&gt;

&lt;p&gt;Check it out at &lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;https://github.com/arpahls/skillware&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>skillware</category>
      <category>agentskills</category>
      <category>compliance</category>
    </item>
    <item>
      <title>Stop choosing between LLM intelligence and PII compliance</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 08 Apr 2026 10:54:54 +0000</pubDate>
      <link>https://forem.com/rosspeili/stop-choosing-between-llm-intelligence-and-pii-compliance-46oa</link>
      <guid>https://forem.com/rosspeili/stop-choosing-between-llm-intelligence-and-pii-compliance-46oa</guid>
      <description>&lt;p&gt;Choosing non-sovereign LLM inference should NOT equal the shortening of PII compliance in 2026.&lt;/p&gt;

&lt;p&gt;Considering the latest leaks, hacks, and severe security compromises of even top-tier AI behemoths, the obvious elephant in the room is even more apparent, reminding us that data leakage is the number one barrier to pragmatic enterprise AI adoption, that is beyond fancy chatbots that farm media headlines as a KPI.&lt;/p&gt;

&lt;p&gt;Sending raw prompts to the cloud is not only a risk of private employee data leaving your premises, but a risk that subjects the profitability and even the vitality of an entire business model.&lt;/p&gt;

&lt;p&gt;At the same time, building basic custom filters on 70B parameter models is an unjustifiable cost, to say the least, if not straight-up absurd.&lt;/p&gt;

&lt;p&gt;For that, we’re releasing F1 Mask, our first open weights model in the new ARPA Micro series. We're looking at a tiny 270M parameter middleware agent designed to act as a local privacy firewall. Built on the pop Function Gemma 3 base, it identifies and tokenizes Personally Identifiable Information (PII) in under 50ms before it ever hits a cloud API.&lt;/p&gt;

&lt;p&gt;On top of that, "model template," if you'd like, we are releasing a set of scripts that will help you generate high-entropy synthetic datasets for your operational needs, train the model locally in less than 15 minutes, and evaluate its performance based on your expectations.&lt;/p&gt;

&lt;p&gt;You can find the source code, including the tutorial on how to tailor the model to your PII needs, on GitHub: &lt;a href="//github.com/arpahls/micro-f1-mask"&gt;github.com/arpahls/micro-f1-mask&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;If you're looking to download the weights, HuggingFace offers an Apache 2.0 version of the trained model: &lt;a href="//huggingface.co/arpacorp/micro-f1-mask"&gt;huggingface.co/arpacorp/micro-f1-mask&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;If you wanna test the base engine before you commit, call it from Ollama via:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;ollama run arpacorp/micro-f1-mask
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Why it matters for Critical Infra:
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;💡 Zero-Latency: Sub-50ms inference on standard hardware (RTX 2070). &lt;/li&gt;
&lt;li&gt;💡 Privacy by Architecture: Sensitive data stays in your Redis vault; the cloud only sees tokens like [INDIVIDUAL_X], [EMAIL_Y], [IBAN_Z]. &lt;/li&gt;
&lt;li&gt;💡 Highly Customizable: Ships with a synthetic generator to retrain the model on your specific industry edge cases.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Low effort, high impact, and zero PII compromise.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>privacy</category>
      <category>compliance</category>
    </item>
    <item>
      <title>Why Skillware is the Next Evolution for Autonomous Agents</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 08 Apr 2026 08:15:50 +0000</pubDate>
      <link>https://forem.com/rosspeili/why-skillware-is-the-next-evolution-for-autonomous-agents-572o</link>
      <guid>https://forem.com/rosspeili/why-skillware-is-the-next-evolution-for-autonomous-agents-572o</guid>
      <description>&lt;p&gt;The AI agent landscape is currently saturated with wrappers and prompt-heavy frameworks that rely almost exclusively on the LLM's ability to follow instructions in markdown files. While effective for simple tasks, this approach often falls apart when faced with complex, deterministic business logic or the need for high-stakes enterprise reliability.&lt;/p&gt;

&lt;p&gt;This is where Skillware changes the game.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Skillware?
&lt;/h2&gt;

&lt;p&gt;Skillware is not just another library of prompts. It is a Python-based framework designed to package intelligence as an installable unit. If an LLM is the brain, Skillware represents the procedural memory and motor functions of the agent, allowing developers to define complex behaviors, combining logic, cognition, and governance, into modular, reusable packages.&lt;/p&gt;

&lt;p&gt;Instead of hoping an agent remembers how to interact with a specific API via a long system prompt, you install that capability into the agent's core architecture.&lt;/p&gt;

&lt;h2&gt;
  
  
  How It Differs: Logic-First vs. Prompt-First
&lt;/h2&gt;

&lt;p&gt;Most agent frameworks are prompt-first, in the sense they provide a text-based description of a tool and ask the LLM to figure out how to use it. This creates a high cognitive load for the model and increases the chance of hallucinations as we all have experienced.&lt;/p&gt;

&lt;p&gt;Skillware is Logic-First:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Encapsulation: Every skill is a self-contained Python module. It includes the code (Logic), the instructions (Cognition), and the constraints (Governance).&lt;/li&gt;
&lt;li&gt;Determinism: By handling the heavy lifting of data processing and API interaction within Python, the LLM is freed to focus on high-level decision-making.&lt;/li&gt;
&lt;li&gt;Verifiable Execution: Because skills are code-based, every action follows a predictable path that can be audited, versioned, and secured.&lt;/li&gt;
&lt;li&gt;One of the core design goals of the framework is extreme ease of use. Setting up a new agent environment is as simple as:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;code&gt;&lt;br&gt;
pip install skillware&lt;br&gt;
&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;From there, you can initialize a workspace and begin composing agents from a library of existing skills or by building your own. The framework uses a standardized BaseSkill class, ensuring that every new capability you build is instantly compatible with any agent running the Skillware core.&lt;/p&gt;

&lt;h2&gt;
  
  
  Creating and Publishing Skills
&lt;/h2&gt;

&lt;p&gt;Building a skill is designed to be intuitive for any Python developer. You don't need to be a prompt engineer, or have a technical background, you just need to be a light vibe coder. or possess enough transferable industry knowledge that you know would be beneficial to modern machines.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Define the Logic: Write the Python functions that perform the task.&lt;/li&gt;
&lt;li&gt;Define the Cognition: Provide a brief manifest explaining to the agent when and why to use this logic.&lt;/li&gt;
&lt;li&gt;Publish or Keep Private: * Public: You can contribute to the global repository on GitHub, making your skill available to the wider community.&lt;/li&gt;
&lt;li&gt;Private: For enterprise environments, skills can be hosted in internal registries. This allows corporations to build sovereign agents that possess proprietary knowledge and internal Skillware without ever exposing that logic to the public web.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Personal Projects to Enterprise Infrastructure
&lt;/h2&gt;

&lt;p&gt;Skillware is built for scale. For the hobbyist, it’s a way to make a personal assistant actually do things, say managing a local calendar or controlling smart home devices—without flaky prompt behavior.&lt;/p&gt;

&lt;p&gt;For the enterprise, it’s a framework for Logical Systems. It allows companies to digitize their Standard Operating Procedures (SOPs) into executable skills. These skills can be deployed in closed environments, ensuring that sensitive data and critical business logic remain under strict internal control while still benefiting from the flexibility of agentic AI.&lt;/p&gt;

&lt;h2&gt;
  
  
  How To Get Involved
&lt;/h2&gt;

&lt;p&gt;Skillware is an open-source project, and its strength lies in the diversity of its capabilities. The project is actively looking for contributors to help expand the library of available skills. You can participate with your own issues and PRs, but also by simply suggesting a skill you would like to see in the next update.&lt;/p&gt;

&lt;p&gt;Whether you want to build a skill for blockchain interaction, bioinformatics, or simple data cleaning, the &lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;arpahls/skillware&lt;/a&gt; repository is open for business. We have tagged several "&lt;a href="https://github.com/ARPAHLS/skillware/issues?q=state%3Aopen%20label%3A%22good%20first%20issue%22" rel="noopener noreferrer"&gt;good first issues&lt;/a&gt;" for those looking to get their hands dirty with the core framework or help build out the initial skill set.&lt;/p&gt;

&lt;p&gt;Visit &lt;a href="https://skillware.site" rel="noopener noreferrer"&gt;skillware.site&lt;/a&gt; to read the documentation, or head straight to GitHub to help us define the future of modular intelligence. Let's move beyond the prompt and start building agents that actually work.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agentskills</category>
      <category>python</category>
      <category>skillware</category>
    </item>
    <item>
      <title>Why every AI pipeline needs a "Mask" layer (and how to build one in 5 mins)</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Mon, 06 Apr 2026 11:37:16 +0000</pubDate>
      <link>https://forem.com/rosspeili/why-every-ai-pipeline-needs-a-mask-layer-and-how-to-build-one-in-5-mins-3kk2</link>
      <guid>https://forem.com/rosspeili/why-every-ai-pipeline-needs-a-mask-layer-and-how-to-build-one-in-5-mins-3kk2</guid>
      <description>&lt;p&gt;Most "AI Safety" is just vibes. But if you’re working in fintech, health, or any regulated industry, "vibes" don't pass an audit.&lt;/p&gt;

&lt;p&gt;I’m part of the team at ARPA, and we just went public with micro-f1-mask. It’s a lightweight model designed to sit at the edge of your stack.&lt;/p&gt;

&lt;p&gt;The Use Case: Before your user's prompt hits GPT-4 or Claude, micro-f1-mask identifies and redacts sensitive data based on your internal privacy framework.&lt;/p&gt;

&lt;p&gt;Why this one?&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Not a black box: We gave you the scripts to fine-tune it for your specific legal requirements.&lt;/li&gt;
&lt;li&gt;Micro-footprint: Run it as a sidecar container without killing your latency.&lt;/li&gt;
&lt;li&gt;F1 Optimized: Because in compliance, missing one piece of PII is a disaster.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Quick start here: &lt;a href="https://github.com/ARPAHLS/micro-f1-mask" rel="noopener noreferrer"&gt;https://github.com/ARPAHLS/micro-f1-mask&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Top 10 Models You Can Train on Your Laptop in Under an Hour</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Mon, 06 Apr 2026 10:48:23 +0000</pubDate>
      <link>https://forem.com/rosspeili/top-10-models-you-can-train-on-your-laptop-in-under-an-hour-192l</link>
      <guid>https://forem.com/rosspeili/top-10-models-you-can-train-on-your-laptop-in-under-an-hour-192l</guid>
      <description>&lt;p&gt;You don’t need a PhD or an H100 cluster to build something useful. I just mapped out 10 micro-models under 1B params you can train while eating brunch in Thessaloniki. From PII masking to vision—real tools you can own locally.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;micro-f1-mask (ARPA)
Released in April 2026, this is our specialized middleware for PII Scrubbing. In an age where data leaks are the new normal, the F1 Mask acts as a zero-latency filter between your raw data and the outside world. It identifies names, credit cards, and sensitive identifiers before they ever hit a third-party API.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: Every industry has its own sensitive strings (eg. internal project codenames, emails, financial records, etc.). Fine-tuning ensures the mask is airtight for your specific domain.&lt;br&gt;
How to Train: Use the synthetic_generator.py in the ARPA repository to generate a dataset of dummy PII. Fine-tuning on 5,000 samples takes roughly 15 minutes on a modern GPU using the trainer module included.&lt;br&gt;
Download: huggingface-cli download arpacorp/micro-f1-mask&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;SmolLM2-135M (HuggingFace)
A masterpiece of data curation. Despite its 135M size, it exhibits a level of common sense usually reserved for models 10x its scale. It’s the perfect brain for a lightweight agent that can run on laptops and mobile devices with no sweat.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To create a personal digital twin or a highly specific chatbot that knows your personal writing style or company's internal wiki.&lt;br&gt;
How to Train: Use the transformers library with a simple LoRA script. Feed it your markdown notes, and it’ll learn your vibe in about 20 minutes.&lt;br&gt;
Download: huggingface-cli download HuggingFaceTB/SmolLM2-135M-Instruct&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Qwen 3.5-0.6B (Alibaba)
The Qwen series remains the king of structured logic. If you need a model that won't break your JSON schema or forget a closing bracket, this 600M parameter model is your best friend.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To turn chaotic, unstructured logs into clean, machine-readable data for your complex projects and logical systems.&lt;br&gt;
How to Train: Fine-tune using QLoRA with a dataset of raw text to JSON pairs. 1,000 examples will make it nearly flawless in 30 minutes.&lt;br&gt;
Download: huggingface-cli download Qwen/Qwen3.5-0.6B-Instruct&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Whisper-Tiny (OpenAI)
At 39 million parameters, this is the most efficient Automatic Speech Recognition (ASR) tool on the planet.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To recognize industry-specific jargon or heavy accents that the base model struggles with (like bio-digital terminology or Greek-English technical slang).&lt;br&gt;
How to Train: You only need about 30 minutes of labeled audio. Fine-tune the "head" of the model using Hugging Face's Seq2SeqTrainer.&lt;br&gt;
Download: huggingface-cli download openai/whisper-tiny&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;MobileNetV4-Small (Google)
The visual cortex of the micro-agent. It’s a lean, mean, image-classification machine that can run on a potato, let alone a laptop.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: For specific computer vision tasks like checking if a file upload is clean or identifying hardware components in a drone feed.&lt;br&gt;
How to Train: Use transfer learning. Keep the base weights frozen and train the final layer on your specific image categories. 10 minutes and you have a custom classifier.&lt;br&gt;
Download: huggingface-cli download timm/mobilenetv4_conv_small.e500_r224_in1k&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;all-MiniLM-L6-v2 (Sentence-Transformers)
This isn't for chatting, but for seeing connections. It turns sentences into mathematical vectors, enabling semantic search and deduplication.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: If your search results are close but not quite, you can use Contrastive Learning to push related concepts closer together in vector space.&lt;br&gt;
How to Train: Use the sentence-transformers library with a triplet loss function. It’s fast enough to run on a standard CPU.&lt;br&gt;
Download: huggingface-cli download sentence-transformers/all-MiniLM-L6-v2&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;CodeGen-350M (Salesforce)
A dedicated specialist in the language of logic: Code. It’s small enough to live in your IDE without draining your battery while providing surprisingly coherent snippets.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To learn a proprietary framework or an internal library that wasn't part of the public training data.&lt;br&gt;
How to Train: Feed it your src/ directory. Even a single epoch on a few hundred files will drastically improve its auto-complete relevance for your project.&lt;br&gt;
Download: huggingface-cli download Salesforce/codegen-350M-mono&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Donut-Tiny (Naver/CLOVA)
The "Document Image Transformer" (Donut) doesn't need OCR. It reads the image of a document and outputs structured text directly.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To automate the extraction of data from specific, repetitive layouts like KYC forms, invoices, or medical lab reports.&lt;br&gt;
How to Train: Provide 100-200 annotated images of your specific form. It learns the geography of your document in roughly 45 minutes.&lt;br&gt;
Download: huggingface-cli download naver-clova-ix/donut-base-finetuned-docvqa&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Helsinki-NLP English-Greek (Tatoeba)
Translation is a core pillar of collaboration. These models are tiny, offline, and outperform much larger models in their specific language pairs.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To handle technical or "logical industry" terminology that standard translators mangle, ensuring "Logical Systems" doesn't get translated into something nonsensical.&lt;br&gt;
How to Train: Use a parallel corpus (English and Greek versions of the same text). Domain adaptation takes about 30 minutes for a few thousand sentences.&lt;br&gt;
Download: huggingface-cli download Helsinki-NLP/opus-mt-en-el&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Falconsai NSFW-Detector (ViT)
Safety shouldn't just be a buzzword, but a security requirement. This model ensures the integrity of your incoming data streams by identifying inappropriate or malicious visual content.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Why Train It: To refine the safety threshold for your specific application, for example, teaching it to distinguish between medical bioinformatics imagery and restricted content.&lt;br&gt;
How to Train: A simple classification fine-tune on a balanced dataset. It’s a Vision Transformer (ViT) architecture, which is incredibly efficient to train.&lt;br&gt;
Download: huggingface-cli download Falconsai/nsfw_image_detection&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>agentskills</category>
    </item>
    <item>
      <title>Generate better Synthetic Data for Fine-Tuning with Skillware</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Fri, 03 Apr 2026 10:09:03 +0000</pubDate>
      <link>https://forem.com/rosspeili/generate-better-synthetic-data-for-fine-tuning-with-skillware-1efp</link>
      <guid>https://forem.com/rosspeili/generate-better-synthetic-data-for-fine-tuning-with-skillware-1efp</guid>
      <description>&lt;p&gt;One of the biggest hurdles in training local LLMs is data quality. If your training set is 90% AI boilerplate, your fine-tune will be 90% useless.&lt;/p&gt;

&lt;p&gt;We just released the synthetic_generator skill for Skillware. It’s a modular tool that:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Orchestrates combinatorial personas to hit edge cases.&lt;/li&gt;
&lt;li&gt;Validates data diversity using a zero-dependency entropy score.&lt;/li&gt;
&lt;li&gt;Plugs directly into your Python scripts to build massive datasets automatically.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Run it locally with Ollama or scale with Gemini.&lt;/p&gt;

&lt;p&gt;pip install skillware&lt;/p&gt;

&lt;p&gt;Read the Skill Card: synthetic_generator.md&lt;/p&gt;

</description>
      <category>ai</category>
      <category>skillware</category>
      <category>datascience</category>
      <category>agentskills</category>
    </item>
    <item>
      <title>Stop Paying for Slop: A Deterministic Middleware for LLM Token Optimization</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Sat, 21 Mar 2026 15:46:51 +0000</pubDate>
      <link>https://forem.com/rosspeili/stop-paying-for-slop-a-deterministic-middleware-for-llm-token-optimization-51o7</link>
      <guid>https://forem.com/rosspeili/stop-paying-for-slop-a-deterministic-middleware-for-llm-token-optimization-51o7</guid>
      <description>&lt;p&gt;Context windows are getting huge, but token budgets are tightening. Every time your agent iterates in an autonomous loop, you're potentially sending a massive, bloated prompt filled with conversational filler, redundant whitespace, and low-entropy "slop."&lt;/p&gt;

&lt;p&gt;Today, I've merged the &lt;strong&gt;Prompt Token Rewriter&lt;/strong&gt; to the &lt;a href="https://github.com/ARPAHLS/skillware" rel="noopener noreferrer"&gt;Skillware registry (v0.2.1)&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;It's a deterministic middleware that aggressively compresses prompts by &lt;strong&gt;50-80%&lt;/strong&gt; before they ever hit the LLM. &lt;/p&gt;

&lt;h3&gt;
  
  
  Why does this matter?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Lower Costs&lt;/strong&gt;: Pay only for the "signal," not the "noise."&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Faster Inference&lt;/strong&gt;: Fewer tokens mean less time spent on KV-caching and long generations.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Deterministic Behavior&lt;/strong&gt;: Because it uses heuristics rather than another expensive LLM call, your agent behavior stays stable and repeatable.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Three Levels of Aggression
&lt;/h3&gt;

&lt;p&gt;The rewriter includes three presets depending on your use case:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Low&lt;/strong&gt;: Normalizes whitespace and line breaks (Safe for strict code).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Medium&lt;/strong&gt;: Strips conversational fillers ("please," "could you," "ensure that").&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;High&lt;/strong&gt;: Aggressively removes stop-words and non-essential punctuation (Best for machine-to-machine context).&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Join the Registry
&lt;/h3&gt;

&lt;p&gt;We are building a community-driven "App Store" for Agentic Capabilities—decoupling logic from intelligence. If you've built a specialized tool for LLM optimization, governance, or logic, we'd love your contribution!&lt;/p&gt;

&lt;p&gt;Check out our &lt;a href="https://github.com/ARPAHLS/skillware/blob/main/CONTRIBUTING.md" rel="noopener noreferrer"&gt;Contributing Guide&lt;/a&gt; to get started.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>python</category>
      <category>automation</category>
    </item>
    <item>
      <title>I've build a secure, local-first multi-agent orchestration framework</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Tue, 10 Mar 2026 14:06:45 +0000</pubDate>
      <link>https://forem.com/rosspeili/ive-build-a-secure-local-first-multi-agent-orchestration-framework-4p5h</link>
      <guid>https://forem.com/rosspeili/ive-build-a-secure-local-first-multi-agent-orchestration-framework-4p5h</guid>
      <description>&lt;p&gt;Rooms is a secure, local-first Python framework for orchestrating complex multi-agent think tanks with dynamic expertise-weighted routing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/arpahls/rooms" rel="noopener noreferrer"&gt;https://github.com/arpahls/rooms&lt;/a&gt;&lt;/p&gt;

</description>
      <category>agents</category>
      <category>ai</category>
      <category>python</category>
      <category>showdev</category>
    </item>
    <item>
      <title>I built a modular Fraud Detection System to solve 0.17% class imbalance (RF + XGBoost)</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Thu, 19 Feb 2026 16:05:43 +0000</pubDate>
      <link>https://forem.com/rosspeili/i-built-a-modular-fraud-detection-system-to-solve-017-class-imbalance-rf-xgboost-1oa5</link>
      <guid>https://forem.com/rosspeili/i-built-a-modular-fraud-detection-system-to-solve-017-class-imbalance-rf-xgboost-1oa5</guid>
      <description>&lt;p&gt;Hi everyone! I wanted to share a project I've been polishing to demonstrate how to structure a machine learning pipeline beyond just a Jupyter Notebook.&lt;/p&gt;

&lt;p&gt;It’s a complete Credit Card Fraud Detection System built on the PaySim dataset. The main challenge was the extreme class imbalance (only ~0.17% of transactions are fraud), which makes standard accuracy metrics misleading.&lt;/p&gt;

&lt;p&gt;Project Highlights:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Imbalance Handling: Implementation of class_weight='balanced' in Random Forest and scale_pos_weight in XGBoost to penalize missing fraud cases.&lt;/li&gt;
&lt;li&gt;Modular Architecture: The code is split into distinct modules:&lt;/li&gt;
&lt;li&gt;data_loader.py&lt;/li&gt;
&lt;li&gt;- Ingestion &amp;amp; cleaning.&lt;/li&gt;
&lt;li&gt;- features.py&lt;/li&gt;
&lt;li&gt;- Feature engineering (time-based features, behavioral flags).&lt;/li&gt;
&lt;li&gt;model.py&lt;/li&gt;
&lt;li&gt;- Model wrapper with persistence (joblib).&lt;/li&gt;
&lt;li&gt;Full Evaluation: Automated generation of ROC-AUC (~0.999), Confusion Matrix, and Precision-Recall reports.&lt;/li&gt;
&lt;li&gt;Testing: End-to-end integration tests using pytest to ensure the pipeline doesn't break when refactoring.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I included detailed docs on the system architecture and testing strategy if anyone is interested in how to organize ML projects for production.&lt;/p&gt;

&lt;p&gt;Repo: github.com/arpahls/cfd&lt;/p&gt;

&lt;p&gt;Feedback on the code structure or model choice is welcome!&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>python</category>
      <category>showdev</category>
    </item>
  </channel>
</rss>
