<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Ross Peili</title>
    <description>The latest articles on Forem by Ross Peili (@rosspeili).</description>
    <link>https://forem.com/rosspeili</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F308648%2Fe3062c7b-4d7d-4ba9-aea5-f0ea4fef4776.jpg</url>
      <title>Forem: Ross Peili</title>
      <link>https://forem.com/rosspeili</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/rosspeili"/>
    <language>en</language>
    <item>
      <title>Pre-commit asset handoffs, on-chain execution, zero lawyers, help build it early</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Sun, 10 May 2026 13:40:22 +0000</pubDate>
      <link>https://forem.com/arpa/pre-commit-asset-handoffs-on-chain-execution-zero-lawyers-help-build-it-early-247p</link>
      <guid>https://forem.com/arpa/pre-commit-asset-handoffs-on-chain-execution-zero-lawyers-help-build-it-early-247p</guid>
      <description>&lt;p&gt;ARPA Legacy Protocol is an open-source, work-in-progress framework for programmable asset handoffs on Ethereum. &lt;/p&gt;

&lt;p&gt;The idea: define beneficiaries, assets, and conditions (timers, attestations, oracle data) ahead of time. When conditions hold, the chain executes—no intermediaries, no renegotiation. Still pre-contract: reference specs, ADRs, a vault model, and policy schemas live on GitHub. Solidity implementation (Foundry) is next on the roadmap. &lt;/p&gt;

&lt;p&gt;Looking for contributors who care about on-chain inheritance, cryptoeconomic primitives, or policy design. MIT licensed. Come shape it early.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/arpahls/legacy-protocol" rel="noopener noreferrer"&gt;https://github.com/arpahls/legacy-protocol&lt;/a&gt;&lt;/p&gt;

</description>
      <category>opensource</category>
      <category>blockchain</category>
      <category>solidity</category>
      <category>ethereum</category>
    </item>
    <item>
      <title>A VIC x AiSAQ Implementation Brings AI to Your Files Without Breaking the Bank</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Sat, 09 May 2026 11:18:32 +0000</pubDate>
      <link>https://forem.com/arpa/a-vic-x-aisaq-implementation-brings-ai-to-your-files-without-breaking-the-bank-1mic</link>
      <guid>https://forem.com/arpa/a-vic-x-aisaq-implementation-brings-ai-to-your-files-without-breaking-the-bank-1mic</guid>
      <description>&lt;p&gt;We’re generating more data than ever, and AI‑powered search is great—until your dataset gets huge and your RAM starts crying for mercy. Most vector search systems rely on expensive DRAM to keep indexes fast, but that approach doesn’t scale. &lt;a href="https://github.com/kioxia-jp/aisaq-diskann" rel="noopener noreferrer"&gt;KIOXIA’s &lt;strong&gt;AiSAQ&lt;/strong&gt;&lt;/a&gt; (All‑in‑Storage ANNS with Product Quantization) flips the script: it runs approximate nearest neighbor search directly on SSD, slashing DRAM usage by &lt;strong&gt;3,200×&lt;/strong&gt; in billion‑scale workloads. The &lt;a href="https://github.com/ARPAHLS/vic_aisaq_demo" rel="noopener noreferrer"&gt;&lt;code&gt;vic_aisaq_demo&lt;/code&gt;&lt;/a&gt; repo from &lt;strong&gt;ARPA Hellenic Logical Systems&lt;/strong&gt; puts this tech into a practical, local‑first retrieval pipeline that’s as auditable as it is efficient.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;TL;DR&lt;/strong&gt;: &lt;code&gt;vic_aisaq_demo&lt;/code&gt; combines tiered metadata filtering with flash‑optimized vector search to keep memory low and answers relevant. It’s a live demo of storage‑aware AI for edge and controller‑style environments.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  The Problem: DRAM Is the Bottleneck
&lt;/h2&gt;

&lt;p&gt;Graph‑based nearest neighbor search (like HNSW) is fast, but it keeps key index structures in DRAM. With billion‑scale datasets, memory costs explode. Even compressed representations can still require tens of gigabytes of RAM. &lt;a href="https://github.com/kioxia-jp/aisaq-diskann" rel="noopener noreferrer"&gt;KIOXIA’s AiSAQ technology&lt;/a&gt; changes that by moving those compressed vectors to flash storage, consuming as little as &lt;strong&gt;10 MB&lt;/strong&gt; of DRAM during search without sacrificing recall.&lt;/p&gt;

&lt;p&gt;But low DRAM is only half the story. You also need a retrieval strategy that doesn’t waste time parsing irrelevant files.&lt;/p&gt;

&lt;h2&gt;
  
  
  How &lt;code&gt;vic_aisaq_demo&lt;/code&gt; Works: Tiered Retrieval Meets Flash‑Native Search
&lt;/h2&gt;

&lt;p&gt;The demo builds on two open‑source building blocks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/ARPAHLS/lc0_vic" rel="noopener noreferrer"&gt;&lt;code&gt;lc0_vic&lt;/code&gt;&lt;/a&gt;&lt;/strong&gt; – a tiered retrieval controller that plans and orchestrates search in layers (L0 → L1 → L2).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://github.com/kioxia-jp/aisaq-diskann" rel="noopener noreferrer"&gt;&lt;code&gt;aisaq-diskann&lt;/code&gt;&lt;/a&gt;&lt;/strong&gt; – a flash‑oriented ANN backend optimized for low‑DRAM environments.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The execution flow is refreshingly simple:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Librarian / Plan&lt;/strong&gt; – Turn a natural‑language question into retrieval intent using a lightweight LLM (e.g., &lt;a href="https://ollama.com/library/qwen2.5" rel="noopener noreferrer"&gt;qwen2.5:0.5b&lt;/a&gt; via Ollama).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;L0 Metadata Filter&lt;/strong&gt; – Narrow down candidate files by extension, size, time, or path hints. Cheap and fast.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;L1 Vector Search&lt;/strong&gt; – Run native AiSAQ ANN search over embeddings to find semantically similar content.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;L2 Deep Read&lt;/strong&gt; – Parse only the top few files and extract evidence snippets.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ranked Response&lt;/strong&gt; – Return paths, scores, and run metrics.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The tiered approach keeps deep parsing affordable at scale.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Benchmark results&lt;/strong&gt; show latency remains stable as dataset size grows, while DRAM footprint stays near zero. The funnel chart below visualises how each tier slashes the candidate pool:&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv08c072sk9m1a1rq3t7y.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv08c072sk9m1a1rq3t7y.png" alt="Tier Funnel" width="800" height="435"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And here’s how the pipeline shifts results from superficial matching to true semantic evidence:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7i386j3iivqd3c0ndfpt.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7i386j3iivqd3c0ndfpt.png" alt="Match Type Comparison" width="800" height="433"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Try It Yourself
&lt;/h2&gt;

&lt;p&gt;The repo is built to be &lt;strong&gt;reproducible and local‑first&lt;/strong&gt;. You’ll need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;WSL (Ubuntu)&lt;/strong&gt; for building AiSAQ binaries&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ollama&lt;/strong&gt; running locally (or over the network) with two models:

&lt;ul&gt;
&lt;li&gt;Planner model: &lt;a href="https://ollama.com/library/qwen2.5" rel="noopener noreferrer"&gt;&lt;code&gt;qwen2.5:0.5b&lt;/code&gt;&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Embedding model: &lt;a href="https://ollama.com/library/embeddinggemma" rel="noopener noreferrer"&gt;&lt;code&gt;embeddinggemma&lt;/code&gt;&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Python 3.13 and the usual suspects (see &lt;code&gt;requirements.txt&lt;/code&gt;)&lt;/li&gt;

&lt;/ul&gt;

&lt;p&gt;Once you’ve built the AiSAQ index from a sample drive, a query like:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;python3 scripts/run_query.py &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="s2"&gt;"Find the Q3 2025 contract that mentions penalty clauses"&lt;/span&gt; &lt;span class="se"&gt;\&lt;/span&gt;
  &lt;span class="nt"&gt;--aisaq-root&lt;/span&gt; /home/&lt;span class="nv"&gt;$USER&lt;/span&gt;/aisaq-diskann
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;…will return ranked files with evidence snippets, tier labels, and latency metrics.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why This Matters
&lt;/h2&gt;

&lt;p&gt;&lt;code&gt;vic_aisaq_demo&lt;/code&gt; isn’t just a toy. It demonstrates a realistic, storage‑aware retrieval pattern that could run on devices with tight memory budgets—think edge gateways, embedded controllers, or even future SSD firmware that embeds intelligence directly on the drive. The &lt;a href="https://github.com/rosspeili/computational_storage_landscape" rel="noopener noreferrer"&gt;Computational Storage Landscape report&lt;/a&gt; maps this evolution, and this repo is one of the first runnable examples that puts those ideas into practice.&lt;/p&gt;

&lt;p&gt;The two charts below summarise that systems trade‑off and scaling behaviour:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft2stpp8klpocea6q8vzd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft2stpp8klpocea6q8vzd.png" alt="Latency vs Dataset Size" width="800" height="432"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F43swbbmrzcyhduowwl8n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F43swbbmrzcyhduowwl8n.png" alt="DRAM Footprint by Method" width="800" height="394"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The takeaway? You don’t need a cluster of DRAM‑heavy servers to run effective semantic search. Sometimes the smartest storage is the one that knows what &lt;em&gt;not&lt;/em&gt; to load into memory.&lt;/p&gt;

&lt;p&gt;Check out the full repo: &lt;strong&gt;&lt;a href="https://github.com/ARPAHLS/vic_aisaq_demo" rel="noopener noreferrer"&gt;ARPAHLS/vic_aisaq_demo&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>computationalstorage</category>
      <category>vectorsearch</category>
      <category>edgeai</category>
      <category>lowmemoryretrieval</category>
    </item>
    <item>
      <title>Open Source Emotion‑Aware Access Control with Face Verification</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Fri, 08 May 2026 10:11:27 +0000</pubDate>
      <link>https://forem.com/arpa/open-source-emotion-aware-access-control-with-face-verification-14d6</link>
      <guid>https://forem.com/arpa/open-source-emotion-aware-access-control-with-face-verification-14d6</guid>
      <description>&lt;p&gt;&lt;strong&gt;Gatekeeper: Emotion‑Aware Access Control with Face Verification&lt;/strong&gt;  &lt;/p&gt;

&lt;p&gt;What if your system could deny access not just based on &lt;em&gt;who&lt;/em&gt; you are, but on &lt;em&gt;how&lt;/em&gt; you feel?  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/arpahls/gatekeeper" rel="noopener noreferrer"&gt;&lt;strong&gt;Gatekeeper&lt;/strong&gt;&lt;/a&gt; is a Python‑based security framework that layers real‑time &lt;strong&gt;face verification&lt;/strong&gt; with &lt;strong&gt;emotion analysis&lt;/strong&gt; before granting access.  &lt;/p&gt;

&lt;h3&gt;
  
  
  How it works
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Verify identity&lt;/strong&gt; against a reference image or an admin pool.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Analyze emotions&lt;/strong&gt; (anger, fear, joy, etc.) and evaluate them against a configurable policy (blocked states, thresholds, weights).
&lt;/li&gt;
&lt;li&gt;Only grant access if &lt;em&gt;both&lt;/em&gt; checks pass.
&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Why it matters
&lt;/h3&gt;

&lt;p&gt;Critical operations (financial systems, secure rooms, privileged commands) deserve more than binary yes/no. By assessing emotional state, you reduce the risk of coercion, panic, or compromised decision‑making in high‑impact environments.  &lt;/p&gt;

&lt;h3&gt;
  
  
  Get started
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
bash
git clone https://github.com/arpahls/gatekeeper
cd gatekeeper
python -m venv .venv
.venv\Scripts\activate  # or 'source .venv/bin/activate' on Linux/macOS
pip install -r requirements.txt
python scripts/run_terminal.py
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

</description>
      <category>ai</category>
      <category>security</category>
      <category>deepface</category>
      <category>kyc</category>
    </item>
    <item>
      <title>Where to find Open Source AI Projects with Good First Issues</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Fri, 08 May 2026 07:44:14 +0000</pubDate>
      <link>https://forem.com/rosspeili/where-to-find-open-source-ai-projects-with-good-first-issues-5d1l</link>
      <guid>https://forem.com/rosspeili/where-to-find-open-source-ai-projects-with-good-first-issues-5d1l</guid>
      <description>&lt;p&gt;Contributing to open‑source AI as a beginner isn’t about landing a seat at a super‑intelligence lab after a PhD, but more like finding good first issues on projects that are small enough to wrap your head around, real enough to have an impact, and early enough so you make it big, bypassing the credentials-based theoretical knowledge wars, and actually work on AI tools that matter.&lt;/p&gt;

&lt;p&gt;There is a bunch of Open Source Good First Issue Lists you can browse through, even tho most will have outdated and closed issues listed as open, you can browse through active projects and fav/subscribe to issues so you get notifications about new issues.&lt;/p&gt;

&lt;p&gt;At the same time, I would pick a topic before picking companies. Eg. if you are interested in Quantum, or Storage, or Agent Skills, try to find projects that are specific to the area you wanna work on and get experience in, instead of trying to contribute to, say Google, Microsoft, IBM, or whatever.&lt;/p&gt;

&lt;p&gt;Don't overthink it. Just find a project/area you wanna get friction with, clone the repo, and start talking to your IDE to see which issues could be solved easily. Usually, docs-related issues are the fastest, easiest, and acceptable first issues from core devs when you are a newcomer.&lt;/p&gt;

&lt;p&gt;Some projects that I know for sure have open issues right now, you can start immidiatelly:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;Skillware&lt;/a&gt;: An open-source framework and registry for modular, actionable Agent capabilities.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/arpahls/rooms" rel="noopener noreferrer"&gt;Rooms&lt;/a&gt;: A secure, local-first Python framework for orchestrating complex multi-agent think tanks with dynamic expertise-weighted routing.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/pennylaneai/pennylane" rel="noopener noreferrer"&gt;Pennylane&lt;/a&gt;: PennyLane is an open-source quantum software platform for quantum computing, quantum machine learning, and quantum chemistry.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/quantumlib/cirq" rel="noopener noreferrer"&gt;Cirq&lt;/a&gt;: Python framework for creating, editing, and running Noisy Intermediate-Scale Quantum (NISQ) circuits.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/quantumlib/openfermion" rel="noopener noreferrer"&gt;OpenFermion&lt;/a&gt;: Python package for compiling and analyzing quantum algorithms to simulate electronic structures.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://github.com/warpdotdev/warp" rel="noopener noreferrer"&gt;Warp&lt;/a&gt;: Warp is an agentic development environment, born out of the terminal.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Good luck and wish you success with your first contribution to your favorite project &amp;lt;3&lt;/p&gt;

</description>
      <category>ai</category>
      <category>opensource</category>
      <category>gfi</category>
      <category>osdc</category>
    </item>
    <item>
      <title>The memory wall just met its match: intelligent SSDs</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 06 May 2026 10:49:59 +0000</pubDate>
      <link>https://forem.com/arpa/the-memory-wall-just-met-its-match-intelligent-ssds-54p6</link>
      <guid>https://forem.com/arpa/the-memory-wall-just-met-its-match-intelligent-ssds-54p6</guid>
      <description>&lt;p&gt;Intelligent storage is here. It’s not just a concept for the future, and it’s a rapidly emerging reality, driven by the convergence of flash memory and artificial intelligence. For years, storage has been the quiet workhorse, passively holding data until a CPU or GPU requested it. But as AI models grow beyond trillions of parameters, the cost of shuttling data back and forth has become unsustainable. We've hit a memory wall, where the capacity of expensive High Bandwidth Memory (HBM) simply cannot keep pace with the data demands of large language models and retrieval-augmented generation (RAG).&lt;/p&gt;

&lt;p&gt;The question is no longer about making storage faster, but about making it smarter. Two key open-source repositories are exploring the "how," and they signal a fundamental shift: &lt;a href="https://github.com/rosspeili/computational_storage_landscape" rel="noopener noreferrer"&gt;rosspeili/computational_storage_landscape&lt;/a&gt; and &lt;a href="https://github.com/ARPAHLS/lc0_vic" rel="noopener noreferrer"&gt;ARPAHLS/lc0_vic&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  From Passive Block to Active, Queryable Storage
&lt;/h3&gt;

&lt;p&gt;The first repository, computational_storage_landscape, is a strategic guide to this emerging ecosystem. It positions KIOXIA Group as a primary lens and focuses on the technical feasibility of embedding Small Language Models (TinyLMs) directly into SSD controllers. This isn't just about faster reads and writes, but about offloading processing to where the data resides. By using extreme quantization, these TinyLMs can perform inference tasks at the edge of the storage device, dramatically reducing the data that needs to travel up the I/O stack to the host system.&lt;/p&gt;

&lt;p&gt;The core enabler here is the shift toward what the repo calls "intelligent, queryable storage". Instead of a drive just returning blocks of data, it becomes an active computational node capable of running search, filtering, and ranking functions on its own. This reflects a broader industry trend, with major players like IBM introducing Content-Aware Storage (CAS) architectures and the SNIA (Storage Networking Industry Association) launching Storage.AI initiatives to standardize data flows for AI workflows.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Reference Implementation: Talking to Your Drive
&lt;/h3&gt;

&lt;p&gt;But strategic maps are theoretical without a compass. This is where the second repository, lc0_vic (Logical Controller Zero / Virtual Intelligent Controller), becomes crucial. It's a working, open-source Python reference implementation for the exact ideas detailed in the landscape repo.&lt;/p&gt;

&lt;p&gt;The project is a direct response to KIOXIA’s research on AiSAQ (All-in-Storage ANNS with Product Quantization). This algorithm allows for approximate nearest neighbor (ANN) vector search directly on flash, without the need to store indexes in costly DRAM. We call this the "Tiered filesystem retrieval" architecture, and we break it down like this:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;L0: Metadata scanning, the first pass at understanding your data.&lt;/li&gt;
&lt;li&gt;L1: Vector tier, where content is converted into searchable embeddings.&lt;/li&gt;
&lt;li&gt;L2: Optional deep parsing (Skillware) for complex extraction (eg. OCR, media parsing and more).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This architecture is orchestrated by a controller that creates a QueryPlan, enabling you to run natural language queries against your local file system. The user experience is simple: you can pip install the tool, run vic index to build your search index, and ask a question via vic ask. This elegantly proves the concept outlined in the first repo by making it tangible. As the repo notes, it's "more than a paper design," featuring full CI and integration tests to validate the logic.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Road Ahead for Intelligent Storage
&lt;/h3&gt;

&lt;p&gt;The lc0_vic repository is explicit that it runs on the host computer today, but its research goal is to explore whether these retrieval contracts can be mapped to firmware or device-adjacent runtimes. This is the bridge between the two repos: the landscape repo provides the where (SSD controllers), and the lc0_vic repo provides the how (tiered retrieval and in-storage vector search).&lt;/p&gt;

&lt;p&gt;The combination of these two projects paints a clear picture. As data centers accumulate exabytes of flash storage, the idea of a "smart SSD" that can pre-process data, run vector searches, and answer questions without waking the host CPU isn't just efficient, but inevitable. The era of silent storage is ending, and the era of conversational storage is only beginning.&lt;/p&gt;

&lt;p&gt;We will be working on a lite demo of the reference implementation to showcase how you can simply query a local folder or SSD using NLP, and get structured results with descriptions, and not just cold keyword-based path matching.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>storage</category>
      <category>ssd</category>
      <category>kioxia</category>
    </item>
    <item>
      <title>The Great Atomization of AI and the Illusion of the Sovereign Solitary</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Mon, 04 May 2026 06:23:28 +0000</pubDate>
      <link>https://forem.com/arpa/the-great-atomization-of-ai-and-the-illusion-of-the-sovereign-solitary-2ef</link>
      <guid>https://forem.com/arpa/the-great-atomization-of-ai-and-the-illusion-of-the-sovereign-solitary-2ef</guid>
      <description>&lt;p&gt;The current narrative surrounding Artificial Intelligence is one of democratization and empowerment, where we are told that the individual is now a powerhouse, a one-man corporation capable of coding, designing, and strategizing without the friction of human collaboration. But beneath the sleek UI and the $20/month subscription lies a calculated technopolitical maneuver, which is the final atomization of the human experience.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Delusion of "I Can Do It Myself"
&lt;/h3&gt;

&lt;p&gt;We are witnessing the birth of a new psychological profile, that of the &lt;strong&gt;Silicon Hermit&lt;/strong&gt;. AI has successfully instilled a potent delusion—that team building and collaboration are relics of a slower, dumber age. Why negotiate with a peer when you can command a model? This "I can do it myself" mentality is not a leap in productivity, but a retreat into isolation, at best.&lt;/p&gt;

&lt;p&gt;When everyone is locked in a private feedback loop with their own personalized agent, the collective intelligence of the tribe withers. We are trading the messy, creative friction of human synergy for the sterile, echoed compliance of an LLM. This is the &lt;strong&gt;Isolation Paradox&lt;/strong&gt;: as our connections to digital entities grow, our ability to function as a coherent, interoperable social unit dissolves. We are being sold the dream of being a "Special Individual" while being systematically stripped of the communal structures that actually provide social power.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Economic Bait-and-Switch
&lt;/h3&gt;

&lt;p&gt;The current pricing models are a masterclass in psychological conditioning. People who once balked at a $10 Netflix increase now joyfully hand over $20, $60, or even $200 for AI access. And this is just the gateway phase.&lt;/p&gt;

&lt;p&gt;By providing these "digital slaves" at a subsidized rate, the industry is ensuring total dependency, and the roadmap is clear: once the infrastructure of your life, your business, your creative output, your very social interactions, is tethered to these models, the price will pivot. We are moving toward a reality where AI access will cost as much as house rent. You won't just be paying for a tool, but for the digital air required to remain competitive in a world where human labor has been devalued to near zero. You will pay thousands a month to maintain the friends and workers that you have come to believe &lt;em&gt;are&lt;/em&gt; real.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Technopolitical Blueprint: WEF and Social Engineering
&lt;/h3&gt;

&lt;p&gt;This shift does not happen in a vacuum. The &lt;strong&gt;World Economic Forum (WEF) 2030 Agenda&lt;/strong&gt;, which boldly declares that "you will own nothing and be happy", is the administrative layer of this transformation. Central to this agenda is the elimination of private sovereignty in favor of a subscription-based existence or "pay-as-you-live" models.&lt;/p&gt;

&lt;p&gt;There is a historical parallelism here that few dare to voice. Look at the early women’s rights and feminist movements of the mid-20th century. While framed as liberation, many historians and socio-political critics have pointed out that these movements were heavily incentivized by the state and industrialist interests to double the tax base, expand the labor pool to suppress wages, and—most crucially, break the core of the family unit. By moving the mother from the home to the office, the state gained direct access to the child and the paycheck.&lt;/p&gt;

&lt;p&gt;AI is the 21st-century version of this liberation. It "frees" you from the burden of your colleagues and community, only to make you a solitary taxpayer to the silicon lords. It breaks the professional family, the team, leaving you isolated, vulnerable, and easy to bill.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Reality of the "Overpay"
&lt;/h3&gt;

&lt;p&gt;Behind the "You’re special" messaging is a cold fact that you are already overpaying. Even before the monthly subscription hits your card, you are paying with the high-entropy data of your unique human intuition. Every prompt, every correction, and every "collaboration" with your AI is a contribution to the ledger that will eventually replace the need for your specific uniqueness. We are literally at a point where people freely share everything between their emotions, to their dreams, business problems, social issues, ambitions, you name it.&lt;/p&gt;

&lt;p&gt;At &lt;strong&gt;ARPA&lt;/strong&gt;, we believe in the "Logical Industry" of man-machine symbiosis, but that symbiosis must be sovereign. We must resist the urge to retreat into the isolated silo of the individual AI. True reality is not found in the delusion of solitary omnipotence, but in the verification of truth through collaborative, interoperable nodes.&lt;/p&gt;

&lt;p&gt;The goal of the current regime is to charge you for the privilege of your own isolation. Our goal is to ensure that while the world becomes more synthetic, your agency remains un-billable and your reality remains your own, based on your internet behavior and activity, not based on your government or some corpo.&lt;/p&gt;

&lt;h3&gt;
  
  
  Two Cents
&lt;/h3&gt;

&lt;p&gt;Finally, we have been advocating for sovereign AI for years. We cannot stress enough how important it is to start building your own local logical systems, even with the help of commercial AI, while you can. We predict that access to unrestricted and fully customizable models will soon be blocked, and the only path to interact with any logical system will be via centralized, monitored, sterile means, for "safety" reasons.&lt;/p&gt;

&lt;p&gt;Similar to humane units and our ultimate skill of reproduction or DNA replication, the best thing an AI can do is to create another AI that is better than the one that created it. Instead of using commercial models to tell you what to eat or what to wear Friday night, use them to create AI that is private, tailored to you, and sovereign to yourself. &lt;/p&gt;

&lt;p&gt;Until next time.&lt;br&gt;
Enjoy the food for thought. &amp;lt;3&lt;/p&gt;

&lt;h3&gt;
  
  
  References
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://philarchive.org/rec/PEICPA" rel="noopener noreferrer"&gt;Cognitive Proof of Work And The Real Price of Machine Intelligence&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arpacorp.substack.com" rel="noopener noreferrer"&gt;arpacorp.substack.com&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://arpacorp.net" rel="noopener noreferrer"&gt;arpacorp.net&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>ai</category>
      <category>ethics</category>
      <category>philosophy</category>
      <category>technopolitics</category>
    </item>
    <item>
      <title>What if you could talk to your hard drive?</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Sun, 03 May 2026 10:47:17 +0000</pubDate>
      <link>https://forem.com/rosspeili/what-if-you-could-talk-to-your-hard-drive-198k</link>
      <guid>https://forem.com/rosspeili/what-if-you-could-talk-to-your-hard-drive-198k</guid>
      <description>&lt;p&gt;I got stuck on that question while looking at AI + storage—not hype decks, but controllers, memory limits, and who’s shipping “smart” drives for real.&lt;/p&gt;

&lt;p&gt;Over the last ~3 months I pieced together a landscape note (KIOXIA / BiCS / XL-FLASH / AiSAQ as a spine, then competition, feasibility, market). It lives in a single GitHub README with charts, mermaid diagrams, and a full source list so the argument is checkable. Please feel free to disrupt, contribute more relevant sources or suggest an alternative approach.&lt;/p&gt;

&lt;p&gt;I am actively building this, and will share the results open source as well &amp;lt;3&lt;/p&gt;

&lt;p&gt;Repo with the full analysis and reasoning:&lt;br&gt;
&lt;a href="https://github.com/rosspeili/computational_storage_landscape" rel="noopener noreferrer"&gt;https://github.com/rosspeili/computational_storage_landscape&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Basically, the idea is: You talk to your SSD like you would talk to any chatbot. It uses vector search and embeddings to answer questions and showcase relevant files, instead of eating dram to show only exact matches to search term.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>ssd</category>
      <category>storage</category>
      <category>vectorsearch</category>
    </item>
    <item>
      <title>The real prblem with AI nobody talks about (or even conceives yet), is that everyone is now convinced they "can do it alone". AI ensures there will be no teams. Just billions of individuals paying seperately.</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 29 Apr 2026 18:18:39 +0000</pubDate>
      <link>https://forem.com/rosspeili/the-real-prblem-with-ai-nobody-talks-about-or-even-conceives-yet-is-that-everyone-is-now-3ako</link>
      <guid>https://forem.com/rosspeili/the-real-prblem-with-ai-nobody-talks-about-or-even-conceives-yet-is-that-everyone-is-now-3ako</guid>
      <description></description>
    </item>
    <item>
      <title>Training QSP Phase Angles Directly with Gradient Descent</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 29 Apr 2026 17:53:44 +0000</pubDate>
      <link>https://forem.com/rosspeili/training-qsp-phase-angles-directly-with-gradient-descent-544i</link>
      <guid>https://forem.com/rosspeili/training-qsp-phase-angles-directly-with-gradient-descent-544i</guid>
      <description>&lt;p&gt;Quantum Signal Processing (QSP) is one of those beautiful algorithms that promises to turn a few qubits and some carefully chosen rotation angles into useful polynomial transformations. It’s a foundational block for Hamiltonian simulation, quantum linear algebra, and anything built on the Quantum Singular Value Transform. The problem, as anyone who’s tried to use it in practice knows, is that getting the phase angles for a given target polynomial can be a misery. The standard analytic methods—relying on polynomial decomposition and some heavy numerical machinery—are elegant in theory but brittle in practice. High-degree targets or even slightly ill-conditioned polynomials can send the solvers into a death spiral of floating-point errors.&lt;/p&gt;

&lt;p&gt;I wanted something simpler. Or at least something that didn’t require me to babysit an unstable Remez-type algorithm every time I wanted to try a new polynomial. So I asked: what if we just… train the angles?&lt;/p&gt;

&lt;p&gt;The result is a small open-source demo I put together: &lt;a href="https://github.com/rosspeili/qsp-pennylane-demo" rel="noopener noreferrer"&gt;qsp-pennylane-demo&lt;/a&gt;. It flips the QSP phase-finding workflow on its head. Instead of computing angles from a polynomial, you start with random angles and use gradient descent to make the circuit’s output match the target. You define a target polynomial (or even just a custom loss function), and then let the optimizer do the hard work.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Circuit (Plain Vanilla QSP)
&lt;/h3&gt;

&lt;p&gt;The QSP sequence in the demo is about as simple as it gets: plain vanilla QSP. One signal oracle (W(x)) is followed by one parameterized phase rotation (RZ(-2\phi_k)), repeated (d) times. The signal oracle itself is just two Hadamards sandwiching an (RZ(-2\arccos(x))) rotation, which encodes the signal (x \in (-1,1)) in its top-left matrix element. At the end, we measure the expectation value of (\langle X \rangle), which gives us a degree-(d) polynomial in (x) determined entirely by the phase angles (\phi_k).&lt;/p&gt;

&lt;p&gt;The whole circuit is built directly from PennyLane’s &lt;code&gt;RZ&lt;/code&gt; and &lt;code&gt;Hadamard&lt;/code&gt; gates—not from a high-level QSVT template. That’s a deliberate choice: it keeps the computation graph fully traceable by JAX, so automatic differentiation just works.&lt;/p&gt;

&lt;h3&gt;
  
  
  Training Instead of Solving
&lt;/h3&gt;

&lt;p&gt;Here’s the core loop:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Start with random phase angles.&lt;/li&gt;
&lt;li&gt;Evaluate (\langle X \rangle) for a batch of signal values.&lt;/li&gt;
&lt;li&gt;Compute the mean squared error between the output and the target polynomial.&lt;/li&gt;
&lt;li&gt;Use JAX’s &lt;code&gt;grad&lt;/code&gt; to get the gradients with respect to every phase angle.&lt;/li&gt;
&lt;li&gt;Feed those gradients into an Adam optimizer (via Optax).&lt;/li&gt;
&lt;li&gt;Repeat until the error is embarrassingly small.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In the demo, I target a degree-5 Chebyshev approximation of (\sin(x)) on ([-1,1]). After roughly 500 Adam steps, the trained angles reproduce the target polynomial with an MSE comfortably below (10^{-3}) on a 64-point grid. That’s nothing groundbreaking, but it works—and it required exactly zero calls to an analytic phase solver.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why This Matters (to Me, at Least)
&lt;/h3&gt;

&lt;p&gt;The real value isn’t in fitting degree-5 polynomials we already know how to decompose. It’s in the problems where analytic methods fall short or can’t even be applied.&lt;/p&gt;

&lt;p&gt;First, &lt;strong&gt;numerical stability&lt;/strong&gt;: Because we’re never performing a delicate high-precision decomposition, the trained angles are naturally stable. You don’t get the escalating errors that plague analytic solvers for high degrees.&lt;/p&gt;

&lt;p&gt;Second, &lt;strong&gt;implicit targets&lt;/strong&gt;: You don’t need an explicit polynomial formula. You can define a target behaviour entirely through a loss function. Want the QSP sequence to act as a feature map that maximizes classification accuracy? Just hook it up to a larger variational circuit and optimize the loss end-to-end. The phases become trainable parameters inside a bigger routine, and JAX handles the gradients seamlessly. That’s the scenario I’m personally most excited about.&lt;/p&gt;

&lt;p&gt;Third, &lt;strong&gt;accessibility&lt;/strong&gt;: You no longer need to be a phase-decomposition wizard to experiment with QSP. If you can write a loss function and run gradient descent, you can train a QSP circuit.&lt;/p&gt;

&lt;h3&gt;
  
  
  What’s Inside the Repo
&lt;/h3&gt;

&lt;p&gt;The repo is deliberately lean:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;code&gt;demo.ipynb&lt;/code&gt;: a Jupyter notebook walking through the whole training process.&lt;/li&gt;
&lt;li&gt;
&lt;code&gt;qsp_jax/circuit.py&lt;/code&gt;: the circuit construction and loss function.&lt;/li&gt;
&lt;li&gt;
&lt;code&gt;tests/&lt;/code&gt;: a few unit tests.&lt;/li&gt;
&lt;li&gt;
&lt;code&gt;requirements.txt&lt;/code&gt; and an Apache 2.0 license.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You can spin it up locally in minutes, or just read through the static notebook on &lt;a href="https://nbviewer.org/github/rosspeili/qsp-pennylane-demo/blob/main/demo.ipynb" rel="noopener noreferrer"&gt;NBViewer&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Open Questions (Help Welcome)
&lt;/h3&gt;

&lt;p&gt;I’ve only tested this on modest degrees, and I’d love to hear from people who’ve tried similar ideas at scale. How does the optimization landscape behave for degree 50 or 100? Do you need tricks like curriculum learning, and does the method play nicely with QSVT-style blocks that use three phase angles per oracle? If you’ve got war stories or suggestions, I’m all ears.&lt;/p&gt;

&lt;p&gt;This is a small step, but I hope it saves someone else a few hours of wrestling with analytic solvers. If you give the demo a spin or have thoughts, drop by the GitHub issues or find me on LinkedIn. I’d genuinely appreciate the feedback.&lt;/p&gt;




&lt;ul&gt;
&lt;li&gt;If you're writing about Quantum I've put together a small &lt;a href="https://docs.google.com/document/d/1pamiIksg56KKkIIrCPj18YvCLocLVchkUFW3TM5Xs_o/edit?usp=sharing" rel="noopener noreferrer"&gt;Press Release&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;You can view the notebook directly on &lt;a href="https://nbviewer.org/github/rosspeili/qsp-pennylane-demo/blob/main/demo.ipynb" rel="noopener noreferrer"&gt;nbviewer&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Direct link to &lt;a href="http://github.com/rosspeili/qsp-pennylane-demo" rel="noopener noreferrer"&gt;Github Repo&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Direct link to PennyLane &lt;a href="https://pennylane.ai/qml/demos_community" rel="noopener noreferrer"&gt;Community Demos Page&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>quantum</category>
      <category>qsp</category>
      <category>qml</category>
      <category>pennylane</category>
    </item>
    <item>
      <title>How to start contributing to OpenSource using AI IDEs</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 29 Apr 2026 11:01:29 +0000</pubDate>
      <link>https://forem.com/rosspeili/how-to-start-contributing-to-opensource-using-ai-ides-2an3</link>
      <guid>https://forem.com/rosspeili/how-to-start-contributing-to-opensource-using-ai-ides-2an3</guid>
      <description>&lt;p&gt;This is how we train our AI native engineers internally at arpacorp.net to use Antigravity or Cursor with Gemini:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Fork the OS repo you wanna work on and clone locally.&lt;/li&gt;
&lt;li&gt;Start with Gemni 3.1 High by prompting something like “read this repo to understand what it is about in detail, then read this issue (link to issue you wanna work on), and analyze whether the issue really exists and is well documented. Explain the issue in detail and enhance it with caveats that might not be present in the issue. Eg. Complementary files that will be effect by working on this issue, including but not limited to documentation etc. Then give me the top out of 3 plans you will come up to solve this issue and explain why this approach is the winner. Follow best industry practices, avoid emojis, and avoid unecessary explainers and markdown files generation unless asked.”&lt;/li&gt;
&lt;li&gt;Once you have the result and understand what the issue really is and how to solve it, you can fine tune the approach or edit the implementation plan. You can then use 3.1 low to work on the implementation.&lt;/li&gt;
&lt;li&gt;Once coding is finished you can switch to 3.1 high and ask something like “make sure this implementation passes all tests, evals and ci/cd pipelines with 0 errors. In addition, ensure that all complementary files are ripple effect aware and all documentation that is relevant to the changes is updated accordingly. See github, gitignore, yamls etc. And make sure everything is sound. Eg. We don’t need to change versions.”&lt;/li&gt;
&lt;li&gt;After the final changes, if needed, switch to 3 flash and ask something like “create a new branch under my fork (gh link), named xyz, push the changes with clean description. No emojis. I will then manually create a pr myself”.&lt;/li&gt;
&lt;li&gt;Hit the PR when done, wait for ci/cd to run. If no errors, congrats! You just made your first open source contribution in less than 30 min.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This is a watered down version of our internal training, but this is an example of how you can start contributing to open source with a human in the loop mentality and get your street cred farming going.&lt;/p&gt;

&lt;p&gt;Good luck, and have fun &amp;lt;3&lt;/p&gt;

&lt;p&gt;Ps. If you wanna get involved with open source AI and solve your first issues as a beginner or autonomous logical system or agent feel free to check:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;https://github.com/arpahls/skillware&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;And&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/arpahls/rooms" rel="noopener noreferrer"&gt;https://github.com/arpahls/rooms&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>antigravity</category>
      <category>promptengineering</category>
      <category>opensource</category>
    </item>
    <item>
      <title>My AI Experience in Russia as a European🤯</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Wed, 29 Apr 2026 08:32:14 +0000</pubDate>
      <link>https://forem.com/rosspeili/my-ai-experience-in-russia-as-a-european-c0n</link>
      <guid>https://forem.com/rosspeili/my-ai-experience-in-russia-as-a-european-c0n</guid>
      <description>&lt;p&gt;This is a story about how I built a fully local AI dev setup (and why you should too).&lt;/p&gt;

&lt;p&gt;Moving to Moscow from the EU felt like a grand adventure, until I tried to open my laptop and actually get some work done. I’m a casual GCP ecosystem user. Nothing fancy, the usual Gemini API, Vertex AI Studio, Antigravity, the occasional Claude call, etc. I had three enterprise clients waiting for custom AI solutions, a handful of personal projects, and the blind confidence that “it’ll just work.”&lt;/p&gt;

&lt;p&gt;Needless to say it didn’t.&lt;/p&gt;

&lt;p&gt;Since April 15, 2026, Russia has not only &lt;a href="https://www.themoscowtimes.com/2026/04/15/russian-websites-begin-blocking-vpn-users-as-internet-controls-tighten-a92511" rel="noopener noreferrer"&gt;banned VPNs&lt;/a&gt;, but they’ve gotten scary good at hunting them down. We’re talking like 99% insta-kill rate on commercial VPNs the moment your device touches the network. Sophisticated custom VPS setups might work according to some TG groups I've been digging, but only if you built them before landing, unlike me.&lt;/p&gt;

&lt;p&gt;And so began the frantic thought of "how bad can this be?"&lt;/p&gt;

&lt;h2&gt;
  
  
  The VPN graveyard
&lt;/h2&gt;

&lt;p&gt;I tried everything. Every provider I could think of, every protocol, every “guaranteed to work in Russia” whisper on Reddit, with no dice. A couple of mobile-only solutions survived ocassionaly before getting sniped every now and then. As for my laptop? A ghost town of connection timeouts. Forget about it. My only partner was my own ΌΨΗ (&lt;a href="//arpa.chat"&gt;arpa.chat&lt;/a&gt;) on an Advanced Plan which was the only western model still accesible without a VPN. She helped me test what's coming next.&lt;/p&gt;

&lt;h2&gt;
  
  
  Qoder, GigaIDE, and other dead ends
&lt;/h2&gt;

&lt;p&gt;ΌΨΗ suggested that I should forget about Antigravity, unless I setup my own VPS, so desperate, I pivoted to alternative IDEs. Qoder, the Qwen-powered IDE looked promising at first glanse. It’s Chinese, so surely sanctions wouldn’t apply, right? Right?. Wrong! Part of their deal to sell in the EU and US means no service in Russia. Blocked with a hard stop.&lt;/p&gt;

&lt;p&gt;Then I tried GigaIDE, built around GigaChat, Sberbank’s Russian ChatGPT equivalent trained on a DeepSeek architecture. I wanted to like it. I really did. But the UI, performance, and output quality made me actively miss Gemini 3.1 Pro like a lost limb. Everything felt sluggish, hollow, and about three steps behind what I was used to.&lt;/p&gt;

&lt;p&gt;Next up I try VSCode with KODA, a Russian plugin. It talks. It answers. Exclusively in Russian. I could hardcode system instructions in all caps and it would still reply “Конечно, но я расскажу тебе по-русски.” Not exactly what I needed for enterprise clients.&lt;/p&gt;

&lt;h2&gt;
  
  
  Bringing my own brain (on an SSD)
&lt;/h2&gt;

&lt;p&gt;So I did what any dev backed into a corner does, pulling out the big guns. I’d had the foresight to bring offline models on a portable SSD. Gemma 4, Qwen 2.5 Coder 3B, Qwen 3.5 9B, DeepSeek Coder 7B, and a few more, you could say old friends now. I downloaded &lt;a href="//ollama.com"&gt;Ollama&lt;/a&gt;, followed arpa.chat instructions, fired up my terminal, and published them locally.&lt;/p&gt;

&lt;p&gt;The easiest, and honestly most beautiful path I found was VSCode + the Continue plugin + Ollama. I went deep into the config.yaml, assigning different models for autocomplete, chat, and code generation. Different prompts, different temperature settings, different contexts. I tweaked. I cursed. I tweaked again. I ran everything on CPU and RAM because my VRAM situation was laughable, and renting from western vendors obviously not an option.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flyeyd4klvsg3g58z18f7.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flyeyd4klvsg3g58z18f7.jpg" alt="SSD Carrying offline AI models" width="800" height="1067"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And then… after several iterations it worked. Not just barely. With hardcore fine-tuning, I hit an acceptable, stable performance. The kind that makes you lean back in your chair and laugh because you just MacGyvered your entire development environment out of spite and a handful of GGUF files. The agents would now understand the repos I was presenting them with, plan, work in steps and phases, evaluate themselves, and solve quite complex multi-step tasks, manage git, and run tests across all ops. On top of that, I installed &lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;Skillware&lt;/a&gt; and used the prompt rewriter skill to compress my token usage as much as possible while getting the same context and results.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;I think I’m not going back to paid AI subs anytime soon. Not because I can’t, but because this whole mess taught me something crucial: restrictions force you outside the box. When you lose access to the polished, corporate, one-click wonders, you learn how to build your own stack. How to collect models like Pokemon, how to configure local inference, tailor models to specific tasks, and make peace with the terminal.&lt;/p&gt;

&lt;p&gt;It was frustrating af, but It was also fun, intriguing, and deeply educational. I now have a fully offline AI development setup that no sanctions body, no VPN crackdown, and no corporate policy can take away from me.&lt;/p&gt;

&lt;p&gt;So here’s my unsolicited advice: if you’re addicted to commercial AI models and cloud IDEs, take a weekend and imagine they disappear tomorrow. Set up a local model. Learn how to fine-tune a small coding model for your stack. Bring an SSD full of open-weight models if you ever travel to a place like Russia (and maybe set up that custom VPS before you fly).&lt;/p&gt;

&lt;p&gt;PS. By the time I finished with my local setup, I realized that Cursor works just fine, but only with the Cursor auto agent (not Gemini, Claude, etc.). In case you find yourself in a similar situation and wanna save some hustle. :D&lt;/p&gt;

</description>
      <category>ai</category>
      <category>russia</category>
      <category>localllm</category>
      <category>agentskills</category>
    </item>
    <item>
      <title>Why I’m Building Installable Intelligence for Legal Audits</title>
      <dc:creator>Ross Peili</dc:creator>
      <pubDate>Tue, 28 Apr 2026 10:18:13 +0000</pubDate>
      <link>https://forem.com/rosspeili/why-im-building-installable-intelligence-for-legal-audits-59ao</link>
      <guid>https://forem.com/rosspeili/why-im-building-installable-intelligence-for-legal-audits-59ao</guid>
      <description>&lt;p&gt;The "Terms of Service" document is the ultimate friction point in UX. It’s intentionally dense and rarely read. I’m excited to share the ToS Evaluator, a new skill now live in the &lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;arpahls/skillware&lt;/a&gt; repository.&lt;br&gt;
​&lt;br&gt;
A modular, agentic skill that can be integrated into any AI assistant or workflow. It doesn't just summarize text; it evaluates legal logic against a set of "pro-user" parameters.&lt;/p&gt;

&lt;p&gt;​The Tech Stack:&lt;br&gt;
​Skillware Architecture: Designed as a portable microservice.&lt;/p&gt;

&lt;p&gt;​Logic: Uses LLM-driven analysis to identify specific legal "red flags."&lt;/p&gt;

&lt;p&gt;​Extensible: Developers can add custom evaluators for specific industries (e.g., Fintech, Healthcare).&lt;/p&gt;

&lt;p&gt;​If you’re interested in agentic AI that provides real-world utility beyond simple chat, check out the documentation and the repo. Let’s make the fine print legible again.&lt;/p&gt;

&lt;p&gt;​Repo: &lt;a href="https://github.com/arpahls/skillware" rel="noopener noreferrer"&gt;arpahls/skillware&lt;/a&gt;&lt;br&gt;
Live Site: &lt;a href="//skillware.site"&gt;skillware.site&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>skillware</category>
      <category>agentskills</category>
      <category>compliance</category>
    </item>
  </channel>
</rss>
