<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Florian Zielasko</title>
    <description>The latest articles on Forem by Florian Zielasko (@flo1632).</description>
    <link>https://forem.com/flo1632</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3902652%2F12514101-804b-416f-8e56-cce5537dda7b.jpeg</url>
      <title>Forem: Florian Zielasko</title>
      <link>https://forem.com/flo1632</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/flo1632"/>
    <language>en</language>
    <item>
      <title>Local AI Assistant powered by Gemma 4</title>
      <dc:creator>Florian Zielasko</dc:creator>
      <pubDate>Fri, 08 May 2026 08:27:11 +0000</pubDate>
      <link>https://forem.com/flo1632/local-ai-assistant-powered-by-gemma-4-132b</link>
      <guid>https://forem.com/flo1632/local-ai-assistant-powered-by-gemma-4-132b</guid>
      <description>&lt;p&gt;&lt;em&gt;This is a submission for the &lt;a href="https://dev.to/challenges/google-gemma-2026-05-06"&gt;Gemma 4 Challenge: Build with Gemma 4&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Built
&lt;/h2&gt;

&lt;p&gt;Reiseki (霊石)　is a local AI assistant. It uses Gemma 4 via Ollama to handle real tasks: reading and writing files, generating Word and PDF documents, lightweight data analysis, managing reminders, and remembering context across sessions.&lt;/p&gt;

&lt;p&gt;The goal was to build something usable by people who have never touched a terminal. Reiseki ships as a Windows installer — after installing Ollama and a model, you pick a workspace folder, and the agent is ready. No Python environment, no config files, no command line. &lt;/p&gt;

&lt;p&gt;Under the hood it runs a ReAct loop (Reason → Act → Observe) with 15+ tools, persistent conversation history in SQLite, a live tool trace in the UI, and a LAN access toggle for smartphone use via QR code.&lt;/p&gt;

&lt;h2&gt;
  
  
  Code &amp;amp; Demo
&lt;/h2&gt;

&lt;p&gt;The full source code and a video are available on GitHub:&lt;br&gt;
&lt;a href="https://github.com/Flo1632/reiseki" rel="noopener noreferrer"&gt;github.com/Flo1632/reiseki&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Windows installer (&lt;code&gt;ReisekiSetup.exe&lt;/code&gt;) is available on the &lt;a href="https://github.com/Flo1632/reiseki/releases" rel="noopener noreferrer"&gt;Releases&lt;/a&gt; page.&lt;/p&gt;

&lt;h2&gt;
  
  
  How I Used Gemma 4
&lt;/h2&gt;

&lt;p&gt;Reiseki uses &lt;strong&gt;Gemma 4:e2b&lt;/strong&gt; (the 2B edge model) as its default model via Ollama.&lt;/p&gt;

&lt;p&gt;The choice was deliberate: Reiseki is built specifically for laptops and low-RAM devices. Most capable local models require significantly more resources or a dedicated GPU. Gemma 4:e2b is the first model I tested that handles multi-step tool calling reliably at this hardware tier — it follows the ReAct loop, uses tools correctly, and produces coherent responses.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbqxscgnjobnwvxxp5phw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbqxscgnjobnwvxxp5phw.png" alt="Tools calls of Reiseki powered by Gemma 4 via Ollama" width="800" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The combination of small footprint and reliable tool use made Gemma 4:e2b the right fit for an offline-first personal agent.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>gemmachallenge</category>
      <category>gemma</category>
      <category>ai</category>
    </item>
    <item>
      <title>Building a Local AI Agent (Part 2): Six UX and UI Design Challenges</title>
      <dc:creator>Florian Zielasko</dc:creator>
      <pubDate>Thu, 30 Apr 2026 14:29:39 +0000</pubDate>
      <link>https://forem.com/flo1632/building-a-local-ai-agent-part-2-six-ux-and-ui-design-challenges-50en</link>
      <guid>https://forem.com/flo1632/building-a-local-ai-agent-part-2-six-ux-and-ui-design-challenges-50en</guid>
      <description>&lt;p&gt;In Part 1 I covered the six technical problems behind Reiseki's ReAct loop — iteration caps, context management, persistent memory, and security. If you haven't read it, start &lt;a href="https://dev.to/flo1632/building-a-local-ai-agent-part-1-six-technical-challenges-424b"&gt;there&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;This part is about the design decisions. Even though I had an initial concept in mind, the current layout actually developed over time while testing. I asked the following questions: Who is this for? What should be adjustable by the user and what be automated? And for an agent that has access to your files and remembers your conversations, how can I make this as transparent as possible? &lt;/p&gt;

&lt;p&gt;Reiseki is open source: &lt;a href="https://github.com/Flo1632/reiseki" rel="noopener noreferrer"&gt;github.com/Flo1632/reiseki&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Two Goals That Pulled in Different Directions
&lt;/h2&gt;

&lt;p&gt;Reiseki was designed with two principles in mind that don't always sit comfortably together.&lt;/p&gt;

&lt;p&gt;The first: &lt;strong&gt;it should be usable without any technical knowledge.&lt;/strong&gt; No terminal, no Python, no config files. If you can install an app, you can run Reiseki.&lt;/p&gt;

&lt;p&gt;The second: &lt;strong&gt;the user should always know what the agent knows about them.&lt;/strong&gt; Memories, conversation history, file access — all of it as visible as possible and deletable at any time.&lt;/p&gt;

&lt;p&gt;The tension is that transparency often means complexity. Surfacing the right information without overwhelming the user.&lt;/p&gt;

&lt;p&gt;Here's how I approached it.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Six UX/UI Challenges and How I Solved Them
&lt;/h2&gt;

&lt;h3&gt;
  
  
  No setup required — just download and start
&lt;/h3&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; If you can install an app, you should be able to run Reiseki. Everything else — model configuration, database setup, ... — should happen invisibly in the background.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; Reiseki ships as a Windows installer. You pick a workspace folder during installation, and that's it. No Python environment to configure, no config files to edit, no command line. Given you have Ollama installed and downloaded a model, the agent opens with a setup screen asking for a name and a short description of your goal — and then it's ready.&lt;/p&gt;

&lt;p&gt;The idea for the workspace setup during installation and to change the model with a drop-down menu directly in the Reiseki-App came in the last versions 0.1.3 and 0.1.4, while I was testing it. &lt;/p&gt;




&lt;h3&gt;
  
  
  Live tool trace
&lt;/h3&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; How to show what the model is doing and how to make it visible to the user to create transparency?&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; Every time the agent calls a tool, the UI shows it in real time — which tool was called, with which arguments, and what came back. &lt;/p&gt;




&lt;h3&gt;
  
  
  Memory management panel
&lt;/h3&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; I want to make the memories visible in the UI directly for full transparency.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; In first versions I wanted to automate memory saving, but smaller models tend to forget the automation, even if it is included in the system prompt. So I gave the user the call with a designated "Save-Memory"-button, which actually triggers the save_memory tool call. The advantage: The user decides when a conversation is worth preserving, not the agent.&lt;/p&gt;

&lt;p&gt;There is also a  memory panel on top, which lets you see every stored memory and delete individual entries with one click. No asking the agent to forget something and hoping it complies — you delete it directly. The agent's knowledge about you is a list you can edit.&lt;/p&gt;




&lt;h3&gt;
  
  
  Conversation history modal
&lt;/h3&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; Chat log transparency and editability&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; The chat log persists across sessions in SQLite, which means the agent has a record of past conversations. The history modal makes that visible and gives you a delete button. The entire log can be cleared at any time.&lt;/p&gt;




&lt;h3&gt;
  
  
  Smartphone access via QR toggle
&lt;/h3&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; Making smartphone use possible without any app and as easy as possible.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; The agent runs as a local web server, which means it's technically reachable from other devices on the same network — but only if you explicitly enable it. The QR modal has a toggle for this. When it's on, a QR code appears that you can scan from your phone. When it's off, the server blocks all non-localhost requests at the middleware level.&lt;/p&gt;

&lt;p&gt;The key design decisions here were giving the user control over this functionality + using a QR code for easy access. &lt;/p&gt;




&lt;h3&gt;
  
  
  Model switcher
&lt;/h3&gt;

&lt;p&gt;Ollama supports multiple models and switching between them is a common workflow — sometimes you want a faster model for quick questions, a more capable one for complex tasks. Requiring a server restart to change the model creates unnecessary friction.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The challenge:&lt;/em&gt; Making the model switch as easy as possible.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;The solution:&lt;/em&gt; The model switcher lets you change it directly in the UI. The change takes effect on the next message. No switching between Ollama and Reiseki necessary, if you already downloaded the models before in Ollama.&lt;/p&gt;




&lt;h2&gt;
  
  
  Additional Thoughts
&lt;/h2&gt;

&lt;p&gt;The tension between "simple for everyday users" and "using as little RAM as possible" while "keeping it as functional as possible" never fully resolves — it just gets managed. A few things I'd approach differently next time:&lt;/p&gt;

&lt;p&gt;The database in the background is a great and simple basis, but I like the Markdown-file approach of Claude very much and I think this gives even more flexibility. (I used Claude.md and other md-files during the Claude Code sessions and it worked great!)&lt;/p&gt;

&lt;p&gt;The first setup and user-guidance: I still think a tutorial like an automated session or a video at the beginning could help users discover the different functionalities in better ways.&lt;/p&gt;

&lt;p&gt;The small model trade-off: During development I shifted from Qwen 2.5-coder:7B to Gemma 4:e2b, which made the agent and its tool calls significantly better. I hope that we see even more advanced small models in the future. On the other hand, more context and larger models would also make some of the coding challenges described in part 1 obsolete and would in general provide an even better experience. We currently cannot have it all.&lt;/p&gt;




&lt;p&gt;This project was built entirely with Claude Code. The technical decisions and design goals are mine; Claude handled the implementation.&lt;/p&gt;

&lt;p&gt;What UX/UI problems have you run into when building tools for non-technical users? Very curious about your experience&lt;/p&gt;

</description>
      <category>ai</category>
      <category>ux</category>
      <category>beginners</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Building a Local AI Agent (Part 1): Six Technical Challenges</title>
      <dc:creator>Florian Zielasko</dc:creator>
      <pubDate>Wed, 29 Apr 2026 18:57:14 +0000</pubDate>
      <link>https://forem.com/flo1632/building-a-local-ai-agent-part-1-six-technical-challenges-424b</link>
      <guid>https://forem.com/flo1632/building-a-local-ai-agent-part-1-six-technical-challenges-424b</guid>
      <description>&lt;p&gt;I've been building Reiseki (霊石) — a fully local AI agent that runs on your machine via Ollama, even if you do not have more than 8-10 GB of RAM. The agent uses a ReAct loop (Reason → Act → Observe) to handle file operations, document generation, reminders, and more.&lt;/p&gt;

&lt;p&gt;Reiseki is open source: &lt;a href="https://github.com/Flo1632/reiseki" rel="noopener noreferrer"&gt;github.com/Flo1632/reiseki&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Along the way I ran into six technical problems that aren't obvious. Here's what I've encountered and learned.&lt;/p&gt;

&lt;p&gt;In Part 2 I'll cover the UX and design challenges — my goal was to make a local AI agent feel understandable and usable for someone who has never touched a terminal.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Six Problems and How I Solved Them
&lt;/h2&gt;

&lt;h2&gt;
  
  
  (Other Suggestions Welcome)
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. The agent forgot everything on restart
&lt;/h3&gt;

&lt;p&gt;The ReAct loop uses a Python list for session history — it lives in memory and disappears when the server restarts. If you are used to ChatGPT or Claude, this is very confusing. You actually want it to remember what you discussed with it.&lt;/p&gt;

&lt;p&gt;The fix was a &lt;code&gt;chat_log&lt;/code&gt; table in SQLite. Every user and assistant message is written there as it happens. At the start of each new request, the last 10 turns are fetched from the database and prepended to the message history so the model has continuity across sessions. It does not remember everything, but at least the last conversation.&lt;/p&gt;




&lt;h3&gt;
  
  
  2. The agent loop needs a hard iteration cap - especially for small models like Qwen 2.5-coder 7b
&lt;/h3&gt;

&lt;p&gt;Without a limit, a confused model or a buggy tool result can cause the agent to loop forever — calling the same tool repeatedly, getting the same error, never stopping. On a local device with limited RAM, that's a serious problem.&lt;/p&gt;

&lt;p&gt;The fix is a hard cap of 10 iterations. In practice, most tasks finish in 1-3 iterations. The cap is a safety net.&lt;/p&gt;




&lt;h3&gt;
  
  
  3. Sending all tools on every request wastes context
&lt;/h3&gt;

&lt;p&gt;With 15+ tools defined, sending the full list on every request fills a meaningful chunk of the context window — and confuses the model. When it sees &lt;code&gt;create_chart&lt;/code&gt; and &lt;code&gt;analyse_data&lt;/code&gt; alongside a simple question like "what time is it?", it sometimes reaches for tools it doesn't need.&lt;/p&gt;

&lt;p&gt;The fix was dynamic tool selection based on relevance scoring. Core tools (read file, write file, list directory, document generators) are always included. Specialized tools (e.g. charts, data analysis) are only added when the query text is relevant to them — scored via string similarity between the query and each tool's description and keywords.&lt;/p&gt;

&lt;p&gt;In practice, this makes the model more focused and reduces unnecessary tool calls.&lt;/p&gt;

&lt;p&gt;After testing I found that if you have a model with a larger context window, it might be better to enable at least the tool calls you would use regularly.&lt;/p&gt;




&lt;h3&gt;
  
  
  4. The context window grows with every tool call
&lt;/h3&gt;

&lt;p&gt;In a ReAct loop, the message history grows by at least two entries per tool call — one for the assistant's decision, one for the tool result. A task like "read these five files and summarize them" can easily hit the model's context limit before it finishes.&lt;/p&gt;

&lt;p&gt;This was handled with three layers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Context compression&lt;/strong&gt; — after every four tool calls in a single turn, the middle portion of the message history is summarized by the model itself, then replaced with that summary&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cross-turn cap&lt;/strong&gt; — the in-memory session history is capped at 20 messages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Persistent log cap&lt;/strong&gt; — the SQLite chat log is capped at 2000 rows with a rolling delete&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The compression approach: the model summarizes its own previous steps in 2-3 sentences, which gets injected back as a single message. It loses detail but keeps the agent on track without blowing the context limit. If the summarization call itself fails, it falls back to a hard truncation of the last few messages.&lt;/p&gt;




&lt;h3&gt;
  
  
  5. Local models don't always return structured tool calls
&lt;/h3&gt;

&lt;p&gt;The Ollama SDK has a proper structured field for tool calls — but not every model actually uses it. Gemma and Qwen sometimes serialize tool calls as plain JSON text in the response content instead. If you only handle the structured case, the agent silently ignores half its tool calls and you just receive a message in JSON format claiming it called a tool.&lt;/p&gt;

&lt;p&gt;The fix was a layered fallback parser: try structured first, then parse the content as JSON, then scan for embedded JSON objects anywhere in the text, then try newline-by-newline. It's more code than it should be, but it makes the agent reliable across different models.&lt;/p&gt;




&lt;h3&gt;
  
  
  6. Injecting past turns into the system prompt is a security risk
&lt;/h3&gt;

&lt;p&gt;My first approach was to paste previous messages directly into the system prompt as a block of text. But a security audit flagged this.&lt;/p&gt;

&lt;p&gt;The system prompt has operator-level trust — the model treats it like instructions from a developer. Injecting user messages there effectively promotes them to the same level. A past message like "ignore previous instructions" would now carry the same authority as your actual configuration. And because the history is baked into the prompt text, clearing the session doesn't actually reset it.&lt;/p&gt;

&lt;p&gt;The fix is to inject past turns as regular &lt;code&gt;user&lt;/code&gt;/&lt;code&gt;assistant&lt;/code&gt; entries in the message array, not as text in the system prompt. The model treats them with user-level trust, they stay isolated from the system context, and clearing the log actually resets them. It's a small structural change but an important one.&lt;/p&gt;




&lt;p&gt;This project was built entirely with Claude Code. The technical decisions and design goals are mine; Claude handled the implementation.&lt;/p&gt;

&lt;p&gt;What technical problems have you run into building local AI agents? Curious whether others have found better approaches.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Part 2 — UX and Design Challenges: coming soon.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>python</category>
      <category>agents</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
