<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Aman Choudhary</title>
    <description>The latest articles on Forem by Aman Choudhary (@aman_choudhary_ca1bdbc12a).</description>
    <link>https://forem.com/aman_choudhary_ca1bdbc12a</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3501089%2F0c7130d4-f82e-475f-b711-2ace4b3d26a5.png</url>
      <title>Forem: Aman Choudhary</title>
      <link>https://forem.com/aman_choudhary_ca1bdbc12a</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/aman_choudhary_ca1bdbc12a"/>
    <language>en</language>
    <item>
      <title>The Quiet Revolution at Google Cloud Next '26: Your Database Can Talk to Your AI Agent — No Bridge Required published</title>
      <dc:creator>Aman Choudhary</dc:creator>
      <pubDate>Tue, 28 Apr 2026 20:47:48 +0000</pubDate>
      <link>https://forem.com/aman_choudhary_ca1bdbc12a/the-quiet-revolution-at-google-cloud-next-26-your-database-can-talk-to-your-ai-agent-no-bridge-cho</link>
      <guid>https://forem.com/aman_choudhary_ca1bdbc12a/the-quiet-revolution-at-google-cloud-next-26-your-database-can-talk-to-your-ai-agent-no-bridge-cho</guid>
      <description>&lt;p&gt;This is a submission for the &lt;a href="https://dev.to/challenges/google-cloud-next-2026-04-22"&gt;Google Cloud NEXT Writing Challenge&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Everyone at Google Cloud Next '26 is talking about Gemini Enterprise Agent Platform. The flashy keynote demo, the "era of the agent is here" declaration, the snowboarder analyzing his own tricks with AI. I get it. It's a great story.&lt;br&gt;
But buried in the 260-announcement list is something that, for developers building real AI applications, might matter more day-to-day: Google Cloud just made it trivially easy to connect AI agents directly to your production databases via fully managed MCP servers.&lt;br&gt;
No proxy. No server to host. No auth plumbing to debug at 2 AM.&lt;br&gt;
Let me explain why this is a bigger deal than it sounds.&lt;/p&gt;

&lt;p&gt;First, The Problem This Solves&lt;/p&gt;

&lt;p&gt;If you've tried building an AI agent that operates on real data — not sample JSON, but actual operational databases — you know the pain. The agent needs to read user records, check inventory, query transaction history. And to do that, you need to:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Stand up an MCP server (or run one locally)&lt;/li&gt;
&lt;li&gt;Handle authentication — API keys? OAuth? IAM? Good luck wiring it all together&lt;/li&gt;
&lt;li&gt;Manage connection pooling so your agent doesn't accidentally nuke your database with connections&lt;/li&gt;
&lt;li&gt;Keep the whole thing running, monitored, and scaled&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Model Context Protocol (MCP), the open standard created by Anthropic, solves the interface problem beautifully — it gives AI models a standardized way to talk to tools and data sources. But the infrastructure problem was still on you.&lt;br&gt;
That's what Google Cloud just solved.&lt;/p&gt;

&lt;p&gt;What Was Announced&lt;br&gt;
At Next '26, Google Cloud announced managed, remote MCP servers that are now generally available for:&lt;/p&gt;

&lt;p&gt;: AlloyDB (PostgreSQL-compatible)&lt;br&gt;
: Cloud SQL&lt;br&gt;
: Spanner&lt;br&gt;
: Firestore&lt;br&gt;
: Bigtable&lt;/p&gt;

&lt;p&gt;And in preview for Memorystore, Database Migration Service, Datastream, Database Center, and more.&lt;br&gt;
There's also a brand new Developer Knowledge MCP server — which connects your IDE directly to Google's documentation, so your coding agent can answer questions and troubleshoot with live, relevant context rather than hallucinating from training data.&lt;br&gt;
The setup is almost shockingly simple:&lt;/p&gt;

&lt;p&gt;bash# &lt;/p&gt;

&lt;p&gt;Enable the Spanner MCP endpoint — one command&lt;br&gt;
gcloud beta services mcp enable spanner.googleapis.com --project=${PROJECT_ID}&lt;/p&gt;

&lt;p&gt;That's it. No server to deploy. The MCP endpoint is live. Then in your agent or IDE config, you point to it:&lt;/p&gt;

&lt;p&gt;json{&lt;br&gt;
  "mcpServers": {&lt;br&gt;
    "spanner": {&lt;br&gt;
      "url": "&lt;a href="https://spanner.googleapis.com/mcp" rel="noopener noreferrer"&gt;https://spanner.googleapis.com/mcp&lt;/a&gt;",&lt;br&gt;
      "authType": "oauth"&lt;br&gt;
    }&lt;br&gt;
  }&lt;br&gt;
}&lt;/p&gt;

&lt;p&gt;And now your agent can query your Spanner database in natural language — from Gemini CLI, Claude, ChatGPT, or any MCP-compliant client.&lt;/p&gt;

&lt;p&gt;Why the Security Model Actually Impresses Me&lt;/p&gt;

&lt;p&gt;My first instinct when I see "connect your AI agent to your production database" is to reach for the fire extinguisher. But Google's implementation here is thoughtful.&lt;/p&gt;

&lt;p&gt;Authentication is handled entirely through IAM — no shared API keys floating around, no connection strings hardcoded anywhere. Agents can only access the specific tables or views the IAM policy explicitly authorizes. Every query is logged through Google Cloud's standard observability stack. Audit trails are automatic.&lt;/p&gt;

&lt;p&gt;This means you can create a dedicated service account for your agent, grant it read-only access to exactly the tables it needs, and revoke it instantly if something goes wrong. That's the kind of security posture that makes it realistic to actually deploy this in production.&lt;/p&gt;

&lt;p&gt;The Spanner + MCP Angle Is Particularly Interesting&lt;/p&gt;

&lt;p&gt;Spanner's managed MCP server isn't just for SQL queries. Because Spanner now has multi-model capabilities — relational, graph, vector search, full-text — the MCP integration surfaces all of those to your agent through natural language.&lt;/p&gt;

&lt;p&gt;Imagine querying a fraud detection graph:&lt;/p&gt;

&lt;p&gt;"Find all accounts that received transfers from account 12345 within the last 48 hours, and check if any of them share a phone number with a flagged account."&lt;/p&gt;

&lt;p&gt;That's a multi-hop graph traversal combined with a relational join. With the Spanner MCP server, your agent generates the SQL+GQL automatically and executes it — no manual query writing.&lt;/p&gt;

&lt;p&gt;Google even published a codelab walking through exactly this fraud detection use case. It's worth working through if you want to see the natural-language-to-graph-query pipeline in action.&lt;/p&gt;

&lt;p&gt;The Open Source Side: MCP Toolbox 1.0&lt;/p&gt;

&lt;p&gt;Alongside the managed servers, Google also released MCP Toolbox for Databases v1.0 — the stable GA of their open-source MCP server that supports 40+ databases, with contributions from 10 vendors. This includes not just Google's databases but also Neo4j, PostgreSQL, MySQL, SQLite, and more.&lt;/p&gt;

&lt;p&gt;So the story here is two-tiered:&lt;/p&gt;

&lt;p&gt;Managed MCP ServersMCP Toolbox &lt;/p&gt;

&lt;p&gt;1.Infrastructure      Zero — Google manages it   Self-hosted&lt;/p&gt;

&lt;p&gt;2.Database support    GCP portfolio              40+ including non-GCP&lt;/p&gt;

&lt;p&gt;3.Auth --             IAM (built-in)             Configurable&lt;/p&gt;

&lt;p&gt;4.Best for --         GCP-native teams           Hybrid / multi-cloud&lt;/p&gt;

&lt;p&gt;Both are genuinely useful for different teams, and they're complementary rather than competing.&lt;/p&gt;

&lt;p&gt;My Honest Take&lt;br&gt;
The marketing around agents tends to focus on what the AI can think and decide. But agents are only as useful as what they can act on. Most enterprise value lives in operational databases — not in PDFs or chat histories. The bottleneck for practical agent deployment isn't model capability. It's data access.&lt;/p&gt;

&lt;p&gt;What Google announced here directly attacks that bottleneck.&lt;/p&gt;

&lt;p&gt;The criticism I'd level: this is still fairly tightly coupled to Google Cloud's own database portfolio for the managed tier. If your production database is RDS PostgreSQL, Aurora, or Cosmos DB, you're on the open source path — which means you're back to managing infrastructure yourself. That's a real limitation for a lot of teams.&lt;/p&gt;

&lt;p&gt;And the "natural language to SQL" reliability question is always there. For analytical queries on well-defined schemas, it works remarkably well. For complex joins across poorly documented legacy schemas? Test carefully before letting an agent loose on production.&lt;/p&gt;

&lt;p&gt;Still — the direction is right. The security model is right. And the zero-infrastructure pitch for GCP databases is genuinely compelling for teams already in the ecosystem. If your data lives in Spanner, AlloyDB, or Firestore, there's no reason not to try this today.&lt;/p&gt;

&lt;p&gt;Getting Started Right Now&lt;br&gt;
The fastest path to experimenting:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Enable Spanner API in a Google Cloud project (free trial credits work):&lt;br&gt;
bash&lt;br&gt;
gcloud services enable spanner.googleapis.com&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Enable the MCP endpoint:&lt;br&gt;
bash&lt;br&gt;
gcloud beta services mcp enable spanner.googleapis.com --project=${PROJECT_ID}&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Install Gemini CLI:&lt;br&gt;
bash&lt;br&gt;
npm install -g @google/gemini-cli&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Configure the Spanner extension and start querying your database in natural language.&lt;br&gt;
Full walkthrough: Managed MCP Servers announcement blog · Spanner MCP Codelab&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The agentic era needs agents that can actually do things. Connecting them to production data — securely, reliably, without standing up a custom server — is table stakes for that future. Google Cloud just made it significantly easier to get there.&lt;br&gt;
That's worth paying attention to, even if it didn't get the keynote slot.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>cloudnextchallenge</category>
      <category>googlecloud</category>
      <category>mcp</category>
    </item>
    <item>
      <title>This is a submission for the Built with Google Gemini: Writing Challenge*</title>
      <dc:creator>Aman Choudhary</dc:creator>
      <pubDate>Sun, 01 Mar 2026 14:00:58 +0000</pubDate>
      <link>https://forem.com/aman_choudhary_ca1bdbc12a/this-is-a-submission-for-the-built-with-google-gemini-writing-challenge-1oe0</link>
      <guid>https://forem.com/aman_choudhary_ca1bdbc12a/this-is-a-submission-for-the-built-with-google-gemini-writing-challenge-1oe0</guid>
      <description>&lt;p&gt;&lt;em&gt;This is a submission for the &lt;a href="https://dev.to/challenges/mlh/built-with-google-gemini-02-25-26"&gt;Built with Google Gemini: Writing Challenge&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Every great project starts with a spark, but the best developers know that the learning doesn't end when the deadline hits. My recent journey as a builder has been defined by two distinct projects that pushed my boundaries: a solo deep-dive into AI security, and a collaborative team build focused on developer productivity.&lt;/p&gt;

&lt;p&gt;Here is a look back at what I built, the roadblocks encountered, and where the code is taking me next.&lt;/p&gt;

&lt;p&gt;What I Built with Google Gemini&lt;br&gt;
Project 1: Hiding in Plain Sight (Multimodal Steganography)&lt;br&gt;
My first recent dive into Gemini was building a Python-based multimodal steganography application. Standard steganography conceals data within the least significant bits of an image, but if an attacker knows the algorithm, the secret is compromised. I wanted to build a system where the AI itself acts as the cryptographic key.&lt;/p&gt;

&lt;p&gt;By integrating Gemini’s multimodal capabilities, the app requires the user to pass the "cover image" to the model. Gemini analyzes the visual context—identifying objects, mood, and specific details—to generate a dynamic, context-aware key. To retrieve the hidden message, the system requires not just the altered image, but Gemini's exact interpretation of it.&lt;/p&gt;

&lt;p&gt;Project 2: Copilot CoLab (VS Code Extension)&lt;br&gt;
While AI is incredible for security, it is equally powerful for workflow orchestration. Most recently, I teamed up with Nabil and Bhumi to build Copilot CoLab, a real-time team collaboration extension for VS Code. Developers lose countless hours context-switching between their IDE, Slack, and Jira. We brought tasks, chat, and presence directly into the editor.&lt;/p&gt;

&lt;p&gt;As the frontend lead (while also contributing to the backend), I built the interface that ties these features together. We integrated Gemini to act as an embedded project manager. By pinging &lt;a class="mentioned-user" href="https://dev.to/gemini"&gt;@gemini&lt;/a&gt; in the team chat, the model can automatically generate a full Work Breakdown Structure (WBS) for a new feature or perform AI-powered bulk task assignments to team members based on the repository's context.&lt;/p&gt;

&lt;p&gt;Demo&lt;br&gt;
You can check out the source code for both projects here. (Tip: I highly recommend embedding a quick 30-second Loom video or a few high-quality screenshots of the CoLab UI and the Steganography terminal output right here before you publish!)&lt;/p&gt;

&lt;p&gt;Multimodal Steganography: &lt;a href="https://github.com/Aman0choudhary/Project-1" rel="noopener noreferrer"&gt;https://github.com/Aman0choudhary/Project-1&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Copilot CoLab: &lt;a href="https://github.com/n4bi10p/copilot-colab" rel="noopener noreferrer"&gt;https://github.com/n4bi10p/copilot-colab&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;What I Learned&lt;br&gt;
Between juggling my responsibilities as a college Cloud Lead and pushing through late-night study sessions for OS and PPS exams, these projects forced a massive evolution in how I write software.&lt;/p&gt;

&lt;p&gt;Technical Breadth: The steganography app required a deep dive into Python's byte-level file manipulation. Copilot CoLab was a completely different beast: it required mastering the VS Code Webview API, bridging frontend states with extension host commands, and keeping everything synced in real-time using Supabase.&lt;/p&gt;

&lt;p&gt;The Shift from Solo to Lead: Leading the frontend for a team meant I couldn't just build in a silo. I had to clearly communicate UI constraints to the backend, document my logic, and iterate based on Nabil and Bhumi's feedback. It taught me that code readability and clear communication are just as important as the logic itself.&lt;/p&gt;

&lt;p&gt;The Macro vs. Micro Perspective: Building the steganography app required thinking small—literally down to the least significant bit of a single pixel. Building Copilot CoLab required thinking big—about human behavior and how teams actually communicate. Great architecture requires respecting both ends of that spectrum.&lt;/p&gt;

&lt;p&gt;Google Gemini Feedback&lt;br&gt;
The Good:&lt;br&gt;
The Google AI Studio interface is phenomenal for rapid prototyping. Being able to drag and drop images and tweak my prompts for the steganography app before writing a single line of Python saved me hours of API debugging. For Copilot CoLab, the speed of the gemini-1.5-flash model was a massive win; it parsed project contexts and assigned tasks incredibly fast, making the &lt;a class="mentioned-user" href="https://dev.to/gemini"&gt;@gemini&lt;/a&gt; chat feel like a truly real-time teammate.&lt;/p&gt;

&lt;p&gt;The Friction (The Bad and the Ugly):&lt;br&gt;
The biggest hurdle was forcing a generative model to act deterministically. Getting Gemini to output the exact same key format every single time for the security app—or perfectly formatted JSON for Copilot CoLab's bulk task assignment—required heavy prompt engineering. In the early stages, the model would sometimes over-explain (e.g., adding conversational fluff or wrapping outputs in markdown blocks), which completely broke our parsers. We had to learn how to aggressively constrain the prompts and implement strict JSON parsing on our end to filter out the noise.&lt;/p&gt;

&lt;p&gt;Looking Forward&lt;br&gt;
Working on these tools showed me how powerful AI can be when applied to real-time human connection and secure verification. Currently, I'm conceptualizing a hyperlocal social discovery mobile app for students and professionals in Pune, focusing on matching people based on shared interests. I am already brainstorming how to implement Gemini into the backend of this new app—perhaps using multimodal logic to verify student IDs or dynamically match users based on their portfolios.&lt;/p&gt;

&lt;p&gt;The hackathons might be over, but the builder's momentum is just getting started.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>geminireflections</category>
      <category>gemini</category>
    </item>
    <item>
      <title>Steganography App (Artful whisper)</title>
      <dc:creator>Aman Choudhary</dc:creator>
      <pubDate>Sun, 14 Sep 2025 11:11:38 +0000</pubDate>
      <link>https://forem.com/aman_choudhary_ca1bdbc12a/steganography-app-artful-whisper-53hd</link>
      <guid>https://forem.com/aman_choudhary_ca1bdbc12a/steganography-app-artful-whisper-53hd</guid>
      <description>&lt;p&gt;&lt;em&gt;This is a submission for the &lt;a href="https://dev.to/challenges/google-ai-studio-2025-09-03"&gt;Google AI Studio Multimodal Challenge&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Built
&lt;/h2&gt;
&lt;h2&gt;
  
  
  Demo
&lt;/h2&gt;

&lt;h2&gt;
  
  
  Multimodal features
&lt;/h2&gt;

&lt;p&gt;ArtfulWhisper is fundamentally multimodal, creating a seamless flow between text and image data to deliver its unique functionality.&lt;/p&gt;

&lt;p&gt;Text-to-Image Generation: The primary multimodal feature is taking a user's text prompt and transforming it into a rich, complex image using the Imagen 3 model. This is the creative heart of the app.&lt;/p&gt;

&lt;p&gt;2.Fusing Text within an Image:The application then takes a second text input (the secret message) and algorithmically embeds it directly into the pixel data of the newly generated image. This goes beyond simple input-output; it's about fusing one modality (text) invisibly inside another (image).&lt;/p&gt;

&lt;p&gt;The user experience is about power. It enhances it by giving the user a sense of control and secrecy that a simple image-and-text app could never provide. The magic isn't in seeing the two modalities work together; it's in knowing that one is invisibly controlling the other. It's a demonstration of how multimodal AI can be used for more than just cute chatbots and summary tools. It can be used to keep secrets.&lt;/p&gt;

</description>
      <category>devchallenge</category>
      <category>googleaichallenge</category>
      <category>ai</category>
      <category>gemini</category>
    </item>
  </channel>
</rss>
