<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Satyajit Mishra</title>
    <description>The latest articles on Forem by Satyajit Mishra (@satyajitmishra).</description>
    <link>https://forem.com/satyajitmishra</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3687676%2F1826846b-daeb-47e7-9fab-625192f5712c.gif</url>
      <title>Forem: Satyajit Mishra</title>
      <link>https://forem.com/satyajitmishra</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/satyajitmishra"/>
    <language>en</language>
    <item>
      <title>🧠Perceptron vs XOR: Why One Math Problem Changed AI Forever</title>
      <dc:creator>Satyajit Mishra</dc:creator>
      <pubDate>Sun, 04 Jan 2026 06:54:46 +0000</pubDate>
      <link>https://forem.com/satyajitmishra/perceptron-vs-xor-why-one-math-problem-changed-ai-forever-53o3</link>
      <guid>https://forem.com/satyajitmishra/perceptron-vs-xor-why-one-math-problem-changed-ai-forever-53o3</guid>
      <description>&lt;p&gt;Perceptron vs XOR: Why One Math Problem Changed AI Forever&lt;br&gt;
Cover Image: [Use an AI-generated or stock image showing neural networks/AI visualization]&lt;br&gt;
Reading time: 8 min read&lt;br&gt;
Tags: #ai #machinelearning #deeplearning #computerscience #beginners #technology #learning&lt;/p&gt;

&lt;p&gt;The Question That Started It All&lt;br&gt;
Imagine you're a researcher in 1969.&lt;br&gt;
You've just built something incredible: a machine that can learn.&lt;br&gt;
It's called the Perceptron, and it's the future of AI.&lt;br&gt;
It can solve problems like:&lt;br&gt;
✅ AND logic&lt;br&gt;
✅ OR logic&lt;br&gt;
✅ Complex pattern recognition&lt;br&gt;
The world is buzzing. Newspapers declare: "Machines Can Think!"&lt;br&gt;
Funding flows in. Scientists are euphoric.&lt;br&gt;
And then someone asks a simple question:&lt;/p&gt;

&lt;p&gt;"Can your Perceptron solve XOR?"&lt;/p&gt;

&lt;p&gt;Everything falls apart.&lt;/p&gt;

&lt;p&gt;What is a Perceptron? (The Simple Version)&lt;br&gt;
Before we understand why XOR broke everything, let's understand the Perceptron.&lt;br&gt;
The Perceptron is how your brain actually makes decisions.&lt;br&gt;
Right now, your brain is doing this:&lt;br&gt;
You're deciding: "Should I keep reading?"&lt;br&gt;
Your brain checks:&lt;/p&gt;

&lt;p&gt;"Is this interesting?" (Input A)&lt;br&gt;
"Do I have time?" (Input B)&lt;br&gt;
"Will I learn something?" (Input C)&lt;/p&gt;

&lt;p&gt;Then it weighs these inputs and makes a YES or NO decision.&lt;br&gt;
The Perceptron does the exact same thing:&lt;br&gt;
Step 1: Take inputs&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Input A: Is it raining? (1 = yes, 0 = no)
Input B: Do I have work? (1 = yes, 0 = no)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Step 2: Assign importance (weights)&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Rain matters 2x more than work

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Step 3: Add everything together and decide&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Total score = (Rain × 2) + (Work × 1)

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;If total &amp;gt; threshold → Output: YES (1)&lt;br&gt;
If total ≤ threshold → Output: NO (0)&lt;br&gt;
That's it. No magic. No complexity. Pure linear logic.&lt;br&gt;
And it worked brilliantly... until it didn't.&lt;/p&gt;

&lt;p&gt;The Problems It Could Solve (AND &amp;amp; OR)&lt;br&gt;
In the 1950s, researchers discovered the Perceptron could solve simple logic problems.&lt;br&gt;
AND Logic&lt;br&gt;
"Output YES only if BOTH inputs are true"&lt;br&gt;
Input AInput BOutput000010100111&lt;br&gt;
Why the Perceptron nailed it: You can draw one straight line separating the 1s from the 0s.&lt;br&gt;
OR Logic&lt;br&gt;
"Output YES if AT LEAST ONE input is true"&lt;br&gt;
Input A &lt;br&gt;
Input B &lt;br&gt;
Output 000011101111&lt;br&gt;
Again, one straight line works perfectly.&lt;br&gt;
Both problems were linearly separable—the Perceptron's entire world.&lt;br&gt;
Researchers were drunk on success.&lt;br&gt;
They believed AI had no limits.&lt;/p&gt;

&lt;p&gt;The Problem That Changed Everything: XOR&lt;br&gt;
Then came XOR (Exclusive OR).&lt;br&gt;
It looks simple. Almost too simple.&lt;br&gt;
XOR Logic&lt;br&gt;
"Output YES only when inputs are DIFFERENT"&lt;br&gt;
Input AInput BOutput000011101110&lt;br&gt;
Harmless, right?&lt;br&gt;
Dead wrong.&lt;br&gt;
Researchers tried to teach the Perceptron XOR.&lt;br&gt;
They tried for weeks. Months. With different methods. Different weights. Everything.&lt;br&gt;
Nothing worked.&lt;br&gt;
The Perceptron simply could not learn XOR.&lt;br&gt;
And nobody understood why.&lt;/p&gt;

&lt;p&gt;Why XOR Broke the Perceptron (The Geometry Secret)&lt;br&gt;
Here's the shocking truth: XOR isn't complicated mathematically.&lt;br&gt;
The problem was geometric.&lt;br&gt;
Imagine plotting the four XOR results on a graph:&lt;br&gt;
Input B (vertical axis)&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;  1 |  1(0,1)   0(1,1)
       |   \       /
     0 |    0(0,0)-1(1,0)
       └─────────────────── Input A
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Look at this pattern:&lt;/p&gt;

&lt;p&gt;The two 1s are in the middle (0,1) and (1,0)&lt;br&gt;
The two 0s are on the outside (0,0) and (1,1)&lt;/p&gt;

&lt;p&gt;Now try to draw one straight line that separates all the 1s from all the 0s.&lt;br&gt;
You can't.&lt;br&gt;
No matter how you angle it, a single straight line will always misclassify at least one point.&lt;br&gt;
Here's why:&lt;br&gt;
The Perceptron only thinks in straight lines.&lt;br&gt;
It says: "Everything above this line is YES. Everything below is NO."&lt;br&gt;
But XOR's solution isn't a line—it's a curved boundary or multiple lines.&lt;br&gt;
Think of it like this:&lt;/p&gt;

&lt;p&gt;You're someone who can only draw straight lines. You're asked to paint the Mona Lisa.&lt;br&gt;
Impossible, right?&lt;br&gt;
That's the Perceptron vs XOR.&lt;/p&gt;

&lt;p&gt;The Biggest Mistake in AI History&lt;br&gt;
Here's where things got dark.&lt;br&gt;
Researchers saw the problem and drew the worst possible conclusion.&lt;br&gt;
Instead of thinking:&lt;/p&gt;

&lt;p&gt;"The Perceptron needs to evolve. Let's find a better approach."&lt;/p&gt;

&lt;p&gt;They thought:&lt;/p&gt;

&lt;p&gt;"If Perceptrons can't solve XOR, maybe AI itself is impossible."&lt;/p&gt;

&lt;p&gt;And they told everyone.&lt;br&gt;
Two MIT researchers, Marvin Minsky and Seymour Papert, published a book called "Perceptrons" (1969).&lt;br&gt;
In it, they outlined the XOR problem and suggested that single-layer neural networks had fundamental, unfixable limitations.&lt;br&gt;
What happened next was devastating:&lt;/p&gt;

&lt;p&gt;Funding dried up 💸&lt;br&gt;
Research slowed ❄️&lt;br&gt;
Scientists abandoned neural networks&lt;br&gt;
The field froze for over a decade&lt;/p&gt;

&lt;p&gt;This dark period became known as the AI Winter.&lt;br&gt;
For years, artificial intelligence was considered a dead end.&lt;/p&gt;

&lt;p&gt;The Truth That Everyone Missed&lt;br&gt;
Here's the irony: The Perceptron wasn't broken. It was just incomplete.&lt;br&gt;
The researchers who gave up missed one crucial insight:&lt;/p&gt;

&lt;p&gt;Humans don't solve every problem with one way of thinking.&lt;/p&gt;

&lt;p&gt;When you encounter something complex, you don't think harder the same way.&lt;br&gt;
You break it down into layers.&lt;br&gt;
You combine simple ideas into bigger ones.&lt;br&gt;
You add depth.&lt;br&gt;
What if machines could do the same?&lt;/p&gt;

&lt;p&gt;The Breakthrough: Adding Another Layer&lt;br&gt;
In the 1980s, someone had a simple but revolutionary idea:&lt;/p&gt;

&lt;p&gt;"What if we stack Perceptrons together?"&lt;/p&gt;

&lt;p&gt;Instead of one layer making a decision, create:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;Layer 1 (Input layer): learns simple patterns&lt;br&gt;
Layer 2 (Hidden layer): combines those patterns&lt;br&gt;
Layer 3 (Output layer): makes the final decision&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;This created something new: a Multi-Layer Perceptron.&lt;br&gt;
And here's what happened:&lt;br&gt;
With multiple layers, the system could now:&lt;/p&gt;

&lt;p&gt;Learn curves, not just straight lines&lt;br&gt;
Combine simple patterns into complex ones&lt;br&gt;
Finally solve XOR&lt;/p&gt;

&lt;p&gt;Let's test it:&lt;br&gt;
Layer 1: Transforms the input space&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Node A: detects "is Input A different from Input B?"&lt;/li&gt;
&lt;li&gt;Node B: detects "are both inputs the same?"&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Layer 2: Combines these patterns&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;If A XOR B (they differ) → Output 1&lt;/li&gt;
&lt;li&gt;Otherwise → Output 0&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Result: ✅ XOR SOLVED&lt;br&gt;
It worked.&lt;br&gt;
And just like that, something magical was born:&lt;br&gt;
✨ Deep Learning&lt;/p&gt;

&lt;p&gt;Why This Matters More Than You Think&lt;br&gt;
Every AI system you use today exists because of this lesson.&lt;/p&gt;

&lt;p&gt;Your phone's face recognition? Deep Learning.&lt;br&gt;
Netflix recommendations? Deep Learning.&lt;br&gt;
ChatGPT, Claude, and every modern language model? Deep Learning.&lt;br&gt;
Google Translate? Deep Learning.&lt;br&gt;
Autonomous vehicles? Deep Learning.&lt;/p&gt;

&lt;p&gt;All of them use layers. Many, many layers.&lt;br&gt;
Modern AI models use 100+ layers, sometimes 1,000+.&lt;br&gt;
And it all started because someone asked:&lt;/p&gt;

&lt;p&gt;"What if the answer isn't to think harder in one way, but to think DEEPER in multiple ways?"&lt;/p&gt;

&lt;p&gt;The Real Lesson (Beyond Technology)&lt;br&gt;
This story teaches something bigger than AI.&lt;br&gt;
It's about how we respond to limitations.&lt;br&gt;
The Wrong Response (What Almost Happened):&lt;/p&gt;

&lt;p&gt;Hit a problem → Assume it's impossible&lt;br&gt;
Give up → Accept defeat&lt;br&gt;
Move on → Miss the breakthrough&lt;/p&gt;

&lt;p&gt;The Right Response (What Eventually Happened):&lt;/p&gt;

&lt;p&gt;Hit a problem → Ask "What am I missing?"&lt;br&gt;
Try a different approach → Experiment relentlessly&lt;br&gt;
Keep learning → Find the breakthrough&lt;br&gt;
Build on it → Change the world&lt;/p&gt;

&lt;p&gt;The difference between these two paths is everything.&lt;br&gt;
In your own life:&lt;br&gt;
When something doesn't work:&lt;/p&gt;

&lt;p&gt;You could see it as a wall, OR&lt;br&gt;
You could see it as an invitation to level up&lt;/p&gt;

&lt;p&gt;When the Perceptron failed at XOR, it wasn't a failure of AI.&lt;br&gt;
It was a signal that AI needed to grow deeper.&lt;/p&gt;

&lt;p&gt;XOR: The Problem That Saved AI&lt;br&gt;
Here's the beautiful irony:&lt;/p&gt;

&lt;p&gt;XOR didn't destroy AI. It accidentally created it.&lt;/p&gt;

&lt;p&gt;If the Perceptron had worked for everything, AI would have hit a wall eventually. Much later. Much harder.&lt;br&gt;
Instead, XOR forced a breakthrough early.&lt;br&gt;
It forced researchers to ask better questions.&lt;br&gt;
It forced the field to evolve.&lt;br&gt;
And by evolving, it became something magnificent.&lt;/p&gt;

&lt;p&gt;The Timeline: From Failure to Revolution&lt;br&gt;
1943: McCulloch-Pitts neuron invented&lt;br&gt;
1958: Rosenblatt invents the Perceptron&lt;br&gt;
1960s: Perceptron solves AND, OR logic&lt;br&gt;
1969: Minsky &amp;amp; Papert reveal XOR limitation&lt;br&gt;
1970-1980: AI Winter (mostly abandoned)&lt;br&gt;
1986: Backpropagation algorithm rediscovered&lt;br&gt;
       (by Rumelhart, Hinton, Williams)&lt;br&gt;
1987-1990: Multi-layer networks proven to solve XOR and beyond&lt;br&gt;
2000s-2010s: Deep Learning revolution (ImageNet, AlexNet, etc.)&lt;br&gt;
2012-Present: Deep Learning dominates AI (GPT models, &lt;br&gt;
              computer vision, etc.)&lt;br&gt;
One tiny logic puzzle led to a 50+ year journey that changed the world.&lt;/p&gt;

&lt;p&gt;The Deeper Meaning&lt;br&gt;
XOR teaches us something profound about growth.&lt;/p&gt;

&lt;p&gt;Limitations aren't failures. They're invitations.&lt;/p&gt;

&lt;p&gt;The Perceptron's limitation wasn't a bug—it was a feature.&lt;br&gt;
It was a compass pointing toward the future.&lt;br&gt;
When you can't solve a problem the way you've been thinking, that's when real innovation happens.&lt;br&gt;
That's when you discover you've been thinking too shallow.&lt;br&gt;
That's when you learn to go deeper.&lt;/p&gt;

&lt;p&gt;Final Thoughts&lt;br&gt;
In 2025, we live in an age of AI everywhere.&lt;br&gt;
But we almost didn't.&lt;br&gt;
We almost gave up because of one simple logic problem: XOR.&lt;br&gt;
We almost decided that machines couldn't learn.&lt;br&gt;
We almost stopped trying.&lt;br&gt;
But someone—many someones—kept asking:&lt;/p&gt;

&lt;p&gt;"What if there's a better way?"&lt;/p&gt;

&lt;p&gt;And there was.&lt;br&gt;
There always is.&lt;/p&gt;

&lt;p&gt;The Lesson for You&lt;br&gt;
Whatever you're facing right now:&lt;br&gt;
If something isn't working, it's not a dead end.&lt;br&gt;
It's a signal that you need to think deeper.&lt;br&gt;
Like the Perceptron, sometimes you can't solve your problem with one strategy.&lt;/p&gt;

&lt;p&gt;You need to add layers&lt;br&gt;
You need to combine approaches&lt;br&gt;
You need to go deeper&lt;/p&gt;

&lt;p&gt;And when you do, you'll find that your greatest limitations were actually your greatest teachers.&lt;br&gt;
Just like XOR was for AI.&lt;/p&gt;

&lt;p&gt;XOR didn't destroy artificial intelligence. It taught it how to think.&lt;br&gt;
What's your XOR?&lt;br&gt;
What problem are you avoiding because you think it's impossible?&lt;br&gt;
Maybe it's just asking you to go deeper. 🚀&lt;/p&gt;

&lt;p&gt;Share Your Thoughts&lt;br&gt;
💬 Which limitation taught you the most?&lt;br&gt;
Drop a comment below. I'd love to hear your story of turning a problem into a breakthrough.&lt;/p&gt;

&lt;p&gt;Subscribe for more AI explained clearly. ✨&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>ai</category>
      <category>beginners</category>
    </item>
    <item>
      <title>I Learned Generative AI Basics Today — Here’s the Simple Truth No One Explained Clearly</title>
      <dc:creator>Satyajit Mishra</dc:creator>
      <pubDate>Wed, 31 Dec 2025 16:13:30 +0000</pubDate>
      <link>https://forem.com/satyajitmishra/i-learned-generative-ai-basics-today-heres-the-simple-truth-no-one-explained-clearly-1ckf</link>
      <guid>https://forem.com/satyajitmishra/i-learned-generative-ai-basics-today-heres-the-simple-truth-no-one-explained-clearly-1ckf</guid>
      <description>&lt;p&gt;Today wasn’t about building the next ChatGPT.&lt;br&gt;
It was about &lt;strong&gt;finally understanding what Generative AI actually&lt;/strong&gt; is — without buzzwords, without hype, without confusion.&lt;/p&gt;

&lt;p&gt;And honestly?&lt;br&gt;
It was simpler than I expected.&lt;/p&gt;

&lt;p&gt;What I Thought Generative AI Was&lt;/p&gt;

&lt;p&gt;Before today, my understanding was messy.&lt;/p&gt;

&lt;p&gt;I thought:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It’s some magical AI that writes perfect code&lt;/li&gt;
&lt;li&gt;It replaces developers&lt;/li&gt;
&lt;li&gt;You need PhD-level math to understand it
Most of us think this way because we only see finished AI products, not the fundamentals behind them.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;What Generative AI Actually Is (In Simple Terms)&lt;/p&gt;

&lt;p&gt;Generative AI is not magic.&lt;/p&gt;

&lt;p&gt;At its core, it does one thing really well:&lt;br&gt;
It learns patterns from data and generates new content based on those patterns.&lt;/p&gt;

&lt;p&gt;That’s it.&lt;br&gt;
Depending on the model, that content can be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Text&lt;/li&gt;
&lt;li&gt;Images&lt;/li&gt;
&lt;li&gt;Code&lt;/li&gt;
&lt;li&gt;Music&lt;/li&gt;
&lt;li&gt;Summaries&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Core Concepts I Learned Today&lt;br&gt;
&lt;strong&gt;1. Data Is Everything&lt;/strong&gt;&lt;br&gt;
Generative AI doesn’t think.&lt;br&gt;
It learns from huge amounts of data.&lt;/p&gt;

&lt;p&gt;Bad data leads to bad output.&lt;br&gt;
Good data leads to useful output.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Models Learn Patterns, Not Facts&lt;/strong&gt;&lt;br&gt;
This was a big realization.&lt;/p&gt;

&lt;p&gt;The model doesn’t know information.&lt;br&gt;
It predicts what comes next based on probability.&lt;/p&gt;

&lt;p&gt;For example:&lt;br&gt;
“The sky is ___” → blue&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Training vs Inference (Very Important)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Training:&lt;/strong&gt;&lt;/em&gt; The model learns from data (heavy and expensive)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Inference:&lt;/em&gt;&lt;/strong&gt; The model generates output from prompts (what we usually use)&lt;/p&gt;

&lt;p&gt;As developers, we mostly interact with inference, not training.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Prompts Matter More Than I Expected&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Same model.&lt;br&gt;
Different prompt.&lt;br&gt;
Completely different result.&lt;/p&gt;

&lt;p&gt;Prompting is basically clear communication with AI.&lt;/p&gt;

&lt;p&gt;Vague prompt → weak output&lt;br&gt;
Clear prompt → surprisingly good output&lt;/p&gt;

&lt;p&gt;What Generative AI Is NOT&lt;/p&gt;

&lt;p&gt;Let’s clear some common myths:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;It doesn’t understand emotions&lt;/li&gt;
&lt;li&gt;It doesn’t think like humans&lt;/li&gt;
&lt;li&gt;It doesn’t replace learning fundamentals&lt;/li&gt;
&lt;li&gt;It’s not always correct&lt;/li&gt;
&lt;li&gt;It’s powerful — but still just a tool.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Why This Matters for Developers and Students&lt;/p&gt;

&lt;p&gt;One thing became very clear to me today:&lt;/p&gt;

&lt;p&gt;Generative AI will not replace developers.&lt;br&gt;
Developers who understand AI will replace those who don’t.&lt;/p&gt;

&lt;p&gt;Learning AI basics helps you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Work faster&lt;/li&gt;
&lt;li&gt;Learn smarter&lt;/li&gt;
&lt;li&gt;Solve problems better&lt;/li&gt;
&lt;li&gt;Stay relevant in the future&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You don’t need to master everything — just understand how it works.&lt;/p&gt;

&lt;p&gt;My Biggest Takeaway&lt;/p&gt;

&lt;p&gt;I stopped being intimidated by AI.&lt;/p&gt;

&lt;p&gt;Once you remove the hype, Generative AI becomes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Logical&lt;/li&gt;
&lt;li&gt;Learnable&lt;/li&gt;
&lt;li&gt;Extremely useful&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And most importantly — approachable.&lt;/p&gt;

&lt;p&gt;If You’re a Beginner, Start Like This&lt;/p&gt;

&lt;p&gt;Learn concepts before tools&lt;/p&gt;

&lt;p&gt;Don’t chase trends, chase clarity&lt;/p&gt;

&lt;p&gt;Use AI to learn, not to skip learning&lt;/p&gt;

&lt;p&gt;Experiment with prompts&lt;/p&gt;

&lt;p&gt;Stay curious, not scared&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>beginners</category>
      <category>productivity</category>
    </item>
  </channel>
</rss>
