<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: John Mwendwa</title>
    <description>The latest articles on Forem by John Mwendwa (@jmwendwa).</description>
    <link>https://forem.com/jmwendwa</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1868345%2Fc33547bc-a92a-4c40-a40d-cbcaf7c8d333.jpg</url>
      <title>Forem: John Mwendwa</title>
      <link>https://forem.com/jmwendwa</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/jmwendwa"/>
    <language>en</language>
    <item>
      <title>A Deep Dive Into Boston’s Airbnb Performance</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Thu, 08 Jan 2026 15:47:19 +0000</pubDate>
      <link>https://forem.com/jmwendwa/a-deep-dive-into-bostons-airbnb-performance-4fmk</link>
      <guid>https://forem.com/jmwendwa/a-deep-dive-into-bostons-airbnb-performance-4fmk</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;The short term rental market is constantly shifting and platforms like Airbnb need more than descriptive statistics to stay ahead.For this project, i took a deep dive into Boston’s Airbnb listings to answer three high-impact business questions:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; What makes some hosts incredibly successful while others struggle?&lt;/li&gt;
&lt;li&gt; Which Boston neighborhoods show strong demand and which ones are oversaturated?&lt;/li&gt;
&lt;li&gt; What drives exceptional guest experiences, and how can Airbnb detect quality risks early?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;My goal wasn’t just to analyze data, but to translate it into real market intelligence that can guide growth strategy, host enablement, and guest experience improvement.&lt;/p&gt;

&lt;h2&gt;
  
  
  The End-to-End Data Pipeline
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Data Preparation in Python
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Loaded the raw Airbnb Boston dataset&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Checked data structure, completeness, and quality&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cleaned missing values, removed duplicates, and treated outliers&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Performed exploratory data analysis (EDA)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Analyzed pricing patterns, room types, correlations, and neighborhood      distribution&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Data Modeling &amp;amp; Analysis in PostgreSQL
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Loaded the cleaned dataset into PostgreSQL using psycopg2&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Designed relational schemas for listings, hosts, and neighborhood data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Used SQL and CTEs to calculate:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Host success metrics&lt;/li&gt;
&lt;li&gt;Neighborhood market health&lt;/li&gt;
&lt;li&gt;Oversaturation indicators&lt;/li&gt;
&lt;li&gt;Guest engagement via reviews-per-month&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Dashboard Creation in Power BI
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Connected PostgreSQL to Power BI&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Built DAX-based KPIs for host performance, market health, and guest experience&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Designed interactive dashboards with slicers and buttons&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Visualized insights for hosts, neighborhoods, and reviews&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Insights
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. High performing hosts share common behaviors
&lt;/h3&gt;

&lt;p&gt;They set competitive prices, maintain high-quality listings, and benefit strongly from central neighborhoods like Back Bay, Beacon Hill, and the West End.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Not all neighborhoods are equal
&lt;/h3&gt;

&lt;p&gt;Areas like Longwood Medical Area, Bay Village, and Back Bay show strong demand.Others Dorchester, Roxbury, Jamaica Plain, Fenway show signs of oversaturation.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Guest experience can be predicted
&lt;/h3&gt;

&lt;p&gt;Listings with high review velocity (more reviews per month) consistently reflect better guest satisfaction and listing quality.&lt;/p&gt;

&lt;h2&gt;
  
  
  Recommendations
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Help hosts succeed
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Provide better pricing guidance to match market demand&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Support new hosts with onboarding and listing improvement tips&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Boost visibility for small or new hosts&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Manage neighborhood growth strategically
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Focus expansion in high-demand areas&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Be cautious adding new listings in saturated neighborhoods&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Encourage improvements (quality, pricing) for hosts in slow demand regions&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Protect guest experience
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Track listings with declining reviews-per-month&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Benchmark each listing against similar ones in the same neighborhood&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Flag risk listings early and intervene before quality drops affect platform reputation&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>analytics</category>
      <category>datascience</category>
      <category>portfolio</category>
      <category>python</category>
    </item>
    <item>
      <title>SQL and I Had Beef 😂So I Built a Trigger</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Mon, 05 May 2025 10:53:24 +0000</pubDate>
      <link>https://forem.com/jmwendwa/sql-and-i-had-beef-so-i-built-a-trigger-26m9</link>
      <guid>https://forem.com/jmwendwa/sql-and-i-had-beef-so-i-built-a-trigger-26m9</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Let me be honest SQL has been giving me a hard time for a while. So, I was getting stuck even with the basics. So I decided to stop running away and build a real project from scratch: a student enrollment system. &lt;/p&gt;

&lt;p&gt;I wanted something simple but useful, where I could understand how real systems are built with SQL,,including &lt;em&gt;relationships&lt;/em&gt;, &lt;em&gt;foreign keys&lt;/em&gt;, and &lt;em&gt;triggers&lt;/em&gt;. This is how it went down.&lt;/p&gt;




&lt;h2&gt;
  
  
  What the System Does
&lt;/h2&gt;

&lt;p&gt;This project is all about managing a small school system. It handles:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Storing student info (name, email, date of birth)&lt;/li&gt;
&lt;li&gt;Keeping track of instructors&lt;/li&gt;
&lt;li&gt;Linking courses to instructors&lt;/li&gt;
&lt;li&gt;Enrolling students to courses&lt;/li&gt;
&lt;li&gt;Logging every enrollment using a &lt;em&gt;trigger&lt;/em&gt;
&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Tables I Created
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Table&lt;/th&gt;
&lt;th&gt;Description&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Students&lt;/td&gt;
&lt;td&gt;Stores student details like name and date of birth&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Instructors&lt;/td&gt;
&lt;td&gt;Stores instructor info&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Courses&lt;/td&gt;
&lt;td&gt;Each course is taught by one instructor&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Enrollments&lt;/td&gt;
&lt;td&gt;Shows which student is taking which course&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Enrollment_Log&lt;/td&gt;
&lt;td&gt;Automatically logs new enrollments using a trigger&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;




&lt;h2&gt;
  
  
  SQL Code I Used
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Students Table
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TABLE Students (
    student_id INT PRIMARY KEY,
    first_name VARCHAR(50),
    last_name VARCHAR(50),
    email VARCHAR(100) UNIQUE,
    date_of_birth DATE
);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  2. Instructors Table
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TABLE Instructors (
    instructor_id INT PRIMARY KEY,
    first_name VARCHAR(50),
    last_name VARCHAR(50),
    email VARCHAR(100) UNIQUE
);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  3. Courses Table
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TABLE Courses (
    course_id INT PRIMARY KEY,
    course_name VARCHAR(100),
    instructor_id INT,
    FOREIGN KEY (instructor_id) REFERENCES Instructors(instructor_id)
);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  4. Enrollments Table
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TABLE Enrollments (
    enrollment_id INT PRIMARY KEY,
    student_id INT,
    course_id INT,
    grade CHAR(2),
    FOREIGN KEY (student_id) REFERENCES Students(student_id),
    FOREIGN KEY (course_id) REFERENCES Courses(course_id)
);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  5. Enrollment Log Table
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TABLE Enrollment_Log (
    log_id INT PRIMARY KEY AUTO_INCREMENT,
    enrollment_id INT,
    log_time TIMESTAMP DEFAULT CURRENT_TIMESTAMP
);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  6. Trigger to Log Enrollments
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE TRIGGER LogNewEnrollment
AFTER INSERT ON Enrollments
FOR EACH ROW
INSERT INTO Enrollment_Log (enrollment_id)
VALUES (NEW.enrollment_id);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Sample Data I Added
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1.Instructors
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;INSERT INTO Instructors (instructor_id, first_name, last_name, email) VALUES 
(1, 'Thomas', 'Ndegwa', 'thomas.ndegwa@example.com'),
(2, 'Lilian', 'Achieng', 'lilian.a@example.com'),
(3, 'George', 'Kariuki', 'george.kariuki@example.com');
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  2. Students
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;INSERT INTO Students (student_id, first_name, last_name, email, date_of_birth) VALUES 
(101, 'Alice', 'Mwende', 'alice.mwende@example.com', '2002-04-12'),
(102, 'Brian', 'Otieno', 'brian.otieno@example.com', '2001-08-05');
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  3. Courses
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;INSERT INTO Courses (course_id, course_name, instructor_id) VALUES 
(201, 'Database Systems', 1),
(202, 'Python Programming', 2),
(203, 'Data Analytics', 3);
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h3&gt;
  
  
  4.Enrollments
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;INSERT INTO Enrollments (enrollment_id, student_id, course_id, grade) VALUES 
(301, 101, 201, 'A'),
(302, 102, 202, 'B');
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  What I Learned
&lt;/h2&gt;

&lt;p&gt;Through this project, I’ve learned:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;How to design related tables using foreign keys&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;How to keep data clean and connected&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;How to use triggers to automate tasks (like logging)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The importance of thinking through the whole structure before writing code&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This is way better than just memorizing;&lt;br&gt;
&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;SELECT * FROM table;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  What I Might Add Later
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Add user logins for students and instructors&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Store attendance and class schedules&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Build a simple dashboard with Power BI &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;This was a small project, but it really helped me understand SQL better. I now see why people say the database is the heart of any system.&lt;br&gt;
If you're learning SQL, I encourage you to try something small like this,it makes everything more clear.&lt;/p&gt;

</description>
      <category>programming</category>
      <category>beginners</category>
      <category>datascience</category>
      <category>sql</category>
    </item>
    <item>
      <title>From Chaos to Clarity:Excel Made Easy.</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Sat, 26 Apr 2025 16:52:19 +0000</pubDate>
      <link>https://forem.com/jmwendwa/from-chaos-to-clarityexcel-made-easy-bhg</link>
      <guid>https://forem.com/jmwendwa/from-chaos-to-clarityexcel-made-easy-bhg</guid>
      <description>&lt;p&gt;Today, Excel is still one of the best tools we use to understand numbers and make good business decisions.&lt;br&gt;
Whether you are checking sales, looking for patterns, or planning ahead, Excel can really help you a lot.&lt;br&gt;
Let me take you through a project iv done recently on sales data;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Data Transformation: Preparing for Analysis&lt;/strong&gt;&lt;br&gt;
The first and most crucial step in any data project is cleaning and preparing the data. In Excel:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Format Dates Properly: Use Excel’s date formatting tools to ensure that fields that have dates data are recognized as real dates.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Ensure Numeric Fields Are Clean: Highlight the sales_value column tap home tab,then number group then number(set format to number).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Handle Missing Data: Highlight whole data,use find &amp;amp; select tab,go to special,then tap blanks.If u find them either delete or fill them using the mean of the data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Organize the Data: Structure the table clearly, with headers and consistent data types, making it easy to perform calculations later.&lt;br&gt;
A clean dataset is the foundation for reliable insights.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;2. Statistical Analysis: Understanding the Numbers&lt;/strong&gt;&lt;br&gt;
Once your data is ready, you can start by finding out the important numbers:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Total Sales Value: Add up all the sales to see how much money the company made.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Average Sales Value: Divide the total sales by the number of sales to find out the average amount per sale.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Total Quantity Sold: Add up the number of items sold.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Average Quantity Sold: Find out the average number of items sold per sale.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;3. Data Analysis: Identifying Patterns and Trends&lt;/strong&gt;&lt;br&gt;
Beyond basic stats, Excel’s PivotTables allow you to break down performance&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Sales by Region: Find out which regions are bringing in the most money.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Sales by Channel: See which sales methods are working best.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Sales by Salesperson: Check which salespeople are performing well and who may need support.&lt;br&gt;
This step helps you see clearly where the strengths and weaknesses are.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;4. Data Visualization: Building Interactive Dashboards&lt;/strong&gt;&lt;br&gt;
A dashboard is a smart way to show your results nicely.&lt;br&gt;
You can use different charts in Excel like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Bar Charts: To compare sales between regions, channels, or salespeople.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Line Charts: To see how sales are changing over time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Pie Charts: To show how sales are shared between different channels or regions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Slicers: To make the dashboard interactive e.g., you can click on a region and see only its data.&lt;br&gt;
A good dashboard makes it easy for anyone to understand the numbers quickly.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;5. Summary and Insights&lt;/strong&gt;&lt;br&gt;
After your analysis and dashboard are ready, it’s time to explain your findings:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Highlight Top Performers: Recognize best performing regions, channels, and salespeople.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Show Areas That Need Help: Point out regions or channels where sales are low.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Give Smart Advice/recommendations;Focus resources on high-performing areas and develop strategies to improve weak spots.&lt;br&gt;
When you explain your findings well, people can make better decisions based on real facts.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;🎯 Conclusion&lt;/strong&gt;&lt;br&gt;
"Good data in → Great insights out."&lt;br&gt;
Invest time upfront cleaning and organizing your data hence the results will always be worth it.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>The Ultimate Guide to Data Science.</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Sun, 25 Aug 2024 16:03:50 +0000</pubDate>
      <link>https://forem.com/jmwendwa/the-ultimate-guide-to-data-science-ad8</link>
      <guid>https://forem.com/jmwendwa/the-ultimate-guide-to-data-science-ad8</guid>
      <description>&lt;p&gt;Data science is transforming how we perceive and interact with information. It combines statistical, technological, and business expertise to transform raw data into meaningful insights. This article gives a thorough explanation of data science, including the fundamentals, important components, tools, and applications.&lt;/p&gt;

&lt;p&gt;Lets brainstorm this,do you ever wonder what is this data science we are talking about?&lt;br&gt;
Data science is about extracting useful knowledge from data. It entails gathering, cleaning, analysing, and interpreting data to find patterns and insights. The idea is to use these insights to make educated decisions and drive strategic initiatives. Data science is a combination of several fields, including:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Statistics are used to analyse and understand data.&lt;/li&gt;
&lt;li&gt;Computer Science: Programming and data manipulation.&lt;/li&gt;
&lt;li&gt;Domain Knowledge: For applying data insights to specific fields.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Components of Data Science
&lt;/h2&gt;

&lt;p&gt;1.Data Collection: Collecting information from a variety of sources, including databases, online &lt;br&gt;
  questionnaires, and sensors. High-quality data collection is required for accurate &lt;br&gt;
  analysis.&lt;br&gt;
2.Data Cleaning and Preparation: Ensure that the data is clean and well-organized. This includes &lt;br&gt;
  processing missing numbers, rectifying errors, and converting data to a &lt;br&gt;
  usable format. &lt;br&gt;
3.Exploratory Data Analysis (EDA) involves summarising and visualising data in order to comprehend its &lt;br&gt;
  structure and uncover trends. Techniques include descriptive statistics, charts, and graphs.&lt;br&gt;
4.Statistical Analysis: Using statistical tools to test hypotheses and reach conclusions. This may &lt;br&gt;
  include procedures such as regression analysis and hypothesis testing.&lt;br&gt;
5.Machine Learning and Modelling: Creating and training algorithms to predict or classify data. Decision &lt;br&gt;
  trees and neural networks are examples of machine learning models that are used to forecast future &lt;br&gt;
  trends. &lt;br&gt;
6.Data Visualisation: Creating visual representations of data to make difficult information more &lt;br&gt;
  understandable. Matplotlib, Tableau, and Power BI are useful tools for presenting data in the form of &lt;br&gt;
  charts, graphs, and interactive dashboards. &lt;br&gt;
7.Deployment and Integration: Using data models in real-world applications and integrating them with &lt;br&gt;
  existing systems to provide real-time decision assistance. &lt;/p&gt;

&lt;h2&gt;
  
  
  Outstanding Tools and Technologies
&lt;/h2&gt;

&lt;p&gt;1.Python is a versatile programming language popular in data science because to its sophisticated &lt;br&gt;
  libraries, which include Pandas for data processing, NumPy for numerical computations, and Matplotlib &lt;br&gt;
  for visualisation.&lt;br&gt;
2.R is a computer language for statistical analysis and visualisation. R provides programmes such as &lt;br&gt;
  ggplot2 to create complex and customisable graphs.&lt;br&gt;
3.Excel is a spreadsheet utility for organising, analysing, and visualising data. It's excellent for &lt;br&gt;
  performing rapid analyses and making simple charts.&lt;br&gt;
4.SQL is essential for searching and managing relational databases, as well as effectively retrieving &lt;br&gt;
  and processing data.&lt;br&gt;
5.Tableau is well-known for its ability to generate interactive and shareable dashboards, making it &lt;br&gt;
  ideal for data visualisation and insight discovery.&lt;br&gt;
6.Power BI is a Microsoft application that works well with other Microsoft products and is used to &lt;br&gt;
  generate dynamic reports and visualisations.&lt;br&gt;
7.GitHub: Used for version control and collaboration, GitHub enables data scientists to manage code, &lt;br&gt;
  track changes, and collaborate on projects.&lt;/p&gt;

&lt;h2&gt;
  
  
  Application of Data Science
&lt;/h2&gt;

&lt;p&gt;Data science is utilised in many fields to foster innovation and efficiency.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Marketing involves predicting customer behaviour and optimising marketing methods.&lt;/li&gt;
&lt;li&gt;Finance entails predicting market trends and controlling risk.&lt;/li&gt;
&lt;li&gt;Healthcare: Optimising patient outcomes and tailoring treatment methods.&lt;/li&gt;
&lt;li&gt;Retailers analyse sales data to optimise inventory and price.&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>THE ULTIMATE GUIDE TO FEATURE ENGINEERING</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Mon, 19 Aug 2024 11:41:27 +0000</pubDate>
      <link>https://forem.com/jmwendwa/the-ultimate-guide-to-feature-engineering-1mpp</link>
      <guid>https://forem.com/jmwendwa/the-ultimate-guide-to-feature-engineering-1mpp</guid>
      <description>&lt;p&gt;"Data Engineering is like you taking all the frustrating parts of being a data analyst and combining them with all the frustrating parts of being a software engineer.&lt;br&gt;
Until payday and you forget about the frustrations for a moment, haha!!!"&lt;/p&gt;

&lt;p&gt;Hey there, are you into machine learning and want to gain most of your date, probably you should try featuring engineering in your endeavors.&lt;br&gt;
Feature engineering essentially the process of converting raw data like numbers, dates, into something that your machine learning model can comprehend and use. &lt;br&gt;
The better the features, the better your model will perform.&lt;br&gt;
Even with the best algorithms, your model is only as good as the information you feed it. Good features equal better accuracy. Giving your model glasses helps it see things well. &lt;/p&gt;

&lt;p&gt;Now let’s take a look at types of features and know what we are missing out on;   &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Numerical Features: Continuous values, like age, wages, or height.
2.Categorical Features: Are distinct groups like gender(male/female), city (Nairobi, Mombasa).
3.Ordinal Features: Are categorized traits that have a natural order, such as education level or grades (A, B, C).
4.Temporal Features: Include date and time-related information, such as timestamps.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Let’s look at Cool Feature Engineering Tricks:&lt;br&gt;
1.Clean Your Data: Remove missing values and outliers.&lt;br&gt;
Missing value handling techniques include imputation (mean, median, mode), the use of algorithms that support missing values, and the removal of rows/columns with missing data.&lt;br&gt;
Outlier Detection and Removal techniques include use of statistical methods (e.g., Z-scores, IQR) or model-based approaches to identify and manage outliers.&lt;br&gt;
2.Encode Categories: Convert categories into numbers &lt;br&gt;
Label Encoding where you transform categories into numerical labels (e.g., Male = 0, Female = 1).&lt;br&gt;
One-Hot Encoding where you create binary columns for each category (e.g., gender_Male and gender_Female).&lt;br&gt;
Target Encoding where you replace categories with the target variable's mean for each category.&lt;br&gt;
3.Scale Features: Ensure that all of your data is on the same scale so that no one feature overpowers another.&lt;br&gt;
Normalization is the process of scaling features to a specific range, which is usually between 0 and 1.&lt;br&gt;
Standardization involves scaling features to have a zero mean and unit variance.&lt;br&gt;
Robust Scaling uses percentiles to scale features, making it less vulnerable to outliers.&lt;br&gt;
4.Create New Features: Merge or alter existing features to create new ones. &lt;br&gt;
For example, multiply two characteristics or generate time-based features such as "day of the week."&lt;/p&gt;

&lt;p&gt;Here are some of the best practices;&lt;br&gt;
1.Understand Your Data: Perform some analysis to determine what's going on.&lt;br&gt;
2.Start Simple: Don't overcomplicate things; instead, start with basic features.&lt;br&gt;
3.Test and iterate: Experiment with different features to find what works best.&lt;br&gt;
4.Avoid Data Leakage: Using future data to forecast the past is like cheating on a test.&lt;br&gt;
5.Document your work: Take notes on what you've done so you can repeat it if necessary.&lt;/p&gt;

&lt;p&gt;That’s a lot we have looked at. Its amazing right? In that mood of excitement, we also have to take caution at these two things so that we be on the safer side;&lt;br&gt;
1.Don't Overfit: Adding too many features may make your model perform well on existing data but poorly on new data.&lt;br&gt;
2.Use Domain Knowledge: Rather than relying just on tools, construct relevant features based on your understanding of the data.&lt;/p&gt;

&lt;p&gt;All the best!&lt;/p&gt;

</description>
    </item>
    <item>
      <title>UNDERSTANDING YOUR DATA: THE ESSENTIALS OF EXPLORATORY DATA ANALYSIS</title>
      <dc:creator>John Mwendwa</dc:creator>
      <pubDate>Sun, 11 Aug 2024 14:02:27 +0000</pubDate>
      <link>https://forem.com/jmwendwa/understanding-your-data-the-essentials-of-exploratory-data-analysis-3bie</link>
      <guid>https://forem.com/jmwendwa/understanding-your-data-the-essentials-of-exploratory-data-analysis-3bie</guid>
      <description>&lt;p&gt;Exploratory Data Analysis (EDA) is an essential phase in data science that allows you to better understand your data, identify trends, and obtain insights.&lt;br&gt;
EDA analyzes and visualizes data to uncover relationships, detect abnormalities, and verify data quality.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Importance of EDA;&lt;/strong&gt;&lt;br&gt;
1.Summarize main characteristics of the data&lt;br&gt;
2.Gain better understanding of the data set&lt;br&gt;
3.Uncover relationships between different variables &lt;br&gt;
4.extract important variables for the problem that is trying to be solved.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Outlier Detection- Identifies odd data points that may impact the analysis.&lt;/li&gt;
&lt;li&gt;Error detection - Assists in identifying and correcting errors in data.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Tools and Techniques;&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Jupyter Notebook: An interactive environment for EDA (has tools like pandas and NumPy)&lt;/li&gt;
&lt;li&gt;Visualization Libraries: Tools like Matplotlib, Seaborn, and SciPy for creating visualizations.&lt;/li&gt;
&lt;li&gt;Summary Statistics: Key metrics to summarize data (e.g., mean, standard deviation).&lt;/li&gt;
&lt;li&gt;Data Transformation: Techniques like normalization to improve data quality.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Steps involved in EDA;&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Data Collection - Collecting data for your analysis.&lt;br&gt;
Collect data from several sources e.g., databases.&lt;br&gt;
Ensure that the data is correct, complete, and relevant to your situation.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Cleaning – Involves Preparing data by removing inconsistencies and inaccuracies.&lt;br&gt;
A) Remove duplicates: Make sure each record is unique and meaningful&lt;br&gt;
&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;
&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# Check for duplicate records
duplicate_mask = data.duplicated()
# Count the number of duplicate records
num_duplicates = duplicate_mask.sum()
print(f"Number of duplicate records: {num_duplicates}")
# View duplicate records
if num_duplicates &amp;gt; 0:
    print("Duplicate records:")
    print(data[duplicate_mask])
# Optionally: Remove duplicate records
# weather_df = weather_df.drop_duplicates()
# Save the cleaned dataset (if duplicates were removed)
# weather_df.to_csv('cleaned_weather_dataset.csv', index=False)

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;B) Handle Missing Values: Determine whether to eliminate or fill in missing data.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;# Check for null &amp;amp; missing values
data.isnull().sum()

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;C) Standardizing data; Ensuring Consistency in data formats&lt;br&gt;
D) Correcting errors; Fixing any data entry mistakes&lt;/p&gt;

&lt;p&gt;3.Data visualization&lt;br&gt;
  Involves using visual tools to investigate the distribution, trends, and relationships in data.&lt;br&gt;
A) Histograms illustrates the distribution of a single variable.&lt;br&gt;
B) Box plots- used to highlight the distribution and identify outliers.&lt;br&gt;
C) Scatter plots- Investigate the correlations between two variables.&lt;br&gt;
D) Bar charts allow you to compare category data.&lt;/p&gt;

&lt;p&gt;4.Statistical analysis.&lt;br&gt;
   Using statistical measures to summarize the data and analyze its key properties.&lt;/p&gt;

&lt;p&gt;A) Summary Statistics: Determine the mean, median, standard deviation, and other metrics.&lt;br&gt;
`     Data.describe()&lt;/p&gt;

&lt;p&gt;`&lt;br&gt;
B) Correlation Analysis: Determine the relationship between variables using heat map plot&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;#heatmaps to identify relationships between different weather parameters.
correlation_matrix = numeric_data.corr()
plt.figure(figsize=(10, 6))
sns.heatmap(correlation_matrix, annot=True, cmap='coolwarm')
plt.title('Correlation Matrix')
plt.show()

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;5.Interpretation and Insight&lt;br&gt;
  Drawing conclusions and produce insights from the analysis.&lt;br&gt;
A) Interpret visualizations and statistics. Understand the data's patterns, trends, and relationships.&lt;br&gt;
B) Generate hypotheses: Create hypotheses based on the EDA to guide future analysis.&lt;br&gt;
C) Document Findings: Clearly document findings, any effects, and any issues discovered in the data.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
