<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Stephen Ndichu</title>
    <description>The latest articles on Forem by Stephen Ndichu (@s_ndichu).</description>
    <link>https://forem.com/s_ndichu</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1172538%2F6670ae26-f353-4b96-b6df-a7ff5f1fb5ca.jpg</url>
      <title>Forem: Stephen Ndichu</title>
      <link>https://forem.com/s_ndichu</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/s_ndichu"/>
    <language>en</language>
    <item>
      <title>Understanding your data: The Essentials of Exploratory Data Analysis (EDA).</title>
      <dc:creator>Stephen Ndichu</dc:creator>
      <pubDate>Fri, 16 Aug 2024 12:02:59 +0000</pubDate>
      <link>https://forem.com/s_ndichu/understanding-your-data-the-essentials-of-exploratory-data-analysis-eda-1hh7</link>
      <guid>https://forem.com/s_ndichu/understanding-your-data-the-essentials-of-exploratory-data-analysis-eda-1hh7</guid>
      <description>&lt;p&gt;Once data has been collected and stored, there's need for its analysis to derive meaningful understanding of it. It is for this reason that exploratory data analysis (EDA) comes into play. As the name suggests, we are &lt;strong&gt;'exploring'&lt;/strong&gt; the data i.e. getting a general overview of it. &lt;/p&gt;

&lt;p&gt;The data collected may either be text, videos or images and will usually be stored in an unstructured manner. Rarely will you find data that is 100% clean i.e. without any anomalies. Additionally, data may be in various formats like &lt;strong&gt;Excel, CSV (comma separated values), Json, Parquet etc.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In the world of data, EDA may also be referred to as &lt;strong&gt;data manipulation&lt;/strong&gt; or &lt;strong&gt;data cleaning&lt;/strong&gt;. Practitioners in the industry emphasize the importance of cleaning data to remove &lt;strong&gt;'junk'&lt;/strong&gt; as this may negatively impact the results as well as predictions. Structured data, usually in tabular format, can be analysed using several techniques and tools (like Excel, Power BI, SQL) but we will focus on Python for this illustration.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;EDA using Python&lt;/strong&gt;&lt;br&gt;
Python programming language is one of the most widely tools in EDA owing to its versatility which allows for its use across multiple industries, be it finance, education, healthcare, mining, hospitality among others.&lt;br&gt;
Inbuilt libraries, namely Pandas and NumPy are highly effective in this regard and work across board (whether using &lt;strong&gt;Anaconda/Jupyter Notebook, Google Collab, or an IDE like Visual Studio)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Below are the common steps and code lines executable when performing EDA:&lt;/p&gt;

&lt;p&gt;First, you'll import the python libraries necessary for manipulation/analysis:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;import pandas as pd&lt;br&gt;
import numpy as np&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Secondly, load the dataset&lt;br&gt;
&lt;strong&gt;df = pd.read_excel('File path')&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Note: df is the standard function for converting tabular data into a data Frame.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Once loaded, you can preview the data using the code:&lt;br&gt;
&lt;strong&gt;df.head()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This will show the first 5 rows of the dataset&lt;br&gt;
Alternatively, you can simply run &lt;strong&gt;df&lt;/strong&gt; which will show a select few rows (both top and bottom) of the entire dataset as well as all the columns therein.&lt;/p&gt;

&lt;p&gt;Thirdly, understand all the datatypes using:&lt;br&gt;
&lt;strong&gt;df.info()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Note: Datatypes include &lt;strong&gt;integers&lt;/strong&gt; (whole numbers), &lt;strong&gt;floats&lt;/strong&gt; (decimals) or &lt;strong&gt;objects&lt;/strong&gt; (qualitative data/descriptive words).&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;At this step, it's advisable to get summary statistics of the data using:&lt;br&gt;
&lt;strong&gt;df.describe()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This will give you stats like &lt;strong&gt;Mean, Mode, Standard Deviation, Maximum/Minimum values and the Quartiles&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Fourthly, identify whether null values exist in the dataset using:&lt;br&gt;
&lt;strong&gt;df.isnull()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This can then be followed by checking for duplicates (repetitive entries) &lt;br&gt;
&lt;strong&gt;df.duplicated()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Other key aspects of EDA are checking how the various variables in a dataset relate with each other (&lt;strong&gt;Correlation&lt;/strong&gt;) and their &lt;strong&gt;distribution&lt;/strong&gt;. &lt;br&gt;
Correlation can be positive or negative and ranges from -1 to 1. Its code is:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;df.corr()&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Note: A correlation figure close to &lt;strong&gt;1&lt;/strong&gt; indicates a &lt;strong&gt;strong positive correlation&lt;/strong&gt;, while a figure close to &lt;strong&gt;-1&lt;/strong&gt; indicates a &lt;strong&gt;strong negative correlation&lt;/strong&gt;.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Distribution checks on how &lt;strong&gt;symmetrical&lt;/strong&gt; or &lt;strong&gt;asymmetrical&lt;/strong&gt; data is, as well as the &lt;strong&gt;skewness&lt;/strong&gt; of the data and it can either be normal, binomial, Bernoulli or Poisson.&lt;/p&gt;

&lt;p&gt;In summary, exploratory data analysis is an important process in gaining a better understanding of the data. It allows for better visualizations and model building.&lt;/p&gt;

</description>
      <category>python</category>
      <category>excel</category>
      <category>powerbi</category>
      <category>sql</category>
    </item>
    <item>
      <title>The Ultimate Guide to Data Analytics: Techniques and Tools.</title>
      <dc:creator>Stephen Ndichu</dc:creator>
      <pubDate>Fri, 16 Aug 2024 09:15:43 +0000</pubDate>
      <link>https://forem.com/s_ndichu/the-ultimate-guide-to-data-analytics-techniques-and-tools-1c0b</link>
      <guid>https://forem.com/s_ndichu/the-ultimate-guide-to-data-analytics-techniques-and-tools-1c0b</guid>
      <description>&lt;p&gt;With the ever-increasing advancements in technology in our day to day lives, there's an increase in the amount of data being generated. This has necessitated the need for data analysts and data scientists to analyse the raw data so as to make sense of it.&lt;br&gt;
There are a variety of specializations in the data industry that beginners need to familiarize themselves with before deciding which one suits them best. They include &lt;strong&gt;data architects, data analysts, data engineers, machine learning engineers, business analysts, analytics engineers&lt;/strong&gt; among others.&lt;br&gt;
Globally, more and more people are seeking to join the field of data science and analytics. There is an abundance of resources that beginners can tap into as they start their journey into data. Below is a guide for those looking to venture into data science and analytics.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data collection/gathering&lt;/strong&gt;&lt;br&gt;
Usually, most organizational data will be stored in databases in structured formats. However, data can be unstructured or semi structured depending on how its collected and stored. Therefore, it becomes useful for data analysts to be knowledgeable of &lt;strong&gt;data structures&lt;/strong&gt; and &lt;strong&gt;data warehousing&lt;/strong&gt;.&lt;br&gt;
Additionally, data analysts may be required to extract data from websites and structure it in a manner that allows for meaningful analysis. Some tools used in &lt;strong&gt;web scrapping&lt;/strong&gt; include &lt;strong&gt;Beautiful Soup and Selenium&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Manipulation&lt;/strong&gt;&lt;br&gt;
This is usually a vital step in data analytics and can also be referred to as &lt;strong&gt;exploratory data analysis (EDA)&lt;/strong&gt;. while manipulating data, it's important to note data exists in various forms (text, videos or images). Here, data is cleaned, processed and transformed with the aim to identify trends and getting a general outlook of the data. With this, data can be grouped, missing values identified and unwanted values removed.&lt;br&gt;
Some key tools for this are &lt;strong&gt;MS Excel and Google Sheets&lt;/strong&gt;, which are universally used in the business sector for computation and analysis. They offer a wide range of inbuilt functions necessary for manipulating data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Maths and Statistics&lt;/strong&gt;.&lt;br&gt;
Data is either &lt;strong&gt;quantitative&lt;/strong&gt; or &lt;strong&gt;qualitative&lt;/strong&gt; i.e., numeric or non-numeric. Data specialists should be able to perform both simple and complex mathematical and statistical operations on data sets to identify historical trends and predict future outcomes.&lt;br&gt;
For these, beginners should know basics of &lt;strong&gt;descriptive statistics, inferential statistics and probability&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Programming&lt;/strong&gt;&lt;br&gt;
Whereas there are many programming languages that can be used in data analytics, beginners should at least know the fundamentals of &lt;strong&gt;R and Python&lt;/strong&gt;. These are easy to grasp and use in addition to being open source.&lt;br&gt;
Python, in particular, is widely used across different sectors/industries and comes with inbuilt libraries for data analysis and visualizations. They include &lt;strong&gt;Pandas, NumPy, Matplotlib, Seaborn&lt;/strong&gt; among others.&lt;br&gt;
Additionally, people starting out in data analytics should learn &lt;strong&gt;SQL (Structured Query Language)&lt;/strong&gt; which manages and manipulates relational databases.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Visualizations&lt;/strong&gt;&lt;br&gt;
In addition to Matplotlib and Seaborn libraries for visualizations, other programs may be used to create graphs and tables that are easily understandable to all users.&lt;br&gt;
The most common tools are &lt;strong&gt;Tableau, Power BI and Looker&lt;/strong&gt;, which are useful for developing visualizations that can be used for deriving insights and enabling better decision making for the stakeholders.&lt;br&gt;
Another emerging aspect of visualizations is &lt;strong&gt;data physicalisation&lt;/strong&gt;. This involves use of tangible interfaces and shape changing displays that would be of use to visually impaired stakeholders.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Machine Learning&lt;/strong&gt;&lt;br&gt;
The patterns and trends identified during data manipulation and analysis can be used to make future predictions. This is what machine learning (ML) entails. Machine learning engineers use data and algorithms to train models in a variety of ways which can either be supervised, unsupervised, semi supervised or reinforcement learning.&lt;br&gt;
Python has inbuilt frameworks for ML like &lt;strong&gt;Scikit Learn and Tensor Flow.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In a nutshell, data analytics encompasses many tools and techniques which complement each other. Ability to work with these tools enhances both individual and organizational skills when handling and interpreting data.&lt;br&gt;
It is worth remembering that data analytics also requires skills in problem solving, communication and domain knowledge.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>analytics</category>
      <category>data</category>
      <category>dataengineering</category>
    </item>
  </channel>
</rss>
