<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Ann Jeffa</title>
    <description>The latest articles on Forem by Ann Jeffa (@jeffa_jeffa_67f700f712ab7).</description>
    <link>https://forem.com/jeffa_jeffa_67f700f712ab7</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3407918%2F1eeeb732-3805-46e3-9664-aafd3db54a09.png</url>
      <title>Forem: Ann Jeffa</title>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/jeffa_jeffa_67f700f712ab7"/>
    <language>en</language>
    <item>
      <title>Bayesians and Frequentists</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 17 Oct 2025 13:34:09 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/bayesians-and-frequentists-51fl</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/bayesians-and-frequentists-51fl</guid>
      <description>&lt;p&gt;As a statistician ,data analyst or scientist, how do you define your probability. That's the ultimate difference between the Bayesians and  Frequentists. &lt;/p&gt;

&lt;h2&gt;
  
  
  Frequentist
&lt;/h2&gt;

&lt;p&gt;Frequentists define probability with the concept of frequencies. To a Frequentist, saying "there's a 30% chance of rain tomorrow" means that in 100 days with identical atmospheric conditions, it would rain on approximately 30 days of them. Probability uses the idea property of repetition on random processes. This way only works correctly for coin flips and dice rolls but becomes strained when dealing with one time events. Frequentist methods rely entirely on the data collected, without using any prior information. Common examples include t-tests, ANOVA, and confidence intervals.&lt;/p&gt;

&lt;h2&gt;
  
  
  Bayesian
&lt;/h2&gt;

&lt;p&gt;The Bayesian approach treats probability as a degree of belief or confidence in an event. It considers parameters as random variables that can change as new data becomes available. Bayesian methods start with a prior belief about a parameter, then update this belief using observed data to produce a posterior probability. This process, based on Bayes’ theorem, allows continuous learning as new information is introduced.&lt;/p&gt;

&lt;h2&gt;
  
  
  Real World Example
&lt;/h2&gt;

&lt;p&gt;Imagine a doctor testing the effectiveness of a new drug.&lt;br&gt;
A Frequentist would collect data from a large number of patients and test whether the difference in recovery rates between the drug and placebo is statistically significant (example, p-value &amp;lt; 0.05).&lt;br&gt;
A Bayesian would start with prior knowledge about similar drugs, then update the probability that the drug is effective as more patient data becomes available.&lt;/p&gt;

&lt;h2&gt;
  
  
  Summary
&lt;/h2&gt;

&lt;p&gt;In summary, the Frequentist approach focuses on objective, long-term patterns that emerge from repeated experiments, while the Bayesian approach centers on updating beliefs as new evidence becomes available. Each method has its own strengths . Frequentist techniques are straightforward and widely used, whereas Bayesian methods are more flexible and allow for continuous learning. Ultimately, the best approach depends on the nature of the data, how much prior information is available, and the specific goals of the analysis.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>discuss</category>
      <category>science</category>
    </item>
    <item>
      <title>Parametric and Non-Parametric Tests</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 17 Oct 2025 11:34:25 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/parametric-and-non-parametric-tests-p1f</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/parametric-and-non-parametric-tests-p1f</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;In Statistics, different tools , methods and tests are used differently with an end purpopse to derive meaningful insights from data . There are  several methods used to determine the choice of these tools and methods.Type of data is the major factor in choosing the test conducted. The tests are divided either as parametric or non-parametric tests. Let's differentiate them.&lt;/p&gt;

&lt;h2&gt;
  
  
  Parametric Tests
&lt;/h2&gt;

&lt;p&gt;Parametric tests are statistical methods that work under the assumption that data follows a specific distribution, most often a normal distribution. They rely on numerical parameters like the mean, standard deviation, and variance to describe and analyze the data. Because they make these assumptions, parametric tests are generally more powerful and accurate when the data fits the expected conditions, allowing researchers to detect even small differences or relationships within their datasets..&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Examples&lt;/strong&gt;:t_test,ANOVA,Pearsons Correlation,Linear Regression.&lt;/p&gt;

&lt;h2&gt;
  
  
  Non-Parametric Tests
&lt;/h2&gt;

&lt;p&gt;Non-parametric tests are often called distribution-free tests because they do not require the data to follow any particular distribution. These tests are used when data is ordinal, categorical, or not normally distributed. Instead of using numerical values directly, non-parametric tests often rank the data or use medians, making them less affected by extreme values or outliers.Non-parametric tests are especially useful for small sample sizes, non-linear relationships, or when the assumptions of parametric tests cannot be met. Although they are less powerful when compared to parametric tests, they offer flexibility and robustness, ensuring valid conclusions even under irregular data conditions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Examples&lt;/strong&gt;:Chi-square test,Spearman's correlation.&lt;/p&gt;

&lt;h2&gt;
  
  
  Differences
&lt;/h2&gt;

&lt;p&gt;The main difference between parametric and non-parametric tests lies in the assumptions about the data. Parametric tests assume normal distribution, equal variances, and continuous data measured on an interval or ratio scale. Non-parametric tests, however, do not rely on these assumptions and are suitable for ranked or categorical data.&lt;/p&gt;

&lt;p&gt;When deciding which test to use, researchers should first check the normality of the data. If the data is normally distributed and sample size is large, a parametric test is appropriate. However, if the data is skewed, ordinal, or has outliers, a non-parametric test will produce more reliable results.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Both parametric and non-parametric tests are essential tools in statistical analysis. Parametric tests are preferred when the data meets their assumptions because they are more efficient and provide stronger conclusions. Non-parametric tests, on the other hand, offer flexibility and reliability when data does not meet the strict assumptions of normality or when dealing with ranked or categorical data.&lt;/p&gt;

</description>
      <category>analytics</category>
      <category>beginners</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Degrees Of Freedom</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 03 Oct 2025 15:24:57 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/degrees-of-freedom-1dah</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/degrees-of-freedom-1dah</guid>
      <description>&lt;h2&gt;
  
  
  Defination
&lt;/h2&gt;

&lt;p&gt;Degrees of freedom are the number of independent values that a statistical analysis can estimate. You can also think of it as the number of values that are free to vary as you estimate parameters.&lt;/p&gt;

&lt;h2&gt;
  
  
  Scenario
&lt;/h2&gt;

&lt;p&gt;Your hockey team has 20 players registered for a weekend tournament. During each match, only 11 players can be on the pitch, while the rest remain as substitutes. A day before the tournament, your best defender and the only goalkeeper inform the coach that they cannot make it. Since the team has no backup goalkeeper, the coach insists that the goalkeeper must attend. Fortunately, there are three defenders available, so the coach replaces the absent defender with another.&lt;/p&gt;

&lt;p&gt;In your team, 11 players must be on the pitch. That’s a fixed requirement.&lt;br&gt;
If the goalkeeper has no substitute, the coach has no freedom in choosing that position, it’s fixed (0 degrees of freedom for the goalkeeper).&lt;br&gt;
For defenders, there are 3 options to fill 1 spot (since one defender is absent but you still have choices). That means the coach has freedom to decide which defender to bring in (1 degree of freedom for that position).&lt;br&gt;
Similarly, for the other positions, depending on how many substitutes are available, the coach has more freedom of choice.&lt;/p&gt;
&lt;h2&gt;
  
  
  How to calculate Degrees Of Freedom in Statistics
&lt;/h2&gt;

&lt;p&gt;The total degrees of freedom is one less than the total sample size&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;DoF=n-1
where (n) is the sample size. 
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That's the general formula .&lt;/p&gt;

&lt;h2&gt;
  
  
  Importance of Degrees of Freedom
&lt;/h2&gt;

&lt;p&gt;Degrees of freedom matter because they affect the shape of probability distributions used in tests. For instance:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The t-distribution changes depending on the degree of freedom, with larger degree of freedom, it approaches the normal distribution.&lt;/li&gt;
&lt;li&gt;In chi-square tests, they determine the critical values needed to decide whether to reject a hypothesis&lt;/li&gt;
&lt;li&gt;In regression analysis, they help calculate residual variance and test model fit.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In short, degrees of freedom ensure that statistical tests adjust for the amount of information and constraints in your data.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Understanding degrees of freedom helps us interpret statistical tests more accurately, ensuring that our conclusions are based on the right amount of free information.&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>datascience</category>
      <category>science</category>
      <category>learning</category>
    </item>
    <item>
      <title>List Comprehension vs Dictionary Comprehension in Python</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 26 Sep 2025 20:35:20 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/list-comprehension-vs-dictionary-comprehension-in-python-40ai</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/list-comprehension-vs-dictionary-comprehension-in-python-40ai</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;In Python, writing clean and efficient code is highly valued. Instead of using long loops, we often rely on comprehensions ,short, elegant ways to build new data structures. The two most common ones are list comprehension and dictionary comprehension. They look similar at first glance, but they serve different purposes.&lt;/p&gt;

&lt;h2&gt;
  
  
  List Comprehension
&lt;/h2&gt;

&lt;p&gt;A list comprehension is a way to create a list by applying an expression to each item in an iterable.&lt;br&gt;
Python loops over the required range returning a new list. List comprehension produces a list of values.&lt;br&gt;
Example&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;squares = [i**2 for i in range(1, 6)]
print(squares)  
[1, 4, 9, 16, 25]
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Dictionary Comprehension
&lt;/h2&gt;

&lt;p&gt;A dictionary comprehension is similar, but instead of producing a list, it produces a dictionary with key–value pairs. Dictionary comprehension produces a mapping of keys to values.&lt;br&gt;
Example&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;squares_dict = {i: i**2 for i in range(1, 6)}
print(squares_dict)
 {1: 1, 2: 4, 3: 9, 4: 16, 5: 25}

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Key Difference&lt;/strong&gt;&lt;br&gt;
In list Comprehension the results are ordered values in a list.&lt;br&gt;
Whereas in dictionary Comprehension the results are key value pairs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Combining Two Lists into a Dictionary with Comprehension&lt;/strong&gt;&lt;br&gt;
Imagine you have two lists:&lt;br&gt;
Method 1: Using zip()&lt;br&gt;
zip() pairs items from both lists together, and the dictionary comprehension builds key–value pairs.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;names = ["Alice", "Bob", "Charlie"]
scores = [85, 90, 95]
result = {zip(names, scores)}
print(result)  
{'Alice': 85, 'Bob': 90, 'Charlie': 95}
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Method 2: Using Indexing&lt;/p&gt;

&lt;p&gt;Here, the comprehension loops over the index and maps the matching items.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;names = ["Alice", "Bob", "Charlie"]
scores = [85, 90, 95]
result = {names[i]: scores[i] for i in range(len(names))}
print(result)
{'Alice': 85, 'Bob': 90, 'Charlie': 95}
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;List comprehension -creates a list of values.&lt;br&gt;
Dictionary comprehension -creates key–value pairs in a dictionary.&lt;br&gt;
We can combine two lists into a dictionary by using zip() or index-based dictionary comprehension.&lt;br&gt;
Comprehensions make Python code shorter, faster, and more readable.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Skewness and Kurtosis</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 26 Sep 2025 19:40:16 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/skewness-and-kurtosis-2a1b</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/skewness-and-kurtosis-2a1b</guid>
      <description>&lt;p&gt;When analyzing data in statistics or data science, it is not enough to only look at measures of central tendency like the mean, median, or mode and variability (like variance or standard deviation). To fully understand the shape of a dataset’s distribution, we use skewness and kurtosis. These two measures describe how data deviates from a perfectly normal (bell-shaped) distribution.&lt;/p&gt;

&lt;h2&gt;
  
  
  Skewness
&lt;/h2&gt;

&lt;p&gt;It measures the asymmetry of a distribution. A perfectly normal distribution has skewness equal to 0, meaning it is symmetric around the mean.&lt;/p&gt;

&lt;p&gt;Types of Skewness:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Positive Skew (Right-skewed):&lt;br&gt;
The right tail is longer than the left.&lt;br&gt;
Most data values are concentrated on the left, but a few very large values pull the mean to the right.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Negative Skew (Left-skewed):&lt;br&gt;
The left tail is longer than the right.&lt;br&gt;
Most data values are concentrated on the right, but a few very small values pull the mean to the left.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Zero Skew:&lt;br&gt;
The distribution is symmetric.&lt;br&gt;
Mean = Median = Mode.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv6sxxcgr37pvod251h75.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv6sxxcgr37pvod251h75.png" alt=" " width="800" height="309"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Kurtosis
&lt;/h2&gt;

&lt;p&gt;Kurtosis measures the “tailedness” of a distribution, or how extreme the outliers are compared to a normal distribution. &lt;br&gt;
Types of Kurtosis:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Leptokurtic
Heavy tails and a sharp peak.
More extreme outliers than a normal distribution.&lt;/li&gt;
&lt;li&gt;Platykurtic 
Light tails and a flatter peak.
Fewer outliers than a normal distribution.&lt;/li&gt;
&lt;li&gt;Mesokurtic 
Normal bell-shaped curve.
Moderate tails and peak.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwprc9ndq8o7gwitl8w6t.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwprc9ndq8o7gwitl8w6t.jpg" alt=" " width="580" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Importance of skewness and kurtosis in data science.
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;Data Analysis:
They reveal whether data follows assumptions of normality.&lt;/li&gt;
&lt;li&gt;Risk Management
In finance, skewness and kurtosis help in understanding market risks — highly skewed or leptokurtic data indicates greater uncertainty.&lt;/li&gt;
&lt;li&gt;Decision-Making:
They help analysts avoid misleading conclusions that come from looking at mean and standard deviation alone.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;br&gt;
Skewness tells us about the symmetry of data.&lt;br&gt;
Kurtosis tells us about the outliers and tail heaviness.&lt;br&gt;
Together, they provide a deeper picture of data distribution beyond averages and variability, helping statisticians, data scientists, and decision-makers draw more accurate insights.&lt;/p&gt;

</description>
      <category>analytics</category>
      <category>beginners</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Similarities Between Stored Procedures and Python Functions</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Fri, 26 Sep 2025 18:24:29 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/similarities-between-stored-procedures-and-python-functions-5fee</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/similarities-between-stored-procedures-and-python-functions-5fee</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;In the field of data science, python and sql are both tools used in day-to-day work . A Stored Procedure is collection of SQL statements that are supposed to be executed at the same time while a Python function is a block of reusable code that performs a specific task and runs only when called . They help make code more organized, reusable, and easier to maintain.&lt;/p&gt;

&lt;p&gt;Although they exist in different environments—stored procedures inside a database system, and Python functions inside application code—they share a number of similarities in concept and purpose.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Groups details in a single unit&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Both stored procedures and Python functions provide a way to group a set of instructions into a single named block.&lt;br&gt;
A stored procedure contains several SQL statements to retrieve, insert, or update data.&lt;br&gt;
A Python function is like a container that holds lines of codes. &lt;br&gt;
Examples&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;CREATE PROCEDURE Get_Employees
AS
BEGIN
    SELECT employee_id, first_name, last_name, department
    FROM employees;
END;

-- Call the procedure
CALL EmployeesByDepartment('Sales');

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;





&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;
def add_numbers(a, b):
    return a + b
result = add_numbers(5, 3)
print("The sum is:", result)

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;2.Reusability&lt;/strong&gt;&lt;br&gt;
Reusability is one of the strongest similarities between the two. Instead of writing the same block of code multiple times, you can define the logic once and call it whenever needed. This reduces duplication and improves efficiency.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.Input Parameters&lt;/strong&gt;&lt;br&gt;
Both stored procedures and Python functions can accept input parameters that make them dynamic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Returning Results&lt;/strong&gt;&lt;br&gt;
Stored procedures can return data set that is query results.&lt;br&gt;
Python functions return values using the return statement.&lt;br&gt;
This makes them useful for producing outputs based on given inputs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.Control Flow Capabilities&lt;/strong&gt;&lt;br&gt;
Both support control flow logic, such as conditionals and loops.&lt;br&gt;
Stored procedures can use IF…ELSE, CASE, or looping constructs like WHILE.&lt;br&gt;
Python functions can use if, for, and while loops.&lt;br&gt;
This allows developers to write logic that adapts to different conditions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6.Modularity and Maintainability&lt;/strong&gt;&lt;br&gt;
By breaking down tasks into smaller reusable blocks, both stored procedures and Python functions promote modularity. This makes applications easier to read, maintain, and debug.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7.Execution method.&lt;/strong&gt;&lt;br&gt;
Both are executed by calling their name with the required parameters.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Stored procedures and Python functions may belong to different environments, but their similarities lie in structure, purpose, and usage. Both serve to encapsulate logic, promote reusability, accept parameters, return results, support control flow, and enhance maintainability.&lt;/p&gt;

&lt;p&gt;In essence, whether you are working inside a database or writing application code, the concepts behind stored procedures and functions remain closely related.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>CROPS ANALYSIS DASHBOARD ON POWER BI</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Wed, 27 Aug 2025 21:07:44 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/crops-analysis-dashboard-on-power-bi-513i</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/crops-analysis-dashboard-on-power-bi-513i</guid>
      <description>&lt;h2&gt;
  
  
  POWER BI ON AGRICULTURE
&lt;/h2&gt;

&lt;p&gt;In today’s world, most economic activities are data-driven, including&lt;br&gt;
 Agriculture. Agriculture is no longer just about intuition and experience. Farmers and agribusinesses are increasingly relying on data analytics to optimize crop yields, reduce costs, and improve sustainability. Microsoft Power BI—a business intelligence platform that turns raw data into interactive, insightful dashboards.&lt;/p&gt;

&lt;p&gt;When applied to the crops sector, Power BI enables stakeholders to monitor everything from production costs to profits, weather impacts, and market demand. Stakeholders are also able to make meaningful and informed decisions from the insights of the data. Below, we explore how Power BI can be used in agriculture, focusing on crops, and cover all aspects of the process.&lt;/p&gt;

&lt;h2&gt;
  
  
  DATA CLEANING AND TRANSFORMATION.
&lt;/h2&gt;

&lt;p&gt;Data Cleaning in Power BI is a critical and a must step in ensuring reliable crop analysis because data often comes in bias or inconsistent formats. Using Power Query remove duplicates, fill or replace missing values, and standardize measurements.&lt;br&gt;
In this case, are the steps I took to clean the crops dataset&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Changing the formats for the different columns to their required format  ,that is ensuring the right data type. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Filled the blanks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Corrected errors&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  KEY METRICS
&lt;/h2&gt;

&lt;p&gt;The KPIs show important information from the data.&lt;br&gt;
The Key Performance Indicators I conducted on the dataset include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Yield in kilograms&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Revenue and Profit by Crop Type&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cost of Production &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Tracking these indicators helps both smallholder farmers and large agribusinesses make data-backed decisions.&lt;/p&gt;

&lt;h2&gt;
  
  
  VISUALIZATIONS
&lt;/h2&gt;

&lt;p&gt;Data visualizations enables one to tell a story about the data.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frr0y9g5qii40poniugy9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frr0y9g5qii40poniugy9.png" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  DATA INSIGHTS FROM THE DATA
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Key Perfomance&lt;/strong&gt;&lt;br&gt;
• Total Revenue: KES 1.19biliion&lt;br&gt;
• Total Profit: KES 1.10bn&lt;br&gt;
• Average Revenue: KES 121.67million&lt;br&gt;
• Average  Profit: KES 197.13k&lt;br&gt;
&lt;strong&gt;Crop Perfomance&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Top Performing Crop is rice. It generated KES 133.01M in revenue with KES 114.94M profit.&lt;/li&gt;
&lt;li&gt;Rice did produce more revenue during the rainy season and poorly during the dry season.&lt;/li&gt;
&lt;li&gt;Sorghum made KES 121.12M revenue with KES 119.18M profit .&lt;/li&gt;
&lt;li&gt;Tomatoes was the least revenue generating crop .Its total revenue was 58.11M and profit of 51.99M.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;County Perfomance&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Nyeri County was the most productive county 162.05M as total revenue and 160.09M as profit even with it not producing the highest yield.&lt;/li&gt;
&lt;li&gt;Nakuru County was the second best performing with a total revenue of132.89M and a total profit of 98.43M.&lt;/li&gt;
&lt;li&gt;Nairobi produced the highest yield of crops at 142k KG and Kericho the least of 92k KG of total crops yield&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;More Insights&lt;/strong&gt;&lt;br&gt;
• Farmers who used the DAP fertilizer and harvested the most yields while those who used CAN got least yields&lt;br&gt;
• Crops seem to perform better on clay soil and produce more yields than the rest but loam soil produces the least yields.&lt;br&gt;
• The organic crop  did best while hybrid did poorly .&lt;br&gt;
• The average crop profit was 197.53M&lt;/p&gt;

&lt;h2&gt;
  
  
  CONCLUSION
&lt;/h2&gt;

&lt;p&gt;Power BI has the potential to revolutionize crop management by turning raw agricultural data into actionable insights. From individual farmers to national policymakers, the tool empowers all players in the agricultural ecosystem to make data-driven, sustainable, and profitable decisions.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>EXCEL’S STRENGTHS AND WEAKNESSES IN PREDICTIVE ANALYSIS AND THE ROLE OF EXCEL IN MAKING DATA DRIVEN BUSINESS DESICIONS</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Mon, 11 Aug 2025 13:13:07 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/excels-strengths-and-weaknesses-in-predictive-analysis-and-the-role-of-excel-in-making-data-driven-2ofk</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/excels-strengths-and-weaknesses-in-predictive-analysis-and-the-role-of-excel-in-making-data-driven-2ofk</guid>
      <description>&lt;h2&gt;
  
  
  INTRODUCTION
&lt;/h2&gt;

&lt;p&gt;Excel offers valuable tools for predictive analysis and plays a significant role in data-driven business decisions, but has also has limitations. Its advantages include easy to use, widely spread &amp;amp; known and the ability to handle data manipulation tasks and wide range of calculations. Otherwise, excel is not able to support complex data models, real-time data, large datasets and collaboration making it less suitable for large scale and advanced predictive analysis as users are limited only to specific conditions.&lt;/p&gt;

&lt;h2&gt;
  
  
  EXCELS STRENGTHS
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Familiarity and Easy to use.&lt;br&gt;
Excel is widely spread and adopted in various industries for its strengths in project management. Its easy-to use interface and familiar functionalities makes it an accessible tool for professionals with low-level skills and hence less time in training but more on execution.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Summarization and Visualizations.&lt;br&gt;
Excel is able to calculate descriptive statistics, create charts and graphs, build interactive dashboards and also use of conditional formatting helps in understanding data patterns and trends.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Advanced Data Analysis Capabilities.&lt;br&gt;
Its ability to use pivot tables and complex statistical functions enables one to analyse deeply. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Flexibility and Customization.&lt;br&gt;
Project managers are able to create and setup their projects to their liking without conforming to a specific style.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Integration with other Microsoft Tools.&lt;br&gt;
Excels seamless integration with other Microsoft Office Applications like Word, PowerPoint enhances its utility in project management.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  EXCEL'S WEAKNESSES
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Large Datasets.&lt;br&gt;
Excel's performance reduces when handling very large datasets or complex calculations therefore wasting time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Difficulty with Advanced Statistical Modelling.&lt;br&gt;
Its functions may not be sufficient for sophisticated predictive modelling techniques, such as regression&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Collaboration.&lt;br&gt;
Managing and collaborating on large, complex spreadsheets can be challenging and prone to errors.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Connections.&lt;br&gt;
Excel does not support other applications apart from  Microsoft Applications.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  ROLE OF EXCEL IN DATA-DRIVEN BUSINESS DECISIONS
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Informed Decision Making.&lt;br&gt;
It provides tools to analyse data, identify trends and generate insights enabling business managers to make more informed decisions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Financial Analysis.&lt;br&gt;
Excel is used for budgeting, financial modelling and forecasting. This helps plan for future of a business.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Performance Tracking.&lt;br&gt;
Excels helps track key performance indicators (KPIs), analyse trends and identify areas of improvement.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Risk Assessment.&lt;br&gt;
The outcome after analysis can be used to assess potential risks and opportunities.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  CONCLUSION
&lt;/h2&gt;

&lt;p&gt;Excel is a powerful tool in data manipulation, visualization and predictive analysis making it suitable for wide range of business applications. However, its limitations in handling complex models, collaboration and connection with other applications is a setback.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>POSTGRESQL INSTALLATION ON LINUX SERVER</title>
      <dc:creator>Ann Jeffa</dc:creator>
      <pubDate>Sun, 03 Aug 2025 16:14:43 +0000</pubDate>
      <link>https://forem.com/jeffa_jeffa_67f700f712ab7/postgresql-installation-on-linux-server-4mjf</link>
      <guid>https://forem.com/jeffa_jeffa_67f700f712ab7/postgresql-installation-on-linux-server-4mjf</guid>
      <description>&lt;p&gt;PostgreSQL is a powerful, open-source object-relational database system known for its robustness, extensibility, and standards compliance.&lt;br&gt;
For one to install postgreSQL on linux ,one must&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You must have a Linux server (like Ubuntu).&lt;/li&gt;
&lt;li&gt;&lt;p&gt;You must have admin rights (called sudo).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;You must know how to use the terminal/command line.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;INSTALLATION PROCESS&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; One must update the system first using commands on the terminal
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;sudo apt update
sudo apt upgrade -y

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Installing postgres, main database and its tools are installed.&lt;br&gt;
&lt;code&gt;sudo apt install postgresql postgresql-contrib -y&lt;br&gt;
&lt;/code&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Checking if postgres is running after installation&lt;br&gt;
&lt;code&gt;sudo systemctl status postgresql&lt;br&gt;
&lt;/code&gt;&lt;br&gt;
Ensure that postgress is running and if not start postgress&lt;br&gt;
&lt;code&gt;sudo systemctl start postgresql&lt;br&gt;
&lt;/code&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;After installing, postgress creates a user called postgress and we should switch to that user.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;code&gt;sudo -i -u postgres&lt;br&gt;
&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;5.Getting into the database&lt;br&gt;
&lt;code&gt;psql&lt;br&gt;
&lt;/code&gt; &lt;br&gt;
 You should set a password for postgress while still in the database&lt;br&gt;
&lt;code&gt;ALTER USER postgres PASSWORD 'mypassword';&lt;br&gt;
&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Your postgress has been installed on the linux server. If you want you can create your own database.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
