<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: John Mambo</title>
    <description>The latest articles on Forem by John Mambo (@am_mambo).</description>
    <link>https://forem.com/am_mambo</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F815052%2Fcf1ee30d-1992-4896-ba2e-1c72c2486499.jpg</url>
      <title>Forem: John Mambo</title>
      <link>https://forem.com/am_mambo</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/am_mambo"/>
    <language>en</language>
    <item>
      <title>Time Series Models for beginners complete guide</title>
      <dc:creator>John Mambo</dc:creator>
      <pubDate>Thu, 26 Oct 2023 14:09:07 +0000</pubDate>
      <link>https://forem.com/am_mambo/time-series-models-for-beginners-complete-guide-o7n</link>
      <guid>https://forem.com/am_mambo/time-series-models-for-beginners-complete-guide-o7n</guid>
      <description>&lt;p&gt;Time series data is a sequence of data points collected at regular time intervals. It has a natural temporal ordering, making it unique compared to cross-sectional data. Characteristics include trend, seasonality, and autocorrelation.&lt;br&gt;
Examples of time series data are stock prices, weather measurements, and sales figures, website traffic, disease spread, and more. Recognizing these data patterns is essential for analysis.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why Time Series Analysis?&lt;/strong&gt;&lt;br&gt;
Time series analysis finds applications in diverse fields such as finance (stock price prediction), economics (GDP forecasting), meteorology (weather prediction), and business (sales forecasting), providing insights and predictions to inform decision-making.&lt;/p&gt;

&lt;h3&gt;
  
  
  Importance of Forecasting
&lt;/h3&gt;

&lt;p&gt;Forecasting is a key motivation for time series analysis. It helps us predict future values based on historical data, allowing businesses and organizations to plan, allocate resources, and adapt to changing conditions.&lt;/p&gt;

&lt;h2&gt;
  
  
  Components of Time Series Data
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Trend: The long-term direction of the data.&lt;/li&gt;
&lt;li&gt;Seasonality: Repeating patterns at regular intervals.&lt;/li&gt;
&lt;li&gt;Residuals: Unexplained fluctuations or noise.&lt;/li&gt;
&lt;li&gt;Stationarity refers to the constancy of statistical properties over time, which is important in time series analysis. Non-stationary data may need transformations to make them suitable for modeling.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Getting Started with Data for Time Series Analysis.
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Data Collection and Cleaning
Begin by collecting reliable data from trusted sources. Clean the data by handling missing values, correcting errors, and ensuring data consistency.&lt;/li&gt;
&lt;li&gt;Handling Missing Values and Outliers
Missing values can disrupt analysis, so impute or remove them. Outliers can distort results, so address them carefully.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Exploratory Data Analysis (EDA) for Time Series
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Visualizing Time Series Data&lt;/strong&gt;: Visualizations like line plots, histograms, and box plots help reveal patterns and trends in the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Autocorrelation and Partial Autocorrelation Functions&lt;/strong&gt;: These functions help identify how current values are correlated with past values, assisting in model selection.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Seasonal Decomposition of Time Series (STL)&lt;/strong&gt;:STL decomposes time series into its components, facilitating trend and seasonality identification.&lt;/p&gt;

&lt;h2&gt;
  
  
  Choosing the Right Model
&lt;/h2&gt;

&lt;p&gt;Understand the various models available, such as ARIMA, exponential smoothing, seasonal models like SARIMA, Prophet, and LSTM networks.&lt;/p&gt;

&lt;h2&gt;
  
  
  ARIMA Modeling for Time Series
&lt;/h2&gt;

&lt;p&gt;Understanding AR, MA, and D components&lt;br&gt;
ARIMA models have AutoRegressive (AR), Moving Average (MA), and Differencing (D) components. Each serves a unique purpose in modeling time series data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Selecting the Right Model Order&lt;/strong&gt;&lt;br&gt;
Choosing the correct order (p, d, q) for ARIMA models is crucial. Tools like AIC and BIC can help in this process.&lt;br&gt;
Parameter Estimation and Model Fitting&lt;br&gt;
Estimate model parameters and fit the ARIMA model to your data.&lt;br&gt;
Diagnosing Model Performance&lt;br&gt;
Evaluate the model using residual analysis and diagnostic tests like the Ljung-Box test.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Exponential Smoothing Models&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;Simple Exponential Smoothing&lt;/strong&gt;: Learn the basics of simple exponential smoothing, suitable for time series with no trend or seasonality.&lt;br&gt;
&lt;strong&gt;Holt-Winters Exponential Smoothing&lt;/strong&gt;: This method accounts for trends and seasonality, making it suitable for more complex time series data.&lt;br&gt;
&lt;strong&gt;Model Selection and Evaluation&lt;/strong&gt;: Select the appropriate exponential smoothing model and evaluate its performance.&lt;/p&gt;

&lt;h2&gt;
  
  
  SARIMA Modeling
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Seasonal ARIMA&lt;/strong&gt;: SARIMA models are an extension of ARIMA that incorporates seasonality.&lt;br&gt;
&lt;strong&gt;Seasonal Differencing and Lag Selection&lt;/strong&gt;: Identify the correct seasonal differencing and lag values for SARIMA modeling.&lt;br&gt;
&lt;strong&gt;Fitting and Forecasting with SARIMA&lt;/strong&gt;: Fit the SARIMA model to your data and use it for making forecasts.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prophet Model
&lt;/h2&gt;

&lt;p&gt;Prophet is a tool developed by Facebook for forecasting with strong seasonality and holiday effects. Learn how to apply Prophet for time series forecasting, especially for business-related data.&lt;br&gt;
Fine-tune Prophet models to improve forecasting accuracy.&lt;/p&gt;

&lt;h2&gt;
  
  
  LSTM Networks for Time Series
&lt;/h2&gt;

&lt;p&gt;LSTM networks are a type of RNN designed to handle sequential data like time series.&lt;br&gt;
Learn how to implement and train LSTM networks for time series forecasting.&lt;br&gt;
Prepare your time series data for LSTM modeling, including scaling and window.&lt;/p&gt;

&lt;h2&gt;
  
  
  Model Evaluation and Validation
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Train-Test Split&lt;/strong&gt;: Split your data into training and testing sets to assess model performance.&lt;br&gt;
&lt;strong&gt;Cross-Validation Techniques&lt;/strong&gt;: Cross-validation helps ensure the model's generalizability by validating its performance on different subsets of data.&lt;br&gt;
&lt;strong&gt;Common Evaluation Metrics&lt;/strong&gt;: Use metrics like Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and others to measure model accuracy.&lt;/p&gt;

&lt;h2&gt;
  
  
  Hyperparameter Tuning and Optimization
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Grid Search and Random Search&lt;/strong&gt;: Optimize model hyperparameters using techniques like grid search and random search to enhance model performance.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fine-tuning Model Performance&lt;/strong&gt;:Refine your models by adjusting hyperparameters and model settings for better accuracy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world Applications and Case Studies&lt;/strong&gt;: Practical Implementations of Time Series Models&lt;br&gt;
Explore real-world use cases, such as stock price prediction, sales forecasting, and more.&lt;/p&gt;

</description>
      <category>dsea</category>
      <category>newbie</category>
      <category>dataengineering</category>
      <category>python</category>
    </item>
    <item>
      <title>Exploratory data Analysis using Visualization Techniques</title>
      <dc:creator>John Mambo</dc:creator>
      <pubDate>Sun, 08 Oct 2023 13:02:51 +0000</pubDate>
      <link>https://forem.com/am_mambo/exploratory-data-analysis-using-visualization-techniques-2lf9</link>
      <guid>https://forem.com/am_mambo/exploratory-data-analysis-using-visualization-techniques-2lf9</guid>
      <description>&lt;p&gt;&lt;strong&gt;Exploratory Data Analysis (EDA)&lt;/strong&gt; refers to the method of studying and exploring record sets to apprehend their predominant traits, discover patterns, locate outliers, and identify relationships between variables. EDA is normally carried out as a preliminary step before undertaking extra formal statistical analyses or modeling.&lt;/p&gt;

&lt;h3&gt;
  
  
  Goals of Exploratory Data Analysis
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;1. Data Cleaning:&lt;/strong&gt; Handling missing values, removing outliers, and ensuring data quality. Data Scientists spend 80% of their time cleaning Data.&lt;br&gt;
&lt;strong&gt;2.Data Exploration:&lt;/strong&gt; Mostly involves identifying patterns from the cleaned Data.&lt;br&gt;
&lt;strong&gt;3. Data Visualization:&lt;/strong&gt; Visualizations consisting of histograms, box plots, scatter plots, line plots, heatmaps, and bar charts assist in identifying styles, trends, and relationships within the facts.&lt;br&gt;
&lt;strong&gt;4.Hypothesis Generation:&lt;/strong&gt; EDA aids in generating studies questions based totally on the preliminary exploration of the data. It facilitates form the inspiration for in addition evaluation and model building.&lt;br&gt;
&lt;strong&gt;5. Correlation and Relationships:&lt;/strong&gt; EDA allows discover relationships and dependencies between variables. Techniques such as correlation analysis, scatter plots, and pass-tabulations offer insights into the power and direction of relationships between variables.&lt;/p&gt;
&lt;h3&gt;
  
  
  Steps Involved in Exploratory Data Analysis.
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;1. Importing and Reading Data:&lt;/strong&gt; involves importing the required libraries used in data cleaning, description, analyzation and visualization.&lt;br&gt;
Lets use the titanic data set that is freely found &lt;a href="https://www.kaggle.com/datasets/hesh97/titanicdataset-traincsv"&gt;here&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns

titanic_data = sns.load_dataset('titanic')
#using pandas 
titanic_data = pd.read_csv("path_to_your_dataset_location")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;2.Understanding the Data:&lt;/strong&gt; Find the characteristiics of your Data, its content and stracture among others.&lt;br&gt;
Example&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;titanic_data.shape
titanic_data.head(10)
titanic_data.dtypes
titanic_data.describe()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;3. Data Preparation:&lt;/strong&gt; Identity and Remove Duplicates, Drop irrelevant data and make it Ready for Analysis.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;#Drop any column with missing data
titanics_data.dropna()
#Fill key missing Data
titanic_data.fillna(value)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;strong&gt;4.Data Exploration:&lt;/strong&gt; Examine statistics, visualize the data distributions and Identify patterns.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.Feature Engineering:&lt;/strong&gt; Involves creating and transforming new variables or features found in the dataset to improve performance of machine learning model.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6.Data Visualization:&lt;/strong&gt; Presenting the insights derived by the features through plots, charts and graphs to communicate findings and story-tell effectively.&lt;/p&gt;

&lt;h2&gt;
  
  
  Types of Exploratory Data Analysis
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1.Univariate Non-graphical:&lt;/strong&gt; Here data consists of one variable and may not deal with relationships.&lt;br&gt;
&lt;strong&gt;2. Univariate Graphical:&lt;/strong&gt; It involves summarizing and visualizing a unmarried variable at a time to understand its distribution, relevant tendency, unfold, and different applicable records. Techniques like histograms, field plots and bar charts information are generally used in univariate analysis&lt;br&gt;
Example Drawing Histograms using the titanic dataset.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import seaborn as sns
import matplotlib.pyplot as plt

titanic_data = sns.load_dataset('titanic')

plt.figure(figsize=(8, 5))
sns.histplot(titanic_data['age'].dropna(), bins=30, kde=True, color='blue')
plt.xlabel('Age')
plt.title('Distribution of Passenger Ages')
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output:&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--9ttFd_hx--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/9g4o88r4ueuxyfnwfc42.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--9ttFd_hx--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/9g4o88r4ueuxyfnwfc42.png" alt="Output Histogram" width="800" height="415"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Multivariate non-Graphical:&lt;/strong&gt; In this type, data arises from more than one variable. it shows the relationship between two or more variables of the data through cross-tabulation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Multivariate Graphical:&lt;/strong&gt; Uses Graphics to display relationships between two or more sets of Data. For example grouped bar plot or bar chart.&lt;br&gt;
Drawing a grouped bar plot using the titanic data&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import seaborn as sns
import matplotlib.pyplot as plt
import pandas as pd

# Load the Titanic dataset
titanic = sns.load_dataset('titanic')

# Grouped bar plot showing survival count by class and sex
sns.set(style="whitegrid")
g = sns.catplot(x="class", hue="sex", col="survived",
                data=titanic, kind="count",
                height=4, aspect=0.7, palette="pastel")

# Customize labels and title
g.set_axis_labels("Class", "Count")
plt.subplots_adjust(top=0.85)
g.fig.suptitle('Survival Count by Class and Sex')

plt.show()

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Output&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--WHwy5Rmo--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/48cyf75y6gh6cw0hakw8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--WHwy5Rmo--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/48cyf75y6gh6cw0hakw8.png" alt="Grouped bar plot" width="800" height="365"&gt;&lt;/a&gt;&lt;br&gt;
In this article, we have mainly focused on using python for our Exploratory data Analysis examples, One can also use R programing language too.&lt;br&gt;
In Summary, EDA is a crucial phase in Every day's Data analysis procedures. It helps reveal valuable knowledge hidden within data driving businesses into asking the right questions, making better decisions through better insights and effective problem solving.&lt;/p&gt;

</description>
      <category>dsea</category>
      <category>newbie</category>
      <category>datascience</category>
      <category>dataengineering</category>
    </item>
    <item>
      <title>Data Science for beginners 2023 - 2024 Complete Roadmap</title>
      <dc:creator>John Mambo</dc:creator>
      <pubDate>Mon, 02 Oct 2023 08:48:54 +0000</pubDate>
      <link>https://forem.com/am_mambo/data-science-for-beginners-2023-2024-complete-roadmap-2dkh</link>
      <guid>https://forem.com/am_mambo/data-science-for-beginners-2023-2024-complete-roadmap-2dkh</guid>
      <description>&lt;p&gt;Data Science is an Interdisciplinary field that focuses on analyzing massive amounts of data to automatically identify inherent patterns, extract underlying models, and make relevant predictions.&lt;br&gt;
It Impacts virtually all areas of the economy, including science, engineering, medicine, banking, finance, sports and the arts hence offering endless opportunities.&lt;br&gt;
With the right Data science Roadmap, dedication, practice and mentorship, one can become a good data scientist. &lt;br&gt;
This Roadmap provides a solid foundation you can rely to kickstart your Data science Career.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;1. Understanding the basics:&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Involves understanding what you want to become: a data scientist, understanding the roles and career paths in this field.&lt;br&gt;
&lt;strong&gt;Some career paths in Data Science include;&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data Analyst&lt;/li&gt;
&lt;li&gt;Data Engineer&lt;/li&gt;
&lt;li&gt;Machine Learning Engineer&lt;/li&gt;
&lt;li&gt;NLP Engineeer&lt;/li&gt;
&lt;li&gt;Business Analyst&lt;/li&gt;
&lt;li&gt;Power BI Engineer&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Scientist&lt;br&gt;
&lt;strong&gt;Major Roles of Data Scientists are;&lt;/strong&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Identify valuable data sources and automate collection processes&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Undertake preprocessing of structured and unstructured data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Analyze large amounts of information to discover trends and patterns&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Build predictive models and machine-learning algorithms&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Combine models through ensemble modeling&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Present information using data visualization techniques&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Propose solutions and strategies to business challenges&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Collaborate with engineering and product development teams&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  2.Mathematics
&lt;/h2&gt;

&lt;p&gt;Mathematical knowledge is essential for data science since it is the building foundation of Machine Learning and Data analysis.&lt;br&gt;
learn;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Statistics&lt;/li&gt;
&lt;li&gt;Probability&lt;/li&gt;
&lt;li&gt;calculus&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  3. Learn Programming
&lt;/h2&gt;

&lt;p&gt;Understand and get hands-on experience in programming languages to implement various algorithms.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Python&lt;/li&gt;
&lt;li&gt;SAS&lt;/li&gt;
&lt;li&gt;R&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  4. Data Science tools
&lt;/h2&gt;

&lt;p&gt;Familiarize yourself with Data science tools such as Jupyter Notebooks, Kaggle Notebooks, Google Colab and their environments for interactive coding and Git for Version Control.&lt;/p&gt;

&lt;h2&gt;
  
  
  5. Database Knowledge
&lt;/h2&gt;

&lt;p&gt;Have a sound database knowledge to deal with the structured data stored in RDBMS.&lt;br&gt;
Learn Data Manipulation Language, Data Definition language and Data Control Language. Get to know about different Database Versions and Types such as;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;MYSQL&lt;/li&gt;
&lt;li&gt;Oracle&lt;/li&gt;
&lt;li&gt;Cassandra&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  6. Data Engineering
&lt;/h2&gt;

&lt;p&gt;Master data engineering skills to clean and process a massive amount of data to avoid missing values&lt;br&gt;
Gain Data Engineering Skills such as;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Data Processing&lt;/li&gt;
&lt;li&gt;Data Wrangling&lt;/li&gt;
&lt;li&gt;SQL&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;em&gt;Take Away: Build atleast 2 Projects&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  7. Machine Learning
&lt;/h2&gt;

&lt;p&gt;Learn and implement machine learning algorithms to create predictive models.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Supervised learning &lt;/li&gt;
&lt;li&gt;Unsupervised learning &lt;/li&gt;
&lt;li&gt;Reinforcement learning&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  8. Deep Learning
&lt;/h2&gt;

&lt;p&gt;Get a thorough understanding of Deep Learning and its algorithms to work with vast volumes of unstructured data.&lt;br&gt;
Focus on;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;TensorFlow &lt;/li&gt;
&lt;li&gt;Artificial Neural Network &lt;/li&gt;
&lt;li&gt;Deep Belief Network &lt;/li&gt;
&lt;li&gt;Generative Adversarial Network&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  9.Big Data
&lt;/h2&gt;

&lt;p&gt;Get an added advantage by learning Hadoop and Spark to easily store, process and manipulate data.&lt;/p&gt;

&lt;h2&gt;
  
  
  10. Data Visualization
&lt;/h2&gt;

&lt;p&gt;Master different data visualization tools to build interactive plots and dashboards to derive business insights.&lt;br&gt;
Learn the following;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Tableau&lt;/li&gt;
&lt;li&gt;Excel&lt;/li&gt;
&lt;li&gt;QlikView&lt;/li&gt;
&lt;li&gt;Power BI&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;em&gt;Take away: Build atleast 2 projects&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  11. Build a Portfolio
&lt;/h2&gt;

&lt;p&gt;A Portfolio is a collection of projects that a professional has worked on. Data Science portfolio can help you showcase your skills and credibility, and also helps you highlight your strengths and abilities.&lt;br&gt;
The best part about building a data science portfolio is that there is no shortage of datasets that you can use to get started.&lt;br&gt;
Get a good platform to host your portfolio. Example, GitHub.&lt;/p&gt;

&lt;h2&gt;
  
  
  12.Personal attributes
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Good Communication Skills&lt;/li&gt;
&lt;li&gt;Good Story telling&lt;/li&gt;
&lt;li&gt;Persuasion skills&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  13.Get Certified
&lt;/h2&gt;

&lt;p&gt;Gives you better chances of being hired.&lt;/p&gt;

&lt;h2&gt;
  
  
  14. Apply for Jobs
&lt;/h2&gt;

&lt;p&gt;Apply for Junior Roles as  Data Scientist or any data Science Career Path of your choice depending on your abilities.&lt;br&gt;
&lt;em&gt;Happy Learning, Cheers!&lt;/em&gt;&lt;/p&gt;

</description>
      <category>lux</category>
      <category>dsea</category>
      <category>datascience</category>
      <category>dataengineering</category>
    </item>
    <item>
      <title>Introduction to Python for Data Engineering</title>
      <dc:creator>John Mambo</dc:creator>
      <pubDate>Tue, 30 Aug 2022 08:45:13 +0000</pubDate>
      <link>https://forem.com/am_mambo/introduction-to-python-for-data-engineering-180d</link>
      <guid>https://forem.com/am_mambo/introduction-to-python-for-data-engineering-180d</guid>
      <description>&lt;p&gt;&lt;strong&gt;Python&lt;/strong&gt; is a high-level, interpreted, general-purpose programming language designed by Guido van Rossum in 1991.&lt;br&gt;
&lt;strong&gt;Python&lt;/strong&gt; is Dynamically-typed and garbage-collected. Garbage collection means gaining memory back that has been &lt;u&gt;allocated and is not currently&lt;/u&gt; in use in any part of the program.&lt;br&gt;
Python also supports multiple programming paradigms including structural, object-oriented, and functional.&lt;/p&gt;

&lt;h2&gt;
  
  
  Features of Python
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Simplicity: Python Syntax is straightforward, easy to read, and code.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Portability: Python code written on Windows machines can run on other platforms such as Unix and Linux systems, and Mac too.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Easy to Debug: Simply by glancing at the code, you can determine where an error is.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;High-Level Language: python does not focus on System architecture or memory management.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Object-Oriented: Python supports Object-oriented language and the concept of classes, objects, inheritance, and encapsulation among others.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;large Standard library: Python has a huge &lt;a href="https://docs.python.org/3/library/index.html"&gt;standard library&lt;/a&gt; that provides modules and functions so that you do not have to write your code for every single thing. &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Applications of Python
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Artificial Intelligence&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Machine learning&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Science, Data Engineering, exploration, and Visualization.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Software Development&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Game Development&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Operating Systems Development&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Robotics&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Language Development&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Installing Python
&lt;/h3&gt;

&lt;p&gt;Download the latest version of Python for your operating system from &lt;a href="https://www.python.org/downloads/"&gt;Python Official Website&lt;/a&gt;. For Windows System users you can read more about setting up a python development on windows-10 from this article by &lt;a href="https://www.digitalocean.com/community/tutorials/install-python-windows-10"&gt;Digitalocean.com&lt;/a&gt;.&lt;br&gt;
If you are using Mac, you can use &lt;a href="https://docs.brew.sh/Homebrew-and-Python"&gt;brew&lt;/a&gt; and with an Ubuntu-based desktop, we would recommend using &lt;a href="https://snapcraft.io/docs"&gt;snap&lt;/a&gt;.&lt;br&gt;
To learn more about getting started with python basics, you can visit &lt;a href="https://docs.python.org/"&gt;Python Official documentation&lt;/a&gt; for more, &lt;a href="https://www.w3schools.com/python/"&gt; w3Schools&lt;/a&gt; or &lt;a href="https://dev.to/grayhat/python-101-introduction-to-python-3kg5"&gt;this blog&lt;/a&gt; among others that help beginners learn.&lt;/p&gt;

&lt;p&gt;If you are setting up an environment for Data Science or &lt;strong&gt;Data Engineering&lt;/strong&gt;, it is straightforward to get started using &lt;strong&gt;&lt;a href="https://anaconda.org/"&gt;Anaconda&lt;/a&gt;&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Engineering&lt;/strong&gt; is the art of building/architecting data platforms, designing and implementing data stores and repositories, data lakes and gathering, importing, cleaning, pre-processing, querying, analyzing data, performance monitoring, evaluation, optimization, and fine-tuning the processes and systems.&lt;/p&gt;

&lt;h3&gt;
  
  
  Critical Aspects of Data Engineering using Python
&lt;/h3&gt;

&lt;p&gt;Now that you got a brief understanding of Python and Data Engineering, we can mention some critical aspects that highlight why Python is essential in Data Engineering. Python for Data Engineering mainly comprises Data Wrangling such as reshaping, aggregating, joining sources of different formats, small-scale ETL, API interaction, and automation.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Python is Popular: Its ubiquity is one of the greatest advantages. In November 2020 it ranked second in the TIOBE Community Index and third in the 2020 Developer Survey of Stack Overflow.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Machine Learning and AI teams also use Python widely: ML, AI, and Data Engineering work closely and have to communicate the same language, Python is the most common one.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;large Standard library: A library is a collection of packages and a Package is a collection of modules. Because of Python's&lt;br&gt;
ease of use and various libraries for accessing and manipulating data and databases, it has become a popular tool to execute ETL jobs. Many teams use Python for Data Engineering rather than an ETL tool because it is more versatile and powerful for these activities.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Python is also used in technologies such as Apache Airflow and libraries for popular tools such as Apache Spark. If you intend to use these tools, it is important to know the Language you utilize.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Common Python Packages used in Data Engineering
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Pandas&lt;/strong&gt;&lt;br&gt;
Pandas is a Python Open-Source Package for manipulating and processing data frames. Pandas handle, read, aggregate, filter, reshape, and export data into various formats quickly and easily.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;SciPy&lt;/strong&gt;&lt;br&gt;
This is a module for Scientific Computing with Python. Data Engineers rely on it in carrying out computations and solving problems.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Beautiful Soup&lt;/strong&gt;&lt;br&gt;
Beautiful Soup is a library for web scraping and data mining. It provides Data Engineers with a tool to extract data from websites such as HTML pages and JSON files.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Pygrametl&lt;/strong&gt;&lt;br&gt;
Is a Python Framework that provides commonly used functionality for the development of Extract-Transform-Load (ETL) processes due to its efficiency.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Petl&lt;/strong&gt;&lt;br&gt;
Petl is a Python library for the broad purpose of extracting, manipulating, and loading data tables. It offers a broad range of functions to convert tables with few lines of code, in addition to supporting data imports from CSV, JSON, and SQL.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Advantages of using Python for Data Engineering over Java
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Ease of use&lt;/strong&gt;: Although both Python and Java are expressive, Python is more user-friendly and concise. Python helps you write short-line codes compared to Java.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Wide range of Applications&lt;/strong&gt;: Python is used in Data Science, Big Data, Data mining, Artificial Intelligence, and Machine Learning. This enables Python to be more preferred in Data Engineering than Java.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Use Cases of Python for Data Engineering
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Data Acquisition: Involves acquiring Data from APIs or through web scraping using python. ETL jobs require Python skills to use platforms such as Airflow.&lt;br&gt;
PyMoDAQ, An open-source Python-based tool is used for modular data acquisition.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Manipulation: Python for Data Engineering provides a PySpark interface that allows manipulation on large Datasets using Spark clusters. Pandas on the other hand can be used to manipulate small datasets.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Modelling: Python is a common language to use when working with teams undertaking Machine Learning, using frameworks such as Tensorflow and Pytorch.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In conclusion, Python is a key Language for Data Engineers and for those who are aspiring to become Data Engineers too. Data Engineers use Python and Python Libraries, packages, and modules in their daily routines to wrangle Data and Create Data Pipelines&lt;/p&gt;

</description>
      <category>codenewbie</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Data Engineering 101: Introduction to Data Engineering</title>
      <dc:creator>John Mambo</dc:creator>
      <pubDate>Fri, 19 Aug 2022 17:49:00 +0000</pubDate>
      <link>https://forem.com/am_mambo/data-engineering-101-introduction-to-data-engineering-357a</link>
      <guid>https://forem.com/am_mambo/data-engineering-101-introduction-to-data-engineering-357a</guid>
      <description>&lt;p&gt;With the tremendous growth of technology throughout the World, Data handling has been a big challenge especially when the data has to be moved from storage to users, from different original formats to new formats without losing its value.&lt;br&gt;
To handle this data, Data Engineering is essential.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;What is Data Engineering?&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Data Engineering is the art of building/architecting data platforms, designing and implementing data stores and repositories, data lakes and gathering, importing, cleaning, pre-processing, querying, analyzing data, performance monitoring, evaluation, optimization, and fine-tuning the processes and systems. It makes Data available for analysis and efficient Data-driven decision-making.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;What is the Role of a Data Engineer?&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;It’s the role of a data engineer to store, extract, transform, load, aggregate, and validate data. &lt;br&gt;
This involves:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Building data pipelines and efficiently storing data for tools that need to query the data. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Analyzing the data, ensuring it adheres to data governance rules and regulations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Understanding the pros and cons of data storage and query options.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Data Engineers deliver:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;The Correct data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;In the Correct Form.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;To the Right People.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;As Efficiently as Possible.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Data Engineers are Responsible for:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Ingesting Data from different Sources.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Optimizing Databases for Analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Removing Corrupted Data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Develop, Construct, test, and Maintain Data Architecture.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Why is Data Engineering important?&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;The Data Engineering lifecycle consists of building data platforms, designing and implementing Data Stores, Repos, and Data Lakes, and gathering, imparting, cleaning, preprocessing, querying, analyzing data, performance monitoring, evaluation, optimization, and tuning the System. &lt;br&gt;
Companies of all sizes have huge amounts of disparate data to comb through to answer critical business questions. Data engineering is designed to support the process, making it possible for consumers of data, such as analysts, data scientists, and executives, to reliably, quickly, and securely inspect all of the data available.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Data Engineering Tools and skills&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Data Engineers use many tools to work with Data. They use a specialized skill set to create end-to-end data pipelines that move Data from Source systems to target destinations.&lt;br&gt;
Data Engineers work with a variety of tools and technologies, such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;ETL Tools&lt;/strong&gt;: ETL (extract, transform, load) tools move data between systems. They access data, then apply rules to “transform” the data through steps that make it more suitable for analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cloud Data Storage&lt;/strong&gt;: Including Amazon S3, Azure Data Lake Storage (ADLS), Google Cloud Storage, etc.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Query Engines&lt;/strong&gt;: Engines run queries against data to return answers. Data engineers may work with engines like Dremio Sonar, Spark, Flink, and others.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Python:&lt;/strong&gt; Python is a general programming language. Data engineers may choose to use Python for ETL tasks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;SQL&lt;/strong&gt;: Structured Query Language (SQL) is the standard language for querying relational databases.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data Engineering Versus Data Science
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mqxmrc4nnd2drxsmt0i.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2mqxmrc4nnd2drxsmt0i.jpeg" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
    </item>
  </channel>
</rss>
