<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Muchamad Faiz</title>
    <description>The latest articles on Forem by Muchamad Faiz (@muchamadfaiz).</description>
    <link>https://forem.com/muchamadfaiz</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F967443%2F78ec4bd7-884f-4bbc-986a-80bc702a9611.jpg</url>
      <title>Forem: Muchamad Faiz</title>
      <link>https://forem.com/muchamadfaiz</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/muchamadfaiz"/>
    <language>en</language>
    <item>
      <title>Scrapping Top Repositories for GitHub Topics</title>
      <dc:creator>Muchamad Faiz</dc:creator>
      <pubDate>Sun, 18 Dec 2022 23:08:09 +0000</pubDate>
      <link>https://forem.com/zettasoft/scrapping-top-repositories-for-github-topics-21l9</link>
      <guid>https://forem.com/zettasoft/scrapping-top-repositories-for-github-topics-21l9</guid>
      <description>&lt;p&gt;GitHub is a popular website for sharing open source projects and code repositories. For example, the tensorflow repository contains the entire source code of the Tensorflow deep learning framework.&lt;/p&gt;

&lt;p&gt;Repositories in GitHub can be tagged using topics. For example, the tensorflow repository has the topics python, machine-learning, deep-learning etc.&lt;/p&gt;

&lt;p&gt;The page &lt;a href="https://github.com/topics" rel="noopener noreferrer"&gt;https://github.com/topics&lt;/a&gt; provides a list of the top topics on Github. In this project, we'll retrive information from this page using web scraping: the process of extracting information from a website in an automated fashion using code. We'll use the Python libraries Requests and Beautiful Soup to scrape data from this page.&lt;/p&gt;

&lt;h2&gt;
  
  
  Project Outline
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;1. We're going to scrape &lt;a href="https://github.com/topics" rel="noopener noreferrer"&gt;https://github.com/topics&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;2. We'll get a list of topics. For each topic, we'll get topic title, topic page URL and topic description&lt;/li&gt;
&lt;li&gt;3. For each topic, we'll get the top 25 repositories in the topic from the topic page&lt;/li&gt;
&lt;li&gt;4. For each repository, we'll grab the repo name, username, stars and repo URL&lt;/li&gt;
&lt;li&gt;5. By the end of the project, we'll create a CSV and XLX file in the following format:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy80s9nc99b51mdfwuzsj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy80s9nc99b51mdfwuzsj.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Instal and import all library needed
&lt;/h3&gt;

&lt;p&gt;Before we begin lets install the library with pip&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install requests
pip install beautifulsoup4 
pip install pandas
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;then, lets import all the library into code editor&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import requests as req
from bs4 import BeautifulSoup
import pandas as pd
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Write a function to download the page
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def get_topic_link(base_url):
    response = req.get(url=base_url)
    page_content = response.text
    print(response.status_code)

    soup = BeautifulSoup(page_content, "html.parser")
    tags = soup.find_all("div", class_ = "py-4 border-bottom d-flex flex-justify-between")
    topic_links = []
    for tag in tags:
        url_end = tag.find("a")["href"]
        topic_link = f"https://www.github.com{url_end}"
        topic_links.append(topic_link)
    return topic_links
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  write a function to extract information
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def get_info_topic(topic_link):
    response1 = req.get(topic_link)
    topic_soup = BeautifulSoup(response1.text, "html.parser")
    topic_tag = topic_soup.find("h1").text.strip() #3D
    topic_desc = topic_soup.find("p").text
    info_topic = {
        "title" : topic_tag,
        "desc" : topic_desc
    }
    return info_topic


def get_info_tags(topic_link):
    response = req.get(topic_link)
    info_soup = BeautifulSoup(response.text, "html.parser")
    repo_tags = info_soup.find_all("div", class_ = "d-flex flex-justify-between flex-items-start flex-wrap gap-2 my-3")
    return repo_tags


def get_info(tag):
    repo_username = tag.find_all("a")[0]
    repo_name = tag.find_all("a")[1]

    url_end = tag.find("a")["href"]
    repo_url = f"https://www.github.com{url_end}"

    repo_star = tag.find("span",{"id":"repo-stars-counter-star"}).text.strip()
    repo_value = int(float(repo_star[:-1]) * 1000)

    topics_data = {
        "repo_name" : "repo_name",
        "repo_username" : repo_username.text.strip(),
        "repo_name" : repo_name.text.strip(),
        "repo_url" : repo_url,
        "repo_star" : repo_value,
        }
    return topics_data
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Create CSV file(s) with the extracted information
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def save_CSV(results):
    df = pd.DataFrame(results)
    df.to_csv("github.csv", index=False)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Create XLX file(s) with the extracted information
&lt;/h2&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def save_XLX(results):
    df = pd.DataFrame(results)
    df.to_excel("github.xlsx", index=False)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Putting it all together
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;we have a function to get the list of topics&lt;/li&gt;
&lt;li&gt;we have a function to create a CSV file for scraped repos from a topics page&lt;/li&gt;
&lt;li&gt;Let's create a function to put them together
&lt;/li&gt;
&lt;/ul&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def main():
    base_url = "https://github.com/topics"
    topic_links = get_topic_link(base_url) # list of url ex: https://github.com/topics/3d, https://github.com/topics/AJAX, etc
    result2 = []
    for topic_link in topic_links: # https://github.com/topics/3d
        print(f"getting info {topic_link}")
        topic_tags = get_info_topic(topic_link) #title, desc
        repo_tags = get_info_tags(topic_link) # some repo tags, so we can use for loop
        result1 = []
        for tag in repo_tags:
            repo_info = get_info(tag)
            result1.append(repo_info)
        for x in result1:
            gabungan = topic_tags | x
            result2.append(gabungan)
        save_CSV(result2)
        save_XLX(result2)


if __name__ == "__main__":
    main()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;We are done here, i hope this simple project can be valuable to your practice in python web scrapping. &lt;/p&gt;

&lt;p&gt;Github :&lt;a href="https://github.com/muchamadfaiz" rel="noopener noreferrer"&gt;https://github.com/muchamadfaiz&lt;/a&gt;&lt;br&gt;
Email : &lt;a href="mailto:muchamadfaiz@gmail.com"&gt;muchamadfaiz@gmail.com&lt;/a&gt;&lt;/p&gt;

</description>
      <category>python</category>
      <category>tutorial</category>
      <category>beginners</category>
    </item>
    <item>
      <title>How to manage browser driver with Web Driver Manager</title>
      <dc:creator>Muchamad Faiz</dc:creator>
      <pubDate>Thu, 15 Dec 2022 21:09:28 +0000</pubDate>
      <link>https://forem.com/zettasoft/how-to-manage-browser-driver-with-web-driver-manager-3k53</link>
      <guid>https://forem.com/zettasoft/how-to-manage-browser-driver-with-web-driver-manager-3k53</guid>
      <description>&lt;p&gt;In this article i will explain how to use Web Driver Manager to manage browser driver easily&lt;/p&gt;

&lt;h2&gt;
  
  
  Background
&lt;/h2&gt;

&lt;p&gt;when we make a project using selenium library every time there is a newer version of the driver we need to download it and we have to set the corresponding executable path explicitly. After that, we make object from the driver and continue with the code we want to run. these steps become complicated as we need to repeat these steps out every time the versions change.So therefore, we use WebDriverManager to easily manage it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Task
&lt;/h2&gt;

&lt;p&gt;Create a simple session to open google.com and search with keyword "web scrapping"&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started
&lt;/h2&gt;

&lt;p&gt;assuming python is intalled on your machine or virtual environment, then you must install WebDriverManager and selenium&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install webdriver-manager
pip install selenium
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;after that on code editor lets import all library we need&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from selenium import webdriver
from selenium.webdriver.common.by import By
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.chrome.service import Service as ChromeService
from webdriver_manager.chrome import ChromeDriverManager
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;next we start the session by instatiating driver object&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;driver = webdriver.Chrome(options=Options, service=ChromeService(ChromeDriverManager().install()))
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;the script above will install driver browser automatically, &lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvf0g46gd0akvsup3m6wd.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvf0g46gd0akvsup3m6wd.jpg" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
then, lets call the driver object to take action on google.com&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;driver.get("https:www.google.com")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;and lets find element search and send key "web scraping" on it&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;search_el = driver.find_element(By.CSS_SELECTOR,"input[title='Search']")
search_el.send_keys("web scrapping")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;if you have arrived at this step, then your screen will be the same as this image&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7brlrtd5j2kom6pchlii.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7brlrtd5j2kom6pchlii.jpg" alt="Web Scraping"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;now we know that WebDriverManager automates the browser setup in the Selenium code and this make the difficult steps to store newer version of driver become automatic so we can focus on our execution script rather than browser settings&lt;/p&gt;

&lt;p&gt;Github :&lt;a href="https://github.com/muchamadfaiz" rel="noopener noreferrer"&gt;https://github.com/muchamadfaiz&lt;/a&gt;&lt;br&gt;
Email : &lt;a href="mailto:muchamadfaiz@gmail.com"&gt;muchamadfaiz@gmail.com&lt;/a&gt;&lt;/p&gt;

</description>
      <category>programming</category>
      <category>beginners</category>
      <category>tutorial</category>
      <category>python</category>
    </item>
    <item>
      <title>Google Image Scraping Using Selenium - Part 1</title>
      <dc:creator>Muchamad Faiz</dc:creator>
      <pubDate>Tue, 13 Dec 2022 09:35:06 +0000</pubDate>
      <link>https://forem.com/zettasoft/google-image-scraping-using-selenium-part-1-1nl7</link>
      <guid>https://forem.com/zettasoft/google-image-scraping-using-selenium-part-1-1nl7</guid>
      <description>&lt;p&gt;If you wanna learn automation scrapping with selenium, then this simple project can be the starting point of your journey. In this tutorial i will explain how to scrape image from google using selenium.&lt;/p&gt;

&lt;h2&gt;
  
  
  Background
&lt;/h2&gt;

&lt;p&gt;on my case i want to scrape image on google with some keyword, lets say "cat" then i will store some links as csv files.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting Started
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What is Selenium
&lt;/h3&gt;

&lt;p&gt;before we go any further we must know what is selenium. Selenium is a tool for controlling web browsers through programs and performing browser automation. It is mainly used as a testing framework for cross-web browser platforms. However, Selenium is also a very capable tool to use for general web automation, as we are able to program it to do what a human user can do on a browser (in this case, to programmatically download images from Google).&lt;/p&gt;

&lt;h3&gt;
  
  
  Scraping with Selenium
&lt;/h3&gt;

&lt;p&gt;So how does Selenium exactly work? well, Selenium provides the mechanisms to locate elements on a web page and it mimic the user behaviour. here is the table for most used attribute and locator&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Ch5FrzJH--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/17xhe3nt2a2kxqb39uee.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Ch5FrzJH--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/17xhe3nt2a2kxqb39uee.jpg" alt="test" width="184" height="121"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These elements can be found in feature Developer Tools on web browsers&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s---dbpyfNk--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/74m2kjk5ozctooxo6kkr.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s---dbpyfNk--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/74m2kjk5ozctooxo6kkr.jpg" alt="Developer Tools" width="880" height="365"&gt;&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;and now lets start coding!&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Set up the necessary libraries required for the script
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install selenium
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;Import Libraries
for this tutorial i will be using google chrome so,
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;from selenium.webdriver import Chrome
from selenium.webdriver.common.by import By
from selenium.webdriver.common.keys import Keys
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;go get into google.com then search with keyword "cat"
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;driver = Chrome()
driver.get("https://www.google.com/")
search_el = driver.find_element(By.XPATH, "//input[@title='Search']")
search_el.send_keys("cat")
search_el.send_keys("Keys.ENTER")
image_el = driver.find_element(By.XPATH, "//a[href]")
image_el.click()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;the google window will pop up with cat image&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--N8k6jjeA--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/pnycc2zam9v6606f2sk1.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--N8k6jjeA--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/pnycc2zam9v6606f2sk1.jpg" alt="test2" width="880" height="367"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Congratulations! You have successfully open a browser and and navigate to cat images automatically, next we will scroll the page and extract the url image. i will cover it in next part&lt;/p&gt;

&lt;p&gt;Github :&lt;a href="https://github.com/muchamadfaiz"&gt;https://github.com/muchamadfaiz&lt;/a&gt;&lt;br&gt;
Email : &lt;a href="mailto:muchamadfaiz@gmail.com"&gt;muchamadfaiz@gmail.com&lt;/a&gt;&lt;/p&gt;

</description>
      <category>python</category>
      <category>scrapping</category>
      <category>tutorial</category>
      <category>beginners</category>
    </item>
    <item>
      <title>This 5 Reasons of Web Scraping May Benefit Your Business</title>
      <dc:creator>Muchamad Faiz</dc:creator>
      <pubDate>Thu, 10 Nov 2022 04:59:18 +0000</pubDate>
      <link>https://forem.com/zettasoft/this-5-reasons-of-web-scraping-may-benefit-your-business-3e4h</link>
      <guid>https://forem.com/zettasoft/this-5-reasons-of-web-scraping-may-benefit-your-business-3e4h</guid>
      <description>&lt;p&gt;As the digital economy expands, the role of web scraping becomes ever more important. In this article i will explain how web scraping can benefit your business&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Real Estate Listing Scraping
&lt;/h3&gt;

&lt;p&gt;In this era, the housing industry more like a complicated system of housing sites having big datas about travel destinations, reviews of places to stay, comments, and user profiles.&lt;br&gt;
Many real estate agents use web scraping to populate their database of available properties for sale or for rent.&lt;/p&gt;

&lt;p&gt;For example, a real estate agency will scrape MLS (private databases that are created, maintained and paid for by real estate professionals to help their clients buy and sell property) listings to build an API that directly populates this information onto their website. This way, they get to act as the agent for the property when someone finds this listing on their site.&lt;/p&gt;

&lt;p&gt;Most listings that you will find on a Real Estate website are automatically generated by an API.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Shopping Sites Comparison
&lt;/h3&gt;

&lt;p&gt;Some websites and applications can help you to easily compare pricing between several retailers for the same product.&lt;/p&gt;

&lt;p&gt;One way that these websites work is by using web scrapers to scrape product data and pricing from each retailer daily. This way, they can provide their users with the comparison data they need.&lt;/p&gt;

&lt;p&gt;High-quality web scraped data obtained in large volumes can be very helpful in analyzing the product trends and forecasting the price.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Lead Generation
&lt;/h3&gt;

&lt;p&gt;web scraping is used by many companies to collect contact information about potential customers or clients. This is incredibly common in the business-to-business space, where potential customers will post their business information publicly online. You can take your target persona: education, company, job title, etc. and then find relevant websites on your niche: physicians on health care providers; plumbers, drycleaner, restaurant, etc from Yell.com; KOL (key opinion leader) from big-leap start-up companies&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Investment Decisions
&lt;/h3&gt;

&lt;p&gt;Investment decisions are complex, as it usually involves a series of process before a sound decision can be made from setting up a hypothetical thesis, experimenting, to researching. The most effective way to test an investment thesis is through historical data analysis. It allows you to gain insights into the root cause of past failures or successes, pitfalls you should have avoided, and future investment returns you might gain.&lt;/p&gt;

&lt;p&gt;However, web scraping does not guarantee improved returns on its own. With some financial knowledge and a little for data gathering and analysis, it can prove to be an invaluable tool in today’s information-driven economy.&lt;/p&gt;

&lt;p&gt;Web data has no limitation. It’s ever-growing and full of information that can influence the market. Knowing how to harness the power of a web scraping tool can pave the way to better returns. &lt;/p&gt;

&lt;h3&gt;
  
  
  5. Product Optimization
&lt;/h3&gt;

&lt;p&gt;Suppose you want to collect customers’ feedback to cross-examine your product and make improvements. The sentiment analysis technique is widely used to analyze customers’ attitudes, whether it is positive, neutral or negative. However, the analysis needs a considerable amount of data. Web scraping can automate the extraction process faster which saves tons of time and effort for such work.&lt;/p&gt;

&lt;p&gt;That’s all, if you have any comment or question please dont mind me.&lt;/p&gt;

</description>
      <category>python</category>
    </item>
  </channel>
</rss>
