How To Perform Web Scraping Using Python Program

Introduction 

 
So, first, let us all know what Web Scrapping is.
 
Web Scraping (also known as Screen Scraping, Web Data Extraction, Web Harvesting, etc.) is a process to scrap (extract) data from web pages, web document, HTML document or XML document in a structured manner and then store in the database, text file as a spreadsheet, as an XML, or in whatever file format we want to save in.
Python Libraries we will be using here -
  1. requests
  2. BeautifulSoup
Let us know in short about these two libraries.
 
Requests
 
Basically, Requests is an Apache2 Licensed HTTP library written in Python. It allows us to send HTTP/1.1 requests. With it, we can add content like headers, form data, multipart files, and parameters. It allows us to access the response data of Python.
 
Google it to know more.
 
BeautifulSoup
 
In short, BeautifulSoup is a library that is written in Python for extracting data from HTML and XML files. This Python library is mostly used while scrapping data from HTML and XML.
 
Google it to know more.
 
Commands to install these two libraries.
 
$ pip install requests==2.18.4
 
$ pip install beautifulsoup4==4.6.0
 
Now, let us name the Python web scrapping program.
 
Program Name - web_scrap.py
 
First, let us know what web_scrap.py program will do. We will be creating this Python program to scrap recent posts from the C# Corner website, i.e., http://www.c-sharpcorner.com.
 
Full Code Snippet to scrap recent posts from c-sharpcorner. 
  1. import requests  
  2. from bs4 import BeautifulSoup  
  3. result = requests.get('http://www.c-sharpcorner.com/')  
  4. print (result.status_code)  
  5. print (result.headers)  
  6. csharpcorner = result.content  
  7. scrap = BeautifulSoup(csharpcorner,'html.parser')  
  8. post = scrap.find_all("div",attrs={'class':'post'})  
  9. for posts in post:  
  10.     postBody = posts.find_all('div',attrs={'class':'media-body'})  
  11.     for i in range(len(postBody)):  
  12.       print postBody[i].find('a', attrs={'class':'title'}).string.strip()  
  13.       print 'Category: ',postBody[i].find('a''').string.strip()  
  14.       print 'By: ',postBody[i].find('a', attrs={'class':'author'}).string.strip()  
  15.       print 'Date and Time: ',postBody[i].find('span','hidden-xs').string.strip()  
  16.       print '----------------------------------------------------------------------------------------------'  
Now, let’s understand the code. Understanding the above code is pretty simple if we break the code in some parts and then try to run the program. In this way, we can also understand the large code programs, just by breaking the code in some part and then try to run.
 
Let us start to understand the code.
  1. To use libraries, we need to import it then we can use it. That’s the reason we have first imported requests and BeautifulSoup.
     
  2. We created a variable result and we are making a GET request to the C-SharpCorner website, using the help of python requests library and we store the response to the variable result.
     
  3. Now, by using the python print function we are checking the response code, to check whether we get a response from the website or not. Response Code 200 means OK, we get a response from c-sharpcorner.com
     
  4. Now, using the print function we are printing out the response header.
     
    What is the response header?
     
    The information, in the form of a text record, that a Web server sends back to a client's browser in response to receiving an HTTP request. The response header contains the date, size, and type of file that the server is sending back to the client and also data about the server itself. The header is attached to the files being sent back to the client.
  1. Now we have created a variable name c-sharpcorner and we are storing the whole website content into it. You can check it if you print out c-sharpcorner variable.
     
  2. Now, we have created a variable name scrap that holds the full content of c-sharpcorner and we have defined HTML.parser because we know BeautifulSoup can parse HTML and XML file.
     
  3. Now we have created a variable name post and by using the BeautifulSoup function find_all we are trying to find all the div on the website but with the class name post. And inside our class name "post", Recent Post resides.
     
  4. Now our variable post stores an array of Recent Post from c-sharpcorner. If you print out variable post you will come to know.
     
  5. Now we write a for loop to loop inside the Recent Post and print out the recent post, but we don’t know how many recent posts are there on the website.
     
  6. So to know this, we created a variable name postBody which targets the div with class name media-body ( which has recent post details like title, author, category, date and time).
     
  7. Now we write another for loop and we find the length of recent posts and targets all the class, id, HTML tag, where our desired data is and in this way we print out all the Recent Posts from c-sharpcorner.
     
  8. This is how we did our web scrapping in a nested loop.
Now, let us run the program using the command:
 
$ python web_scrap.py 
 
Python
 
I hope you guys like this tutorial.
 
In this way, we can target HTML tag, class, id and scrap any part of data from the HTML web page.
 
You all can also store the scraped data into a file instead of printing it out in the terminal.
 
Want to know how to store data in a file using Python, comment down below.
 
Thank You!!


Similar Articles