8000 Create README.md · githubhosting/Python-Automation@eadb93b · GitHub
[go: up one dir, main page]

Skip to content
8000

Commit eadb93b

Browse files
Create README.md
1 parent 96d103e commit eadb93b

File tree

1 file changed

+27
-0
lines changed

1 file changed

+27
-0
lines changed

README.md

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,27 @@
1+
# Scrapping Drop
2+
3+
## Modules used:
4+
5+
### MechanicalSoup:
6+
A Python library for automating interaction with websites. MechanicalSoup automatically stores and sends cookies, follows redirects, and can follow links and submit forms.
7+
MechanicalSoup provides a similar API, built on Python giants Requests (for HTTP sessions) and BeautifulSoup (for document navigation).
8+
9+
### os:
10+
Python OS module provides the facility to establish the interaction between the user and the operating system. It offers many useful OS functions that are used to perform OS-based tasks and get related information about the operating system.
11+
It is possible to automatically perform many operating system tasks. The OS module in Python provides functions for creating and removing a directory (folder), fetching its contents, changing and identifying the current directory, etc.
12+
13+
### wget:
14+
Wget is a convenient solution for downloading files over HTTP and FTP protocols. It works well with Python in recursively downloading multiple files, and the process can easily be automated to save you time.Using the proper parameters, Wget can operate as a web crawler. Instead of downloading a single file, it can recursively download files linked from a specific web page until all the links have been exhausted or until it reaches a user-specified recursion depth. In this scenario, Wget saves the downloaded files in a directory structure that resembles the server they have been downloaded from.
15+
16+
# Working:
17+
* First the program takes the input from the user in the console
18+
* Then it stores it in search_item
19+
* Now the Search_item is queried to the browser with the help of the Mechanical Soup Module
20+
* The browser will provide the information and appropriate images with respect to the searched item
21+
* Mechanical Soup module will scrap the HTML code from the results
22+
* That code is filtered with the image links followed by the tag “img”
23+
* The links of the images are stored in the list image_source
24+
* User is promoted to enter the image format that needs to downloaded
25+
* os module uses the given input format and downloads in that format
26+
* Now os module is used to create a folder to store the downloaded images
27+
* wget module downloads all the image links one by one in the desired path

0 commit comments

Comments
 (0)
0