Skip to content

zegron/WebCrawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

3 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Website Crawler (GUI Version)

A simple Python-based website crawler that lets you input a URL, crawl all internal links on that domain, display results in a GUI, and export them to a CSV file.

This project uses Tkinter for the interface and BeautifulSoup + Requests for web crawling.


πŸͺŸ Features

  • πŸ–₯️ Clean GUI built with Tkinter
  • 🌐 Crawls all internal links within the same domain
  • πŸ“Š Progress bar and live status updates while crawling
  • πŸ’Ύ Saves all discovered URLs to links.csv in the project folder
  • βš™οΈ Can be packaged into a standalone Windows .exe with PyInstaller

πŸš€ How to Run

  1. Install dependencies:
    pip install requests beautifulsoup4

Run the app:

bash Copy code python site_crawler_gui.py Enter any website URL (e.g., https://example.com) and click Start Crawling.

All discovered links will appear in the results box and be saved to links.csv.

🧱 Build a Windows App (Optional) To build an .exe version (no console window):

bash Copy code pyinstaller --noconsole --onefile "site_crawler_gui.py" You’ll find the executable in the dist/ folder.

πŸ“‚ Project Structure bash Copy code WebCrawler/ β”‚ β”œβ”€β”€ site_crawler_gui.py # Main application file β”œβ”€β”€ links.csv # Generated after crawling β”œβ”€β”€ README.md # Project overview (this file) └── CHANGELOG.md # Version history and updates πŸ§‘β€πŸ’» Requirements Python 3.9+

Packages: requests, beautifulsoup4

About

Simple Website Crawler - Command Line and GUI Version for WIN10/11

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages