An incredibly fast web crawler designed for OSINT (Open Source Intelligence) data extraction.
Photon is a Python-based web crawler engineered for speed and designed specifically for OSINT (Open Source Intelligence) operations. It automates the process of scanning websites to discover and extract valuable information such as emails, URLs, API keys, subdomains, and files. It solves the problem of manual, time-consuming reconnaissance by providing a fast, automated, and highly configurable tool for security professionals and researchers.
Security researchers, penetration testers, ethical hackers, and digital investigators who need to perform thorough web reconnaissance and data gathering for audits, threat intelligence, or vulnerability assessments.
Developers choose Photon for its exceptional speed, comprehensive data extraction capabilities, and high degree of customization. Its unique selling point is being a dedicated OSINT crawler with built-in intelligence for finding sensitive data, support for plugins like Wayback machine integration, and the ability to run seamlessly via Docker.
Incredibly fast crawler designed for OSINT.
Open-Awesome is built by the community, for the community. Submit a project, suggest an awesome list, or help improve the catalog on GitHub.
Photon uses smart thread management and refined logic for top-tier crawling speed, as highlighted in the 'Genius' section of the README, ensuring efficient reconnaissance.
It extracts a wide range of data including URLs, emails, secret keys, and subdomains, specifically designed for security intelligence, with organized saving and JSON export options.
Offers extensive options like timeout delays and regex-based exclusions, allowing precise targeting and customization, as detailed in the usage wiki linked from the README.
Includes a lightweight Docker image for easy deployment and plugins for Wayback machine and DNSDumpster, enhancing functionality without complex setup.
Lacks a graphical user interface, which can be a barrier for non-technical users or those preferring visual tools for configuration and monitoring.
Admitted in the README as crawling can be resource-intensive, potentially limiting use on low-spec systems or causing performance issues during large-scale scans.
Under heavy development with frequent updates, which might introduce breaking changes or bugs, as noted in the 'Frequent & Seamless Updates' section.