✂️🕷️ Spider-Cut is a Network Mapper Framework (NMAP Framework)

Overview

Build with Python Latest Release Last Comitment Python Version


Spider-Cut is a Network Mapper Framework (NMAP Framework)

Installation    |    Usage    |    Creators    |    Donate

Installation

# Kali Linux | WSL

# clone the repo
$ git clone https://github.com/XFORWORKS/SpiderCut

# change the working directory to SpiderCut
$ cd SpiderCut
  
# install the requirements  
$ python -m pip install -r requirements.txt

# run the bigger installation process
$ python setup.py

Usage

# Kali Linux | WSL

$ python spidercut.py --help
usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                              [-ex] [-a]

Spider-Cut is a Network Mapper Framework (NMAP Framework)

optional arguments:
  -h, --help       show this help message and exit
  -v, --version    shows the current Version of Spider-Cut
  -s, --single     scan a single Target
  -m, --multiple   scan multiple Targets
  -rH, --rhost     scan a range of Hosts
  -sn, --subnet    scan an entire Subnet
  -rh, --ranhost   scan random Hosts
  -eX, --exclude   Excluding Targets from a scan
  -ex, --excllist  Excluding Targets using a list
  -a, --agscan     perform an Aggressive scan

Created By XFORWORKS (Retr0 & Ctript0)

To run the script:

# Kali Linux | WSL

$ python spidercut.py

Creators

The Script is designed by Cript0 (like the banners, the colors, etc...)

Everything other is from Retr0 (like the commands, the framework idea, etc...)

Donate

PayPal

Releases

Latest Release

You might also like...
Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Django and Vue.js

Gerapy Distributed Crawler Management Framework Based on Scrapy, Scrapyd, Scrapyd-Client, Scrapyd-API, Django and Vue.js. Documentation Documentation

Async Python 3.6+ web scraping micro-framework based on asyncio
Async Python 3.6+ web scraping micro-framework based on asyncio

Ruia 🕸️ Async Python 3.6+ web scraping micro-framework based on asyncio. ⚡ Write less, run faster. Overview Ruia is an async web scraping micro-frame

A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Transistor, a Python web scraping framework for intelligent use cases.
Transistor, a Python web scraping framework for intelligent use cases.

Web data collection and storage for intelligent use cases. transistor About The web is full of data. Transistor is a web scraping framework for collec

Web crawling framework  based on asyncio.
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

PyQuery-based scraping micro-framework.

demiurge PyQuery-based scraping micro-framework. Supports Python 2.x and 3.x. Documentation: http://demiurge.readthedocs.org Installing demiurge $ pip

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo.

Crawler do site Fundamentus.com com o uso do framework scrapy, tanto da aba detalhada como a de resumo. (Todas as infomações)

A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Python framework to scrape Pastebin pastes and analyze them
Python framework to scrape Pastebin pastes and analyze them

pastepwn - Paste-Scraping Python Framework Pastebin is a very helpful tool to store or rather share ascii encoded data online. In the world of OSINT,

Releases(1.0)
  • 1.0(Dec 25, 2021)

    Build with Python Last Comitment Python Version


    Spider-Cut is a Network Mapper Framework (NMAP Framework)

    Installation

    # Kali Linux | WSL
    
    # clone the repo
    $ git clone https://github.com/XFORWORKS/SpiderCut
    
    # change the working directory to SpiderCut
    $ cd SpiderCut
      
    # install the requirements  
    $ python -m pip install -r requirements.txt
    
    # run the bigger installation process
    $ python setup.py
    

    Usage

    # Kali Linux | WSL
    
    $ python spidercut.py --help
    usage: spidercut.py [-h] [-v] [-s] [-m] [-rH] [-sn] [-rh] [-eX]
                                  [-ex] [-a]
    
    Spider-Cut is a Network Mapper Framework (NMAP Framework)
    
    optional arguments:
      -h, --help       show this help message and exit
      -v, --version    shows the current Version of Spider-Cut
      -s, --single     scan a single Target
      -m, --multiple   scan multiple Targets
      -rH, --rhost     scan a range of Hosts
      -sn, --subnet    scan an entire Subnet
      -rh, --ranhost   scan random Hosts
      -eX, --exclude   Excluding Targets from a scan
      -ex, --excllist  Excluding Targets using a list
      -a, --agscan     perform an Aggressive scan
    
    Created By XFORWORKS (Retr0 & Cript0)
    

    To run the script:

    # Kali Linux | WSL
    
    $ python spidercut.py
    

    Creators

    The Script is designed by Cript0 (like the banners, the colors, etc...)

    Everything other is from Retr0 (like the commands, the framework idea, etc...)

    Donate

    PayPal

    Source code(tar.gz)
    Source code(zip)
    SpiderCut.rar(171.16 KB)
    SpiderCut.zip(226.92 KB)
Owner
XforWorks
Nothing to see here...
XforWorks
Subscrape - A Python scraper for substrate chains

subscrape A Python scraper for substrate chains that uses Subscan. Usage copy co

ChaosDAO 14 Dec 15, 2022
A simple code to fetch comments below an Instagram post and save them to a csv file

fetch_comments A simple code to fetch comments below an Instagram post and save them to a csv file usage First you have to enter your username and pas

2 Jul 14, 2022
Scrape puzzle scrambles from csTimer.net

Scroodle Selenium script to scrape scrambles from csTimer.net csTimer runs locally in your browser, so this doesn't strain the servers any more than i

Jason Nguyen 1 Oct 29, 2021
淘宝、天猫半价抢购,抢电视、抢茅台,干死黄牛党

taobao_seckill 淘宝、天猫半价抢购,抢电视、抢茅台,干死黄牛党 依赖 安装chrome浏览器,根据浏览器的版本找到对应的chromedriver下载安装 web版使用说明 1、抢购前需要校准本地时间,然后把需要抢购的商品加入购物车 2、如果要打包成可执行文件,可使用pyinstalle

2k Jan 05, 2023
A distributed crawler for weibo, building with celery and requests.

A distributed crawler for weibo, building with celery and requests.

SpiderClub 4.8k Jan 03, 2023
Tool to scan for secret files on HTTP servers

snallygaster Finds file leaks and other security problems on HTTP servers. what? snallygaster is a tool that looks for files accessible on web servers

Hanno Böck 2k Dec 28, 2022
Pro Football Reference Game Data Webscraper

Pro Football Reference Game Data Webscraper Code Copyright Yeetzsche This is a simple Pro Football Reference Webscraper that can either collect all ga

6 Dec 21, 2022
Libextract: extract data from websites

Libextract is a statistics-enabled data extraction library that works on HTML and XML documents and written in Python

499 Dec 09, 2022
自动完成每日体温上报(Github Actions)

体温上报助手 简介 每天 10:30 GMT+8 自动完成体温上报,如想修改定时运行的时间,可修改 .github/workflows/SduHealthReport.yml 中 schedule 属性。 如果当日有异常,请手动在小程序端/PC 端填写!

Teng Zhang 23 Sep 15, 2022
Auto Join: A GitHub action script to automatically invite everyone to the organization who star your repository.

Auto Invite To The Organization By Star A GitHub Action script to automatically invite everyone to your organization that stars your repository. What

Max Base 11 Dec 11, 2022
Web scrapping tool written in python3, using regex, to get CVEs, Source and URLs.

searchcve Web scrapping tool written in python3, using regex, to get CVEs, Source and URLs. Generates a CSV file in the current directory. Uses the NI

32 Oct 10, 2022
This tool crawls a list of websites and download all PDF and office documents

This tool crawls a list of websites and download all PDF and office documents. Then it analyses the PDF documents and tries to detect accessibility issues.

AccessibilityLU 7 Sep 30, 2022
robobrowser - A simple, Pythonic library for browsing the web without a standalone web browser.

RoboBrowser: Your friendly neighborhood web scraper Homepage: http://robobrowser.readthedocs.org/ RoboBrowser is a simple, Pythonic library for browsi

Joshua Carp 3.7k Dec 27, 2022
A simple, configurable and expandable combined shop scraper to minimize the costs of ordering several items

combined-shop-scraper A simple, configurable and expandable combined shop scraper to minimize the costs of ordering several items. Features Define an

2 Dec 13, 2021
Github scraper app is used to scrape data for a specific user profile created using streamlit and BeautifulSoup python packages

Github Scraper Github scraper app is used to scrape data for a specific user profile. Github scraper app gets a github profile name and check whether

Siva Prakash 6 Apr 05, 2022
Grab the changelog from releases on Github

release-notes-scraper This simple script can be used to grab the release notes for projects from github that do not keep a CHANGELOG, but publish thei

Dan Čermák 4 Apr 01, 2022
Linkedin webscraping - Linkedin web scraping with python

linkedin_webscraping This is the first step of a full project called "LinkedIn J

Pedro Dib 4 Apr 24, 2022
Scraping script for stats on covid19 pandemic status in Chiba prefecture, Japan

About 千葉県の地域別の詳細感染者統計(Excelファイル) をCSVに変換し、かつ地域別の日時感染者集計値を出力するスクリプトです。 Requirement POSIX互換なシェル, e.g. GNU Bash (1) curl (1) python = 3.8 pandas = 1.1.

Conv4Japan 1 Nov 29, 2021
京东秒杀商品抢购Python脚本

Jd_Seckill 非常感谢原作者 https://github.com/zhou-xiaojun/jd_mask 提供的代码 也非常感谢 https://github.com/wlwwu/jd_maotai 进行的优化 主要功能 登陆京东商城(www.jd.com) cookies登录 (需要自

Andy Zou 1.5k Jan 03, 2023
ChromiumJniGenerator - Jni Generator module extracted from Chromium project

ChromiumJniGenerator - Jni Generator module extracted from Chromium project

allenxuan 4 Jun 12, 2022