薅薅乐 - JD 测试脚本

Overview

薅薅乐

安裝

使用docker

  • docker一键安装: docker run -d --name jd classmatelin/hhl:latest.

使用

  • 进入容器: docker exec -it jd bash

  • 获取JD_COOKIES: python get_jd_cookies.py, 扫描登录成功后控制台会打印JD_COOKIES.

  • vim /scripts/conf/config.yaml, 填入上一步获取的JD_COOKIES。

    debug: true
    
    # JD_COOKIES配置, 一行一个, -符号是必须的。
    jd_cookies: 
      - pt_pin=jd_78b;pt_key=AAJgyqEMOsFQr5a0ucVzJepxU;
    
    # 此处省略更多配置
  • 配置好JD_COOKIES, 随便运行一个脚本检查配置, 如: python update_share_code.py.

  • 内置定时任务, 配好JD_COOKIES自动会执行脚本。

  • 消息通知支持TG机器人, 需要在config.yaml文件中填写 tg_bot_token和tg_user_id配置.

脚本列表

脚本名称 脚本描述 完成进度
jd_big_winner.py 京东极速版->大赢家翻翻乐 100%
update_share_code.py 更新助力码 100%
jd_bean_change.py 资产变动通知 100%
get_jd_cookies.py 获取京东的COOKIES 100%
jd_cute_pet.py 京东APP->东东萌宠 100%
jd_factory.py 京东APP->东东工厂 100%
jd_farm.py 京东APP->东东农场 100%
jd_lucky_turntable.py 幸运大转盘 100%
jd_pet_dog 京东APP-> 宠汪汪 10%
jd_planting_bean.py 京东APP->种豆得豆 100%
jd_ranking_list 京东APP->排行榜 100%
jd_shark_bean 京东APP->摇金豆 100%
jd_sign_collection 京东签到合集 100%
jr_daily_task_goose.py 京东金融->天天提鹅 100%
jr_pet_pig 京东金融->养猪猪 100%
jr_money_tree 京东金融->摇钱树 100%
jd_bean_lottery.py 京东APP->签到领京豆->摇金豆->京豆夺宝 100%
jd_earn_bean.py 微信小程序-赚金豆 100%
jd_cash.py 京东APP-领现金 100%
jd_burning_summer.py 京东APP-燃动夏季 100%
jd_wishing_pool.py 京东APP-京东众筹-许愿池 100%
jd_good_morning.py 京东APP-领金豆->早起福利 100%
jd_gold_creator.py 京东APP-京东排行榜->金榜创造营 100%
jd_beauty.py 京东APP->美丽研究院 0%
jx_factory.py 京喜APP->京喜工厂 0%
jx_farm.py 京喜APP->京喜农场 0%
jd_esports_manager.py 京东APP->电竞经理 0%

配置说明

  • 请查看/scripts/conf/config.yaml中的配置项。

特别声明:

  • 本仓库发布的jd_scripts项目中涉及的任何解锁和解密分析脚本,仅用于测试和学习研究,禁止用于商业用途,不能保证其合法性,准确性,完整性和有效性,请根据情况自行判断.

  • 本项目内所有资源文件,禁止任何公众号、自媒体进行任何形式的转载、发布。

  • ClassmateLin对任何脚本问题概不负责,包括但不限于由任何脚本错误导致的任何损失或损害.

  • 间接使用脚本的任何用户,包括但不限于建立VPS或在某些行为违反国家/地区法律或相关法规的情况下进行传播, ClassmateLin对于由此引起的任何隐私泄漏或其他后果概不负责.

  • 请勿将jd_scripts项目的任何内容用于商业或非法目的,否则后果自负.

  • 如果任何单位或个人认为该项目的脚本可能涉嫌侵犯其权利,则应及时通知并提供身份证明,所有权证明,本人将在收到认证文件后删除相关脚本.

  • 任何以任何方式查看此项目的人或直接或间接使用该jd_scripts项目的任何脚本的使用者都应仔细阅读此声明。ClassmateLin保留随时更改或补充此免责声明的权利。一旦使用并复制了任何相关脚本或jd_scripts项目的规则,则视为您已接受此免责声明.

您必须在下载后的24小时内从计算机或手机中完全删除以上内容.
您使用或者复制了本仓库且本人制作的任何脚本,则视为已接受此声明,请仔细阅读

Owner
ClassmateLin
ClassmateLin
Web-Scrapper using Python and Flask

Web-Scrapper "[초급]Python으로 웹 스크래퍼 만들기" 코스 -NomadCoders 기초적인 Python 문법강의부터 시작하여 웹사이트의 html파일에서 원하는 내용을 Scrapping해서 출력, csv 파일로 저장, flask를 이용한 간단한 웹페이지

윤성도 1 Nov 10, 2021
UdemyBot - A Simple Udemy Free Courses Scrapper

UdemyBot - A Simple Udemy Free Courses Scrapper

Gautam Kumar 112 Nov 12, 2022
A dead simple crawler to get books information from Douban.

Introduction A dead simple crawler to get books information from Douban. Pre-requesites Python 3 Install dependencies from requirements.txt (Optional)

Yun Wang 1 Jan 10, 2022
12306抢票脚本

12306抢票脚本

罐子里的茶 457 Jan 05, 2023
TikTok Username Swapper/Claimer/etc

TikTok-Turbo TikTok Username Swapper/Claimer/etc I wanted to create it as fast as possible but i eventually gave up and recoded it many many many many

Kevin 12 Dec 19, 2022
This repo has the source code for the crawler and data crawled from auto-data.net

This repo contains the source code for crawler and crawled data of cars specifications from autodata. The data has roughly 45k cars

Tô Đức Anh 5 Nov 22, 2022
a way to scrape a database of all of the isef projects

ISEF Database This is a simple web scraper which gets all of the projects and abstract information from here. My goal for this is for someone to get i

William Kaiser 1 Mar 18, 2022
Minimal set of tools to conduct stealthy scraping.

Stealthy Scraping Tools Do not use puppeteer and playwright for scraping. Explanation. We only use the CDP to obtain the page source and to get the ab

Nikolai Tschacher 88 Jan 04, 2023
Visual scraping for Scrapy

Portia Portia is a tool that allows you to visually scrape websites without any programming knowledge required. With Portia you can annotate a web pag

Scrapinghub 8.7k Jan 05, 2023
OSTA web scraper, for checking the status of school buses in Ottawa

OSTA-La-Vista OSTA web scraper, for checking the status of school buses in Ottawa. Getting Started Using a Raspberry Pi, download Python 3, and option

1 Jan 28, 2022
Unja is a fast & light tool for fetching known URLs from Wayback Machine

Unja Fetch Known Urls What's Unja? Unja is a fast & light tool for fetching known URLs from Wayback Machine, Common Crawl, Virus Total & AlienVault's

Sheryar 10 Aug 07, 2022
一些爬虫相关的签名、验证码破解

cracking4crawling 一些爬虫相关的签名、验证码破解,目前已有脚本: 小红书App接口签名(shield)(2020.12.02) 小红书滑块(数美)验证破解(2020.12.02) 海南航空App接口签名(hnairSign)(2020.12.05) 说明: 脚本按目标网站、App命

XNFA 90 Feb 09, 2021
Scraping Top Repositories for Topics on GitHub,

0.-Webscrapping-using-python Scraping Top Repositories for Topics on GitHub, Web scraping is the process of extracting and parsing data from websites

Dev Aravind D Satprem 2 Mar 18, 2022
A web Scraper for CSrankings.com that scrapes University and Faculty list for a particular country

A look into what we're building Demo.mp4 Prerequisites Python 3 Node v16+ Steps to run Create a virtual environment. Activate the virtual environment.

2 Jun 06, 2022
A package designed to scrape data from Yahoo Finance.

yahoostock A package designed to scrape data from Yahoo Finance. Installation The most simple installation method is through PIP. pip install yahoosto

Rohan Singh 2 May 28, 2022
Web crawling framework based on asyncio.

Web crawling framework for everyone. Written with asyncio, uvloop and aiohttp. Requirements Python3.5+ Installation pip install gain pip install uvloo

Jiuli Gao 2k Jan 05, 2023
Scrapy uses Request and Response objects for crawling web sites.

Requests and Responses¶ Scrapy uses Request and Response objects for crawling web sites. Typically, Request objects are generated in the spiders and p

Md Rashidul Islam 1 Nov 03, 2021
Scrape plants scientific name information from Agroforestry Species Switchboard 2.0.

Agroforestry Species Switchboard 2.0 Scraper Scrape plants scientific name information from Species Switchboard 2.0. Requirements python = 3.10 (you

Mgs. M. Rizqi Fadhlurrahman 2 Dec 23, 2021
This app will let you continuously scrape certain parts of LeasePlan and extract data of cars becoming available for lease.

LeasePlan - Scraper This app will let you continuously scrape certain parts of LeasePlan and extract data of cars becoming available for lease. It has

Rodney 4 Nov 18, 2022
Web Scraping OLX with Python and Bsoup.

webScrap WebScraping first step. Authors: Paulo, Claudio M. First steps in Web Scraping. Project carried out for training in Web Scrapping. The export

claudio paulo 5 Sep 25, 2022