联通手机营业厅自动做任务、签到、领流量、领积分等。

Overview

UnicomTask

简介

👯 😄 📫

联通手机营业厅自动完成每日任务,领流量、签到获取积分等,月底流量不发愁。

功能

  • 沃之树领流量、浇水(12M日流量)
  • 每日签到(1积分+翻倍4积分+第七天1G流量日包)
  • 天天抽奖,每天三次免费机会(随机奖励)
  • 游戏中心每日打卡(连续打卡,积分递增至最高7,第七天1G流量日包)
  • 游戏中心宝箱100M任务(100M日流量+随机奖励并翻倍)
  • 4G流量包看视频、下软件任务(90M+150M七日流量)
  • 每日领取100定向积分
  • 积分抽奖,每天最多抽30次(中奖几率渺茫)
  • 冬奥积分活动(第1和7天,可领取600定向积分,其余领取300定向积分,有效期至下月底)
  • 邮件推送运行结果

Github Actions 部署方法

1.fork本项目

项目地址:srcrs/UnicomTask

2.准备需要的参数

手机号、服务密码、appID。

其中appId的获取:

  • 安卓用户可在文件管理 --> Unicom/appid 文件中获取。

  • 苹果用户可抓取客户端登录接口获取。

https://m.client.10010.com/mobileService/login.htm

3.将必要参数填到Secrets

Secrets中的NameValue格式如下:

Name Value 说明
USERNAME_COVER 18566669999 手机号(必须)
PASSWORD_COVER 123456 服务密码(必须)
APPID_COVER xxxxxxxxx appId(必须)
EMAIL_COVER [email protected] 邮箱(可选)
LOTTERY_NUM 填写正整数 抽奖次数(可选)

4.开启Actions

默认Actions处于禁止状态,在Actions选项中开启Actions功能,把那个绿色的长按钮点一下。如果看到左侧工作流上有黄色!号,还需继续开启。

5.进行一次push操作

push操作会触发工作流运行。

删除掉README.md中的 😄 即可。完成后,每天早上7:30将自动完成每日任务。

同步上游代码

在最新的代码中,已经加上自动同步上游代码的action,将会定时在每周五16点执行,文件地址在.github/workflows/auto_merge.yml

同时您也可以安装pull应用,也可实现自动同步上游代码。

申明

本项目仅用于学习。

参考项目

mixool/HiCnUnicom,感谢该项目对于登录部分的思路

You might also like...
Releases(UnicomTask-v1.7)
a way to scrape a database of all of the isef projects

ISEF Database This is a simple web scraper which gets all of the projects and abstract information from here. My goal for this is for someone to get i

William Kaiser 1 Mar 18, 2022
Pseudo API for Google Trends

pytrends Introduction Unofficial API for Google Trends Allows simple interface for automating downloading of reports from Google Trends. Only good unt

General Mills 2.6k Dec 28, 2022
A repository with scraping code and soccer dataset from understat.com.

UNDERSTAT - SHOTS DATASET As many people interested in soccer analytics know, Understat is an amazing source of information. They provide Expected Goa

douglasbc 48 Jan 03, 2023
This repo has the source code for the crawler and data crawled from auto-data.net

This repo contains the source code for crawler and crawled data of cars specifications from autodata. The data has roughly 45k cars

Tô Đức Anh 5 Nov 22, 2022
A Spider for BiliBili comments with a simple API server.

BiliComment A spider for BiliBili comment. Spider Usage Put config.json into config directory, and then python . ./config/config.json. A example confi

Hao 3 Jul 05, 2021
A modern CSS selector implementation for BeautifulSoup

Soup Sieve Overview Soup Sieve is a CSS selector library designed to be used with Beautiful Soup 4. It aims to provide selecting, matching, and filter

Isaac Muse 151 Dec 23, 2022
CRI Scrape is a tool for get general info about Italian Red Cross in GAIA Platform

CRI Scrape CRI Scrape is a tool for get general info about Italian Red Cross in GAIA Platform Disclaimer This code is only for educational purpose. So

Vincenzo Cardone 0 Jul 23, 2022
API which uses discord to scrape NameMC searches/droptime/dropping status of minecraft names

NameMC Scrape API This is an api to scrape NameMC using message previews generated by discord. NameMC makes it a pain to scrape their website, but som

Twilak 2 Dec 22, 2021
Google Scholar Web Scraping

Google Scholar Web Scraping This is a python script that asks for a user to input the url for a google scholar profile, and then it writes publication

Suzan M 1 Dec 12, 2021
A webdriver-based script for reserving Tsinghua badminton courts.

AutoReserve A webdriver-based script for reserving badminton courts. 使用说明 下载 chromedriver 选择当前Chrome对应版本 安装 selenium pip install selenium 更改场次、金额信息dat

Payne Zhang 4 Nov 09, 2021
A simple django-rest-framework api using web scraping

Apicell You can use this api to search in google, bing, pypi and subscene and get results Method : POST Parameter : query Example import request url =

Hesam N 1 Dec 19, 2021
Python script who crawl first shodan page and check DBLTEK vulnerability

🐛 MASS DBLTEK EXPLOIT CHECKER USING SHODAN 🕸 Python script who crawl first shodan page and check DBLTEK vulnerability

Divin 4 Jan 09, 2022
This is a python api to scrape search results from a url.

googlescrape Installation Installation is simple! # Stable version pip install googlescrape Examples from googlescrape import client scrapeClient=cli

1 Dec 15, 2022
Footballmapies - Football mapies for learning webscraping and use of gmplot module in python

Footballmapies - Football mapies for learning webscraping and use of gmplot module in python

1 Jan 28, 2022
A Pixiv web crawler module

Pixiv-spider A Pixiv spider module WARNING It's an unfinished work, browsing the code carefully before using it. Features 0004 - Readme.md updated, co

Uzuki 1 Nov 14, 2021
Proxy scraper. Format: IP | PORT | COUNTRY | TYPE

proxy scraper 🔎 Installation: git clone https://github.com/ebankoff/proxy_scraper Required pip libraries (pip install library name): lxml beautifulso

Eban'ko 19 Dec 07, 2022
Haphazard scripts for scraping bitcoin/bitcoin data from GitHub

This is a quick-and-dirty tool used to scrape bitcoin/bitcoin pull request and commentary data. Each output/pr number folder contains comments.json:

James O'Beirne 8 Oct 12, 2022
Python script for crawling ResearchGate.net papers✨⭐️📎

ResearchGate Crawler Python script for crawling ResearchGate.net papers About the script This code start crawling process by urls in start.txt and giv

Mohammad Sadegh Salimi 4 Aug 30, 2022
SearchifyX, predecessor to Searchify, is a fast Quizlet, Quizizz, and Brainly webscraper with various stealth features.

SearchifyX SearchifyX, predecessor to Searchify, is a fast Quizlet, Quizizz, and Brainly webscraper with various stealth features. SearchifyX lets you

28 Dec 20, 2022
LSpider 一个为被动扫描器定制的前端爬虫

LSpider LSpider - 一个为被动扫描器定制的前端爬虫 什么是LSpider? 一款为被动扫描器而生的前端爬虫~ 由Chrome Headless、LSpider主控、Mysql数据库、RabbitMQ、被动扫描器5部分组合而成。

Knownsec, Inc. 321 Dec 12, 2022