Google Maps crawler using Selenium

Overview

Google Maps Crawler using Selenium

antifragile project python version GitHub Code style: black try/except style: tryceratops Open in Visual Studio Code Follow guilatrova

Built as part of the Antifragile Dev Project

Selenium crawler that browses Google Maps as a regular user and stores the data in an object.


Sample

Sample

Extracted data example:

 Place(                                                                                                            │ │
│ │ │   name='Pizza Me Santos',                                                                                    │ │
│ │ │   address='Av. Washington Luis, 565 - loja 05 - Boqueirão, Santos - SP, 11055-001',                          │ │
│ │ │   business_hours={                                                                                           │ │
│ │ │   │   'Wednesday': '6–10:30PM',                                                                              │ │
│ │ │   │   'Thursday': '6–10:30PM',                                                                               │ │
│ │ │   │   'Friday': '6–11PM',                                                                                    │ │
│ │ │   │   'Saturday': '6–11PM',                                                                                  │ │
│ │ │   │   'Sunday': '6–10:30PM',                                                                                 │ │
│ │ │   │   'Monday': '6–10:30PM',                                                                                 │ │
│ │ │   │   'Tuesday': '6–10:30PM'                                                                                 │ │
│ │ │   },                                                                                                         │ │
│ │ │   photo_link='https://lh5.googleusercontent.com/p/AF1QipMyVkKioODaU0A_ogHPXosm_QcMndZN6I6YHIDo=w408-h272-k-no│ │
│ │ │   rate='5.0',                                                                                                │ │
│ │ │   reviews='16 reviews',                                                                                      │ │
│ │ │   extra_attrs={                                                                                              │ │
│ │ │   │   'Menu': 'Menu\npizzame-santos.goomer.app',                                                             │ │
│ │ │   │   'Website: pizzame-santos.goomer.app ': 'pizzame-santos.goomer.app',                                    │ │
│ │ │   │   'Phone: (13) 3385-0059 ': '(13) 3385-0059',                                                            │ │
│ │ │   │   'Plus code: 2MHC+WF Boqueirão, Santos - State of São Paulo': '2MHC+WF Boqueirão, Santos - State of São │ │
│ │ Paulo'                                                                                                         │ │
│ │ │   },                                                                                                         │ │
│ │ │   traits={                                                                                                   │ │
│ │ │   │   'Service options': ['No-contact delivery', 'Delivery', 'Takeaway', 'Dine-in'],                         │ │
│ │ │   │   'Accessibility': ['Wheelchair-accessible entrance'],                                                   │ │
│ │ │   │   'Offerings': ['Organic dishes', 'Vegetarian options'],                                                 │ │
│ │ │   │   'Dining options': ['Dessert'],                                                                         │ │
│ │ │   │   'Amenities': ['Good for kids'],                                                                        │ │
│ │ │   │   'Atmosphere': ['Casual'],                                                                              │ │
│ │ │   │   'Crowd': ['Groups'],                                                                                   │ │
│ │ │   │   'Planning': ['Accepts reservations'],                                                                  │ │
│ │ │   │   'Payments': ['Credit cards']                                                                           │ │
│ │ │   }                                                                                                          │ │
│ │ )
Owner
Guilherme Latrova
Sportist, Creator, Software writer, Coffee appreciator, Lucky husband and God servant :)
Guilherme Latrova
Command line program to download documents from web portals.

command line document download made easy Highlights list available documents in json format or download them filter documents using string matching re

16 Dec 26, 2022
A high-level distributed crawling framework.

Cola: high-level distributed crawling framework Overview Cola is a high-level distributed crawling framework, used to crawl pages and extract structur

Xuye (Chris) Qin 1.5k Jan 04, 2023
Consulta de CPF e CNPJ na Receita Federal com Web-Scraping

Repositório contendo scripts Python que realizam a consulta de CPF e CNPJ diretamente no site da Receita Federal.

Josué Campos 5 Nov 29, 2021
SkyScrapers: A collection of variety of Scraping Apps

SkyScrapers Collection of variety of Web Scraping Apps The web-scrapers involved

Biplov Pokhrel 3 Feb 17, 2022
一款利用Python来自动获取QQ音乐上某个歌手所有歌曲歌词的爬虫软件

QQ音乐歌词爬虫 一款利用Python来自动获取QQ音乐上某个歌手所有歌曲歌词的爬虫软件,默认去除了所有演唱会(Live)版本的歌曲。 使用方法 直接运行python run.py即可,然后输入你想获取的歌手名字,然后静静等待片刻。 output目录下保存生成的歌词和歌名文件。以周杰伦为例,会生成两

Yang Wei 11 Jul 27, 2022
Simple tool to scrape and download cross country ski timings and results from live.skidor.com

LiveSkidorDownload Simple tool to scrape and download cross country ski timings and results from live.skidor.com Usage: Put the python file in a dedic

0 Jan 07, 2022
Python script to check if there is any differences in responses of an application when the request comes from a search engine's crawler.

crawlersuseragents This Python script can be used to check if there is any differences in responses of an application when the request comes from a se

Podalirius 13 Dec 27, 2022
DaProfiler allows you to get emails, social medias, adresses, works and more on your target using web scraping and google dorking techniques

DaProfiler allows you to get emails, social medias, adresses, works and more on your target using web scraping and google dorking techniques, based in France Only. The particularity of this program i

Dalunacrobate 347 Jan 07, 2023
A Pixiv web crawler module

Pixiv-spider A Pixiv spider module WARNING It's an unfinished work, browsing the code carefully before using it. Features 0004 - Readme.md updated, co

Uzuki 1 Nov 14, 2021
Google Developer Profile Badge Scraper

Google Developer Profile Badge Scraper It is a Google Developer Profile Web Scraper which scrapes for specific badges in a user's Google Developer Pro

Hemant Sachdeva 2 Feb 22, 2022
Scraping weather data using Python to receive umbrella reminders

A Python package which scrapes weather data from google and sends umbrella reminders to specified email at specified time daily.

Edula Vinay Kumar Reddy 1 Aug 23, 2022
feapder 是一款简单、快速、轻量级的爬虫框架。以开发快速、抓取快速、使用简单、功能强大为宗旨。支持分布式爬虫、批次爬虫、多模板爬虫,以及完善的爬虫报警机制。

feapder 是一款简单、快速、轻量级的爬虫框架。起名源于 fast、easy、air、pro、spider的缩写,以开发快速、抓取快速、使用简单、功能强大为宗旨,历时4年倾心打造。支持轻量爬虫、分布式爬虫、批次爬虫、爬虫集成,以及完善的爬虫报警机制。 之

boris 1.4k Dec 29, 2022
A simple reddit scraper to get memes (only images) from r/ProgrammerHumor.

memey A simple reddit scraper to get memes (only images) from r/ProgrammerHumor. Note Only works if you have firefox installed (yet). Instructions foo

2 Nov 16, 2021
A simple code to fetch comments below an Instagram post and save them to a csv file

fetch_comments A simple code to fetch comments below an Instagram post and save them to a csv file usage First you have to enter your username and pas

2 Jul 14, 2022
京东云无线宝积分推送,支持查看多设备积分使用情况

JDRouterPush 项目简介 本项目调用京东云无线宝API,可每天定时推送积分收益情况,帮助你更好的观察主要信息 更新日志 2021-03-02: 查询绑定的京东账户 通知排版优化 脚本检测更新 支持Server酱Turbo版 2021-02-25: 实现多设备查询 查询今

雷疯 199 Dec 12, 2022
A Python Covid-19 cases tracker that scrapes data off the web and presents the number of Cases, Recovered Cases, and Deaths that occurred because of the pandemic.

A Python Covid-19 cases tracker that scrapes data off the web and presents the number of Cases, Recovered Cases, and Deaths that occurred because of the pandemic.

Alex Papadopoulos 1 Nov 13, 2021
自动完成每日体温上报(Github Actions)

体温上报助手 简介 每天 10:30 GMT+8 自动完成体温上报,如想修改定时运行的时间,可修改 .github/workflows/SduHealthReport.yml 中 schedule 属性。 如果当日有异常,请手动在小程序端/PC 端填写!

Teng Zhang 23 Sep 15, 2022
京东抢茅台,秒杀成功很多次讨论,天猫抢购,赚钱交流等。

Jd_Seckill 特别声明: 请添加个人微信:19972009719 进群交流讨论 目前群里很多人抢到【扫描微信添加群就好,满200关闭群,有喜欢薅信用卡羊毛的也可以找我交流】 本仓库发布的jd_seckill项目中涉及的任何脚本,仅用于测试和学习研究,禁止用于商业用途,不能保证其合法性,准确性

50 Jan 05, 2023
Demonstration on how to use async python to control multiple playwright browsers for web-scraping

Playwright Browser Pool This example illustrates how it's possible to use a pool of browsers to retrieve page urls in a single asynchronous process. i

Bernardas Ališauskas 8 Oct 27, 2022
A training task for web scraping using python multithreading and a real-time-updated list of available proxy servers.

Parallel web scraping The project is a training task for web scraping using python multithreading and a real-time-updated list of available proxy serv

Kushal Shingote 1 Feb 10, 2022