EOD Historical Data Python Library (Unofficial)

Overview

EOD Historical Data Python Library (Unofficial)

https://eodhistoricaldata.com

Installation

python3 -m pip install eodhistoricaldata

Note

Demo API key below is provided by EOD Historial Data for testing purposes https://eodhistoricaldata.com/financial-apis/new-real-time-data-api-websockets

Usage

None: """Main""" websocket = WebSocketClient( # Demo API key for testing purposes api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="crypto", symbols=["BTC-USD"] #api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="forex", symbols=["EURUSD"] #api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="us", symbols=["AAPL"] ) websocket.start() message_count = 0 while True: if websocket: if ( message_count != websocket.message_count ): print(websocket.message) message_count = websocket.message_count sleep(0.25) # output every 1/4 second, websocket is realtime if __name__ == "__main__": main() ">
"""Sample script"""

from time import sleep
from eodhistoricaldata import WebSocketClient

def main() -> None:
    """Main"""

    websocket = WebSocketClient(
        # Demo API key for testing purposes
        api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="crypto", symbols=["BTC-USD"]
        #api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="forex", symbols=["EURUSD"]
        #api_key="OeAFFmMliFG5orCUuwAKQ8l4WWFQ67YX", endpoint="us", symbols=["AAPL"]
    )
    websocket.start()

    message_count = 0
    while True:
        if websocket:
            if (
                message_count != websocket.message_count
            ):
                print(websocket.message)
                message_count = websocket.message_count
                sleep(0.25)  # output every 1/4 second, websocket is realtime

if __name__ == "__main__":
    main()
You might also like...
TE-dependent analysis (tedana) is a Python library for denoising multi-echo functional magnetic resonance imaging (fMRI) data
TE-dependent analysis (tedana) is a Python library for denoising multi-echo functional magnetic resonance imaging (fMRI) data

tedana: TE Dependent ANAlysis TE-dependent analysis (tedana) is a Python library for denoising multi-echo functional magnetic resonance imaging (fMRI)

Hatchet is a Python-based library that allows Pandas dataframes to be indexed by structured tree and graph data.
Hatchet is a Python-based library that allows Pandas dataframes to be indexed by structured tree and graph data.

Hatchet Hatchet is a Python-based library that allows Pandas dataframes to be indexed by structured tree and graph data. It is intended for analyzing

 🧪 Panel-Chemistry - exploratory data analysis and build powerful data and viz tools within the domain of Chemistry using Python and HoloViz Panel.
🧪 Panel-Chemistry - exploratory data analysis and build powerful data and viz tools within the domain of Chemistry using Python and HoloViz Panel.

🧪📈 🐍. The purpose of the panel-chemistry project is to make it really easy for you to do DATA ANALYSIS and build powerful DATA AND VIZ APPLICATIONS within the domain of Chemistry using using Python and HoloViz Panel.

Tuplex is a parallel big data processing framework that runs data science pipelines written in Python at the speed of compiled code

Tuplex is a parallel big data processing framework that runs data science pipelines written in Python at the speed of compiled code. Tuplex has similar Python APIs to Apache Spark or Dask, but rather than invoking the Python interpreter, Tuplex generates optimized LLVM bytecode for the given pipeline and input data set.

Python data processing, analysis, visualization, and data operations

Python This is a Python data processing, analysis, visualization and data operations of the source code warehouse, book ISBN: 9787115527592 Descriptio

Catalogue data - A Python Scripts to prepare catalogue data

catalogue_data Scripts to prepare catalogue data. Setup Clone this repo. Install

fds is a tool for Data Scientists made by DAGsHub to version control data and code at once.
fds is a tool for Data Scientists made by DAGsHub to version control data and code at once.

Fast Data Science, AKA fds, is a CLI for Data Scientists to version control data and code at once, by conveniently wrapping git and dvc

A data parser for the internal syncing data format used by Fog of World.
A data parser for the internal syncing data format used by Fog of World.

A data parser for the internal syncing data format used by Fog of World. The parser is not designed to be a well-coded library with good performance, it is more like a demo for showing the data structure.

Functional Data Analysis, or FDA, is the field of Statistics that analyses data that depend on a continuous parameter.
Comments
  • Syntax issue with query Parameter in get_calendar_ functions

    Syntax issue with query Parameter in get_calendar_ functions

    Hello,

    When using the get_calendar_XXX, functions we cannot use the query parameters defined by EOD as the word "from" is forbidden by Python, for instance : earning=client.get_calendar_earnings(from='2022-11-01', to='2022-11-30')

    will raise an issue.

    Should I pass the argument differently ?

    opened by ATCBGroup 1
  • dependency on matplotlib but it is not installed with pip

    dependency on matplotlib but it is not installed with pip

    dependency on matplotlib but it is not installed with pip

    [email protected]:~/git/traderai/eod$ cat test.py
    from eodhd import APIClient
    api = APIClient("DEMO")
    
    [email protected]:~/git/traderai/eod$ python3 test.py
    Traceback (most recent call last):
      File "/home/mshamber/.local/lib/python3.8/site-packages/eodhd/eodhdgraphs.py", line 5, in <module>
        import matplotlib.pyplot as plt
    ModuleNotFoundError: No module named 'matplotlib'
    
    [email protected]:~/git/traderai/eod$ python3 -m pip install eodhd
    Requirement already satisfied: eodhd in /home/mshamber/.local/lib/python3.8/site-packages (1.0.8)
    Requirement already satisfied: websocket-client==1.3.3 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (1.3.3)
    Requirement already satisfied: rich==12.5.1 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (12.5.1)
    Requirement already satisfied: websockets==10.3 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (10.3)
    Requirement already satisfied: numpy==1.21.6 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (1.21.6)
    Requirement already satisfied: pandas==1.3.5 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (1.3.5)
    Requirement already satisfied: requests==2.28.1 in /home/mshamber/.local/lib/python3.8/site-packages (from eodhd) (2.28.1)
    Requirement already satisfied: commonmark<0.10.0,>=0.9.0 in /home/mshamber/.local/lib/python3.8/site-packages (from rich==12.5.1->eodhd) (0.9.1)
    Requirement already satisfied: typing-extensions<5.0,>=4.0.0; python_version < "3.9" in /home/mshamber/.local/lib/python3.8/site-packages (from rich==12.5.1->eodhd) (4.3.0)
    Requirement already satisfied: pygments<3.0.0,>=2.6.0 in /home/mshamber/.local/lib/python3.8/site-packages (from rich==12.5.1->eodhd) (2.13.0)
    Requirement already satisfied: python-dateutil>=2.7.3 in /home/mshamber/.local/lib/python3.8/site-packages (from pandas==1.3.5->eodhd) (2.8.2)
    Requirement already satisfied: pytz>=2017.3 in /home/mshamber/.local/lib/python3.8/site-packages (from pandas==1.3.5->eodhd) (2022.5)
    Requirement already satisfied: charset-normalizer<3,>=2 in /home/mshamber/.local/lib/python3.8/site-packages (from requests==2.28.1->eodhd) (2.1.1)
    Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3/dist-packages (from requests==2.28.1->eodhd) (2.8)
    Requirement already satisfied: certifi>=2017.4.17 in /usr/lib/python3/dist-packages (from requests==2.28.1->eodhd) (2019.11.28)
    Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3/dist-packages (from requests==2.28.1->eodhd) (1.25.8)
    Requirement already satisfied: six>=1.5 in /home/mshamber/.local/lib/python3.8/site-packages (from python-dateutil>=2.7.3->pandas==1.3.5->eodhd) (1.16.0)
    
    opened by opme 1
Releases(1.0.8)
Owner
Michael Whittle
Solution Architect
Michael Whittle
This is a repo documenting the best practices in PySpark.

Spark-Syntax This is a public repo documenting all of the "best practices" of writing PySpark code from what I have learnt from working with PySpark f

Eric Xiao 447 Dec 25, 2022
InDels analysis of CRISPR lines by NGS amplicon sequencing technology for a multicopy gene family.

CRISPRanalysis InDels analysis of CRISPR lines by NGS amplicon sequencing technology for a multicopy gene family. In this work, we present a workflow

2 Jan 31, 2022
An interactive grid for sorting, filtering, and editing DataFrames in Jupyter notebooks

qgrid Qgrid is a Jupyter notebook widget which uses SlickGrid to render pandas DataFrames within a Jupyter notebook. This allows you to explore your D

Quantopian, Inc. 2.9k Jan 08, 2023
Randomisation-based inference in Python based on data resampling and permutation.

Randomisation-based inference in Python based on data resampling and permutation.

67 Dec 27, 2022
VevestaX is an open source Python package for ML Engineers and Data Scientists.

VevestaX Track failed and successful experiments as well as features. VevestaX is an open source Python package for ML Engineers and Data Scientists.

Vevesta 24 Dec 14, 2022
follow-analyzer helps GitHub users analyze their following and followers relationship

follow-analyzer follow-analyzer helps GitHub users analyze their following and followers relationship by providing a report in html format which conta

Yin-Chiuan Chen 2 May 02, 2022
MoRecon - A tool for reconstructing missing frames in motion capture data.

MoRecon - A tool for reconstructing missing frames in motion capture data.

Yuki Nishidate 38 Dec 03, 2022
University Challenge 2021 With Python

University Challenge 2021 This repository contains: The TeX file of the technical write-up describing the University / HYPER Challenge 2021 under late

2 Nov 27, 2021
Pizza Orders Data Pipeline Usecase Solved by SQL, Sqoop, HDFS, Hive, Airflow.

PizzaOrders_DataPipeline There is a Tony who is owning a New Pizza shop. He knew that pizza alone was not going to help him get seed funding to expand

Melwin Varghese P 4 Jun 05, 2022
Spectral Analysis in Python

SPECTRUM : Spectral Analysis in Python contributions: Please join https://github.com/cokelaer/spectrum contributors: https://github.com/cokelaer/spect

Thomas Cokelaer 280 Dec 16, 2022
ETL pipeline on movie data using Python and postgreSQL

Movies-ETL ETL pipeline on movie data using Python and postgreSQL Overview This project consisted on a automated Extraction, Transformation and Load p

Juan Nicolas Serrano 0 Jul 07, 2021
Vectorizers for a range of different data types

Vectorizers for a range of different data types

Tutte Institute for Mathematics and Computing 69 Dec 29, 2022
Produces a summary CSV report of an Amber Electric customer's energy consumption and cost data.

Amber Electric Usage Summary This is a command line tool that produces a summary CSV report of an Amber Electric customer's energy consumption and cos

Graham Lea 12 May 26, 2022
Gathering data of likes on Tinder within the past 7 days

tinder_likes_data Gathering data of Likes Sent on Tinder within the past 7 days. Versions November 25th, 2021 - Functionality to get the name and age

Alex Carter 12 Jan 05, 2023
Evidence enables analysts to deliver a polished business intelligence system using SQL and markdown.

Evidence enables analysts to deliver a polished business intelligence system using SQL and markdown

915 Dec 26, 2022
The Dash Enterprise App Gallery "Oil & Gas Wells" example

This app is based on the Dash Enterprise App Gallery "Oil & Gas Wells" example. For more information and more apps see: Dash App Gallery See the Dash

Austin Caudill 1 Nov 08, 2021
PATC: Introduction to Big Data Analytics. Practical Data Analytics for Solving Real World Problems

PATC: Introduction to Big Data Analytics. Practical Data Analytics for Solving Real World Problems

1 Feb 07, 2022
songplays datamart provide details about the musical taste of our customers and can help us to improve our recomendation system

Songplays User activity datamart The following document describes the model used to build the songplays datamart table and the respective ETL process.

Leandro Kellermann de Oliveira 1 Jul 13, 2021
DataPrep — The easiest way to prepare data in Python

DataPrep — The easiest way to prepare data in Python

SFU Database Group 1.5k Dec 27, 2022
Py-price-monitoring - A Python price monitor

A Python price monitor This project was focused on Brazil, so the monitoring is

Samuel 1 Jan 04, 2022