Learn machine learning the fun way, with Oracle and RedBull Racing

Overview

Red Bull Racing Analytics Hands-On Labs

License: UPL Quality gate

Introduction

Are you interested in learning machine learning (ML)? How about doing this in the context of the exciting world of F1 racing?! Get your ML skills bootstrapped here with Oracle and Red Bull Racing!

Red Bull F1 Race Car

This tutorial teaches ML analytics with a series of hands-on labs (HOLs) using the Data Science service in Oracle Cloud Infrastructure.

You'll learn how to get data from some public data sources, then how to analyze this data using some of the latest ML techniques. In the process you'll build ML models and test them out in a predictor app.

Getting Started

There is some infrastructure that must be deployed before you can enjoy this tutorial. See the Terraform documentation for more information.

After the OCI infrastructure is deployed, proceed with the beginner's tutorial to start through the ML labs.

Prerequisites

You must have an OCI account. Click here to create a new cloud account.

This solution is designed to work with several OCI services, allowing you to quickly be up-and-running:

There are required OCI resources (see the Terraform documentation for more information) that are needed for this tutorial.

Notes/Issues

None at this time.

URLs

Contributing

This project is open source. Please submit your contributions by forking this repository and submitting a pull request! Oracle appreciates any contributions that are made by the open source community.

License

Copyright (c) 2021 Oracle and/or its affiliates.

Licensed under the Universal Permissive License (UPL), Version 1.0.

See LICENSE for more details.

Comments
  • Refactored Terraform code

    Refactored Terraform code

    • Compatible with ORM, Cloud Shell and Terraform CLI
    • Updated README to include instructions for all three methods
    • Refactored, removing unnecessary resources (Vault, public Subnet, etc.).
    • Added a nerd knob so that it could use an existing Group (rather than create a new one)
    • Fixed ORM RegEx filters to allow dashes (-) and underscores (_), for the names
    opened by timclegg 2
  • Issue with hands on lab guide - launchapp.sh missing

    Issue with hands on lab guide - launchapp.sh missing

    https://github.com/oracle-devrel/redbull-analytics-hol/tree/main/beginners#beginners-hands-on-lab

    In Starting The Web Application it reads:

    cd /home/opc/redbull-analytics-hol/beginners/web ./launchapp.sh start

    However is launchapp.sh is missing, for example

    (redbullenv) cd /home/opc/redbull-analytics-hol/beginners/web (redbullenv) ./launchapp.sh start bash: ./launchapp.sh: No such file or directory

    opened by raekins 1
  • fix: Updating schema.yaml syntax

    fix: Updating schema.yaml syntax

    Making the variable notation follow what the doc syntax shows (https://docs.oracle.com/en-us/iaas/Content/ResourceManager/Concepts/terraformconfigresourcemanager_topic-schema.htm)

    opened by timclegg 1
  • Exploratory Data Analysis Merge Issue

    Exploratory Data Analysis Merge Issue

    Hello I have been encountering an issue while running the lab. The Jupyter notebook 03.f1_analysis_EDA.ipynb has the following issue on cell number 5:


    ValueError Traceback (most recent call last) in ----> 1 df1 = pd.merge(races,results,how='inner',on=['raceId']) 2 df2 = pd.merge(df1,quali,how='inner',on=['raceId','driverId','constructorId']) 3 df3 = pd.merge(df2,drivers,how='inner',on=['driverId']) 4 df4 = pd.merge(df3,constructors,how='inner',on=['constructorId']) 5 df5 = pd.merge(df4,circuit,how='inner',on=['circuitId'])

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in merge(left, right, how, on, left_on, right_on, left_index, right_index, sort, suffixes, copy, indicator, validate) 85 copy=copy, 86 indicator=indicator, ---> 87 validate=validate, 88 ) 89 return op.get_result()

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in init(self, left, right, how, on, left_on, right_on, axis, left_index, right_index, sort, suffixes, copy, indicator, validate) 654 # validate the merge keys dtypes. We may need to coerce 655 # to avoid incompatible dtypes --> 656 self._maybe_coerce_merge_keys() 657 658 # If argument passed to validate,

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in _maybe_coerce_merge_keys(self) 1163 inferred_right in string_types and inferred_left not in string_types 1164 ): -> 1165 raise ValueError(msg) 1166 1167 # datetimelikes must match exactly

    ValueError: You are trying to merge on object and int64 columns. If you wish to proceed you should use pd.concat

    I’m using an oracle automatic deployment provided by oracle as part of their environment. I do not have a lot of experience with Python but one possible ible solution is to read the numeric values form the csv file as integer or float but I’m almost certain the solution might be a little more elaborated than that 😉. Anyway thanks for your time. I’m really excited to test your solution and finish the lab. Thanks again.

    opened by yankodavila 2
  • Has the PAR for the stack deploy image expired.

    Has the PAR for the stack deploy image expired.

    Cannot deploy stack as getting PAR expired message.

    2021/11/07 10:50:11[TERRAFORM_CONSOLE] [INFO] Error Message: work request did not succeed, workId: ocid1.coreservicesworkrequest.oc1.eu-amsterdam-1.abqw2ljrwz2n7qqj7ghdwtnlrqol355oumc7a6coushvgdrebskspaewh7ea, entity: image, action: CREATED. Message: Import image not found: PAR is invalid (maybe is expired or deleted), please check.

    PAR in stack file is https://objectstorage.eu-frankfurt-1.oraclecloud.com/p/khhPjc_IMuyBOMfZUcJajIzCpoZ5aC-D7VMCU__GVZRlIQueXLIIcaaqLOZIuT1a/n/emeasespainsandbox/b/publichol/o/redbullhol-20210809-1523

    opened by Mel-A-M 1
Releases(v0.1.8)
Owner
Oracle DevRel
Oracle DevRel
BigDL - Evaluate the performance of BigDL (Distributed Deep Learning on Apache Spark) in big data analysis problems

Evaluate the performance of BigDL (Distributed Deep Learning on Apache Spark) in big data analysis problems.

Vo Cong Thanh 1 Jan 06, 2022
yt is an open-source, permissively-licensed Python library for analyzing and visualizing volumetric data.

The yt Project yt is an open-source, permissively-licensed Python library for analyzing and visualizing volumetric data. yt supports structured, varia

The yt project 367 Dec 25, 2022
Hidden Markov Models in Python, with scikit-learn like API

hmmlearn hmmlearn is a set of algorithms for unsupervised learning and inference of Hidden Markov Models. For supervised learning learning of HMMs and

2.7k Jan 03, 2023
A probabilistic programming library for Bayesian deep learning, generative models, based on Tensorflow

ZhuSuan is a Python probabilistic programming library for Bayesian deep learning, which conjoins the complimentary advantages of Bayesian methods and

Tsinghua Machine Learning Group 2.2k Dec 28, 2022
Full ELT process on GCP environment.

Rent Houses Germany - GCP Pipeline Project: The goal of the project is to extract data about house rentals in Germany, store, process and analyze it u

Felipe Demenech Vasconcelos 2 Jan 20, 2022
An ETL framework + Monitoring UI/API (experimental project for learning purposes)

Fastlane An ETL framework for building pipelines, and Flask based web API/UI for monitoring pipelines. Project structure fastlane |- fastlane: (ETL fr

Dan Katz 2 Jan 06, 2022
vartests is a Python library to perform some statistic tests to evaluate Value at Risk (VaR) Models

gg I wasn't satisfied with any of the other available Gemini clients, so I wrote my own. Requires Python 3.9 (maybe older, I haven't checked) and opti

RAFAEL RODRIGUES 5 Jan 03, 2023
An Aspiring Drop-In Replacement for NumPy at Scale

Legate NumPy is a Legate library that aims to provide a distributed and accelerated drop-in replacement for the NumPy API on top of the Legion runtime. Using Legate NumPy you do things like run the f

Legate 502 Jan 03, 2023
Analyzing Covid-19 Outbreaks in Ontario

My group and I took Covid-19 outbreak statistics from ontario, and analyzed them to find different patterns and future predictions for the virus

Vishwaajeeth Kamalakkannan 0 Jan 20, 2022
COVID-19 deaths statistics around the world

COVID-19-Deaths-Dataset COVID-19 deaths statistics around the world This is a daily updated dataset of COVID-19 deaths around the world. The dataset c

Nisa Efendioğlu 4 Jul 10, 2022
.npy, .npz, .mtx converter.

npy-converter Matrix Data Converter. Expand matrix for multi-thread, multi-process Divid matrix for multi-thread, multi-process Support: .mtx, .npy, .

taka 1 Feb 07, 2022
Convert monolithic Jupyter notebooks into Ploomber pipelines.

Soorgeon Join our community | Newsletter | Contact us | Blog | Website | YouTube Convert monolithic Jupyter notebooks into Ploomber pipelines. soorgeo

Ploomber 65 Dec 16, 2022
A Python package for Bayesian forecasting with object-oriented design and probabilistic models under the hood.

Disclaimer This project is stable and being incubated for long-term support. It may contain new experimental code, for which APIs are subject to chang

Uber Open Source 1.6k Dec 29, 2022
Statistical & Probabilistic Analysis of Store Sales, University Survey, & Manufacturing data

Statistical_Modelling Statistical & Probabilistic Analysis of Store Sales, University Survey, & Manufacturing data Statistical Methods for Decision Ma

Avnika Mehta 1 Jan 27, 2022
Python reader for Linked Data in HDF5 files

Linked Data are becoming more popular for user-created metadata in HDF5 files.

The HDF Group 8 May 17, 2022
Very basic but functional Kakuro solver written in Python.

kakuro.py Very basic but functional Kakuro solver written in Python. It uses a reduction to exact set cover and Ali Assaf's elegant implementation of

Louis Abraham 4 Jan 15, 2022
Generates a simple report about the current Covid-19 cases and deaths in Malaysia

Generates a simple report about the current Covid-19 cases and deaths in Malaysia. Results are delay one day, data provided by the Ministry of Health Malaysia Covid-19 public data.

Yap Khai Chuen 7 Dec 15, 2022
Elementary is an open-source data reliability framework for modern data teams. The first module of the framework is data lineage.

Data lineage made simple, reliable, and automated. Effortlessly track the flow of data, understand dependencies and analyze impact. Features Visualiza

898 Jan 09, 2023
Validated, scalable, community developed variant calling, RNA-seq and small RNA analysis

Validated, scalable, community developed variant calling, RNA-seq and small RNA analysis. You write a high level configuration file specifying your in

Blue Collar Bioinformatics 917 Jan 03, 2023
Python utility to extract differences between two pandas dataframes.

Python utility to extract differences between two pandas dataframes.

Jaime Valero 8 Jan 07, 2023