HeatNet is a python package that provides tools to build, train and evaluate neural networks designed to predict extreme heat wave events globally on daily to subseasonal timescales.

Related tags

Deep Learningheatnet
Overview

HeatNet

HeatNet is a python package that provides tools to build, train and evaluate neural networks designed to predict extreme heat wave events globally on daily to subseasonal timescales. It also includes preprocessing tools for atmospheric reanalysis data from the Copernicus Climate Data Store.

Dependencies

HeatNet relies on the DLWP-CS project, described in Weyn et al. (2020), and inherits all of its dependencies.

HeatNet requires installation of

  • TensorFlow >= 2.0, to build neural networks and data generators.
  • netCDF4, to read and write netCDF4 datasets.
  • xarray, to seamlessly manipulate datasets and data arrays.
  • dask, to support parallel xarray computations and streaming computation on datasets that don't fit into memory.
  • h5netcdf, which provides a flexible engine for xarray I/O operations.
  • NumPy for efficient array manipulation.
  • cdsapi, to enable downloading data from the Copernicus Climate Data Store.
  • TempestRemap, for mapping functions from latitude-longitude grids to cubed-sphere grids.

Modules

  • data: Classes and methods to download, preprocess and generate reanalysis data for model training.
  • model: Model architectures, custom losses and model estimators with descriptive metadata.
  • eval: Methods to evaluate model predictions, and compare against persistence or climatology.
  • test: Unit tests for classes and methods in the package.

License

HeatNet is distributed under the GNU General Public License Version 3, which means that any software modifying or relying on the HeatNet package must be distributed under the same license. Consult the full notice to understand your rights.

Installation guide

The installation of heatnet and its dependencies has been tested with the following configuration on both Linux and Mac personal workstations:

  • Create a new Python 3.7 environment using [conda] (https://www.anaconda.com/products/individual).

  • In the terminal, activate the environment,
    conda activate .

  • Install TensorFlow v2.3,
    pip install tensorflow==2.3

  • Install xarray,
    pip install xarray

  • Install netCDF4,
    conda install netCDF4

  • Install TempestRemap,
    conda install -c conda-forge tempest-remap

  • Install h5netcdf,
    conda install -c conda-forge h5netcdf

  • Install pygrib (Optional),
    pip install pygrib

  • Install cdsapi,
    pip install cdsapi

  • Install h5py v2.10.0,
    pip install h5py==2.10.0

  • Finally, install dask,
    pip install dask

  • The DLWP package is not currently published, so the source code must be downloaded from its GitHub repository. It is recommended to download this package in the same parent directory as HeatNet,
    git clone https://github.com/jweyn/DLWP-CS.git

  • If you want to plot results using Basemap, which is a slightly fragile (and deprecated) package, the following configuration is compatible with this setup:
    conda install basemap
    pip install -U matplotlib==3.2

Disclaimers

This is not an officially supported Google Product.

Owner
Google Research
Google Research
Use unsupervised and supervised learning to predict stocks

AIAlpha: Multilayer neural network architecture for stock return prediction This project is meant to be an advanced implementation of stacked neural n

Vivek Palaniappan 1.5k Dec 26, 2022
Official repository for the ICCV 2021 paper: UltraPose: Synthesizing Dense Pose with 1 Billion Points by Human-body Decoupling 3D Model.

UltraPose: Synthesizing Dense Pose with 1 Billion Points by Human-body Decoupling 3D Model Official repository for the ICCV 2021 paper: UltraPose: Syn

MomoAILab 92 Dec 21, 2022
WPPNets: Unsupervised CNN Training with Wasserstein Patch Priors for Image Superresolution

WPPNets: Unsupervised CNN Training with Wasserstein Patch Priors for Image Superresolution This code belongs to the paper [1] available at https://arx

Fabian Altekrueger 5 Jun 02, 2022
Implementation of SwinTransformerV2 in TensorFlow.

SwinTransformerV2-TensorFlow A TensorFlow implementation of SwinTransformerV2 by Microsoft Research Asia, based on their official implementation of Sw

Phan Nguyen 2 May 30, 2022
PyTorch implementation for ComboGAN

ComboGAN This is our ongoing PyTorch implementation for ComboGAN. Code was written by Asha Anoosheh (built upon CycleGAN) [ComboGAN Paper] If you use

Asha Anoosheh 139 Dec 20, 2022
Predict multi paths to a moving person depending on his trajectory history.

Multi-future Trajectory Prediction The project is about using the Multiverse model to make possible multible-future trajectory prediction for a seen p

Said Gamal 1 Jan 18, 2022
GPOEO is a micro-intrusive GPU online energy optimization framework for iterative applications

GPOEO GPOEO is a micro-intrusive GPU online energy optimization framework for iterative applications. We also implement ODPP [1] as a comparison. [1]

瑞雪轻飏 8 Sep 10, 2022
NeuTex: Neural Texture Mapping for Volumetric Neural Rendering

NeuTex: Neural Texture Mapping for Volumetric Neural Rendering Paper: https://arxiv.org/abs/2103.00762 Running Run on the provided DTU scene cd run ba

Fanbo Xiang 67 Dec 28, 2022
PyTorch DepthNet Training on Still Box dataset

DepthNet training on Still Box Project page This code can replicate the results of our paper that was published in UAVg-17. If you use this repo in yo

Clément Pinard 115 Nov 21, 2022
Code for the Population-Based Bandits Algorithm, presented at NeurIPS 2020.

Population-Based Bandits (PB2) Code for the Population-Based Bandits (PB2) Algorithm, from the paper Provably Efficient Online Hyperparameter Optimiza

Jack Parker-Holder 22 Nov 16, 2022
Simple helper library to convert a collection of numpy data to tfrecord, and build a tensorflow dataset from the tfrecord.

numpy2tfrecord Simple helper library to convert a collection of numpy data to tfrecord, and build a tensorflow dataset from the tfrecord. Installation

Ryo Yonetani 2 Jan 16, 2022
LSTM model trained on a small dataset of 3000 names written in PyTorch

LSTM model trained on a small dataset of 3000 names. Model generates names from model by selecting one out of top 3 letters suggested by model at a time until an EOS (End Of Sentence) character is no

Sahil Lamba 1 Dec 20, 2021
BankNote-Net: Open dataset and encoder model for assistive currency recognition

BankNote-Net: Open Dataset for Assistive Currency Recognition Millions of people around the world have low or no vision. Assistive software applicatio

Microsoft 13 Oct 28, 2022
Cluttered MNIST Dataset

Cluttered MNIST Dataset A setup script will download MNIST and produce mnist/*.t7 files: luajit download_mnist.lua Example usage: local mnist_clutter

DeepMind 50 Jul 12, 2022
ReferFormer - Official Implementation of ReferFormer

The official implementation of the paper: Language as Queries for Referring Video Object Segmentation Language as Queries for Referring Video Object S

Jonas Wu 232 Dec 29, 2022
Anchor-free Oriented Proposal Generator for Object Detection

Anchor-free Oriented Proposal Generator for Object Detection Gong Cheng, Jiabao Wang, Ke Li, Xingxing Xie, Chunbo Lang, Yanqing Yao, Junwei Han, Intro

jbwang1997 56 Nov 15, 2022
GenGNN: A Generic FPGA Framework for Graph Neural Network Acceleration

GenGNN: A Generic FPGA Framework for Graph Neural Network Acceleration Stefan Abi-Karam*, Yuqi He*, Rishov Sarkar*, Lakshmi Sathidevi, Zihang Qiao, Co

Sharc-Lab 19 Dec 15, 2022
A Large-Scale Dataset for Spinal Vertebrae Segmentation in Computed Tomography

A Large-Scale Dataset for Spinal Vertebrae Segmentation in Computed Tomography

ICT.MIRACLE lab 75 Dec 26, 2022
Implementation for paper "Towards the Generalization of Contrastive Self-Supervised Learning"

Contrastive Self-Supervised Learning on CIFAR-10 Paper "Towards the Generalization of Contrastive Self-Supervised Learning", Weiran Huang, Mingyang Yi

Weiran Huang 13 Nov 30, 2022
Graph Transformer Architecture. Source code for

Graph Transformer Architecture Source code for the paper "A Generalization of Transformer Networks to Graphs" by Vijay Prakash Dwivedi and Xavier Bres

NTU Graph Deep Learning Lab 561 Jan 08, 2023