[ICML 2021] DouZero: Mastering DouDizhu with Self-Play Deep Reinforcement Learning | 斗地主AI

Overview

[ICML 2021] DouZero: Mastering DouDizhu with Self-Play Deep Reinforcement Learning

Logo

Downloads Downloads

DouZero is a reinforcement learning framework for DouDizhu (斗地主), the most popular card game in China. It is a shedding-type game where the player’s objective is to empty one’s hand of all cards before other players. DouDizhu is a very challenging domain with competition, collaboration, imperfect information, large state space, and particularly a massive set of possible actions where the legal actions vary significantly from turn to turn. DouZero is developed by AI Platform, Kwai Inc. (快手).

Community:

  • Slack: Discuss in DouZero channel.
  • QQ Group: Join our QQ group 819204202. Password: douzeroqqgroup

Cite this Work

For now, please cite our Arxiv version:

Zha, Daochen, et al. "DouZero: Mastering DouDizhu with Self-Play Deep Reinforcement Learning." arXiv preprint arXiv:2106.06135 (2021).

@article{zha2021douzero,
  title={DouZero: Mastering DouDizhu with Self-Play Deep Reinforcement Learning},
  author={Zha, Daochen and Xie, Jingru and Ma, Wenye and Zhang, Sheng and Lian, Xiangru and Hu, Xia and Liu, Ji},
  journal={arXiv preprint arXiv:2106.06135},
  year={2021}
}

What Makes DouDizhu Challenging?

In addition to the challenge of imperfect information, DouDizhu has huge state and action spaces. In particular, the action space of DouDizhu is 10^4 (see this table). Unfortunately, most reinforcement learning algorithms can only handle very small action spaces. Moreover, the players in DouDizhu need to both compete and cooperate with others in a partially-observable environment with limited communication, i.e., two Peasants players will play as a team to fight against the Landlord player. Modeling both competing and cooperation is an open research challenge.

In this work, we propose Deep Monte Carlo (DMC) algorithm with action encoding and parallel actors. This leads to a very simple yet surprisingly effective solution for DouDizhu. Please read our paper for more details.

Installation

Clone the repo with

git clone https://github.com/kwai/DouZero.git

Make sure you have python 3.6+ installed. Install dependencies:

cd douzero
pip3 install -r requirements.txt

We recommend installing the stable version of DouZero with

pip3 install douzero

or install the up-to-date version (it could be not stable) with

pip3 install -e .

Training

We assume you have at least one GPU available. Run

python3 train.py

This will train DouZero on one GPU. To train DouZero on multiple GPUs. Use the following arguments.

  • --gpu_devices: what gpu devices are visible
  • --num_actors_devices: how many of the GPU deveices will be used for simulation, i.e., self-play
  • --num_actors: how many actor processes will be used for each device
  • --training_device: which device will be used for training DouZero

For example, if we have 4 GPUs, where we want to use the first 3 GPUs to have 15 actors each for simulating and the 4th GPU for training, we can run the following command:

python3 train.py --gpu_devices 0,1,2,3 --num_actors_devices 3 --num_actors 15 --training_device 3

For more customized configuration of training, see the following optional arguments:

--xpid XPID           Experiment id (default: douzero)
--save_interval SAVE_INTERVAL
                      Time interval (in minutes) at which to save the model
--objective {adp,wp}  Use ADP or WP as reward (default: ADP)
--gpu_devices GPU_DEVICES
                      Which GPUs to be used for training
--num_actor_devices NUM_ACTOR_DEVICES
                      The number of devices used for simulation
--num_actors NUM_ACTORS
                      The number of actors for each simulation device
--training_device TRAINING_DEVICE
                      The index of the GPU used for training models
--load_model          Load an existing model
--disable_checkpoint  Disable saving checkpoint
--savedir SAVEDIR     Root dir where experiment data will be saved
--total_frames TOTAL_FRAMES
                      Total environment frames to train for
--exp_epsilon EXP_EPSILON
                      The probability for exploration
--batch_size BATCH_SIZE
                      Learner batch size
--unroll_length UNROLL_LENGTH
                      The unroll length (time dimension)
--num_buffers NUM_BUFFERS
                      Number of shared-memory buffers
--num_threads NUM_THREADS
                      Number learner threads
--max_grad_norm MAX_GRAD_NORM
                      Max norm of gradients
--learning_rate LEARNING_RATE
                      Learning rate
--alpha ALPHA         RMSProp smoothing constant
--momentum MOMENTUM   RMSProp momentum
--epsilon EPSILON     RMSProp epsilon

Evaluation

The evaluation can be performed with GPU or CPU (GPU will be much faster). Pretrained model is available at Google Drive or 百度网盘, 提取码: 4624. Put pre-trained weights in baselines/. The performance is evaluated through self-play. We have provided pre-trained models and some heuristics as baselines:

  • random: agents that play randomly (uniformly)
  • rlcard: the rule-based agent in RLCard
  • SL (baselines/sl/): the pre-trained deep agents on human data
  • DouZero-ADP (baselines/douzero_ADP/): the pretrained DouZero agents with Average Difference Points (ADP) as objective
  • DouZero-WP (baselines/douzero_WP/): the pretrained DouZero agents with Winning Percentage (WP) as objective

Step 1: Generate evaluation data

python3 generate_eval_data.py

Some important hyperparameters are as follows.

  • --output: where the pickled data will be saved
  • --num_games: how many random games will be generated, default 10000

Step 2: Self-Play

python3 evaluate.py

Some important hyperparameters are as follows.

  • --landlord: which agent will play as Landlord, which can be random, rlcard, or the path of the pre-trained model
  • --landlord_up: which agent will play as LandlordUp (the one plays before the Landlord), which can be random, rlcard, or the path of the pre-trained model
  • --landlord_down: which agent will play as LandlordDown (the one plays after the Landlord), which can be random, rlcard, or the path of the pre-trained model
  • --eval_data: the pickle file that contains evaluation data

For example, the following command evaluates DouZero-ADP in Landlord position against random agents

python3 evaluate.py --landlord baselines/douzero_ADP/landlord.ckpt --landlord_up random --landlord_down random

The following command evaluates DouZero-ADP in Peasants position against RLCard agents

python3 evaluate.py --landlord rlcard --landlord_up baselines/douzero_ADP/landlord_up.ckpt --landlord_down baselines/douzero_ADP/landlord_down.ckpt

By default, our model will be saved in douzero_checkpoints/douzero every half an hour. We provide a script to help you identify the most recent checkpoint. Run

sh get_most_recent.sh douzero_checkpoints/douzero/

The most recent model will be in most_recent_model.

Core Team

Acknowlegements

Comments
  • 执行loss.backward()几个小时报错out of memory

    执行loss.backward()几个小时报错out of memory

    应该出现了OOM,报错如下: Exception in thread batch-and-learn-1: Traceback (most recent call last): File "/usr/lib/python3.7/threading.py", line 917, in _bootstrap_inner self.run() File "/usr/lib/python3.7/threading.py", line 865, in run self._target(*self._args, **self._kwargs) File "/home/zzy/robot/DouZeroV2/douzero/dmc/dmc.py", line 238, in batch_and_learn _stats = learn(position, models, learner_model.get_model(position), batch, optimizers[position], flags, position_lock) File "/home/zzy/robot/DouZeroV2/douzero/dmc/dmc.py", line 101, in learn loss.backward() File "/home/zzy/.local/lib/python3.7/site-packages/torch/_tensor.py", line 396, in backward torch.autograd.backward(self, gradient, retain_graph, create_graph, inputs=inputs) File "/home/zzy/.local/lib/python3.7/site-packages/torch/autograd/__init__.py", line 175, in backward allow_unreachable=True, accumulate_grad=True) # Calls into the C++ engine to run the backward pass

    显存 2G,CUDA Version: 11.7, 启动参数已经调的比较小了: python3 train.py --gpu_devices 0 --load_model --batch_size 16 --num_actor_devices 1 --num_actors 1 --num_threads 1 --training_device 0

    opened by SvenNJ 7
  • 关于models.py 几个问题

    关于models.py 几个问题

    ` self.lstm = nn.LSTM(162, 128, batch_first=True)

    self.dense1 = nn.Linear(373 + 128, 512) # 输入501,输出 512 self.dense2 = nn.Linear(512, 512) self.dense3 = nn.Linear(512, 512) self.dense4 = nn.Linear(512, 512) self.dense5 = nn.Linear(512, 512) self.dense6 = nn.Linear(512, 1) `

    1、这个输入维度是162 ,我理解的是 z_batch 有 num_legal_actions行-5列-162个元素,也就是对应输入的162,不知道是不是这个意思 2、隐藏维度128 不清楚 有什么含义,可以随意设置吗 3、这里设计了6层全连接网路,不清楚 有什么含义,可以随意设置吗

    opened by SvenNJ 7
  • Why is the first GPU over-consumed?

    Why is the first GPU over-consumed?

    I run the code with the following command, spotting that there are 9 processes occupying the first GPU. Why would that be the case?

    python3 train.py --gpu_devices 0,1,2,3 --num_actor_devices 3 --num_actors 3 --training_device 3
    

    The initialization logs look fine to me image

    Here's a snapshot of the result of nvidia-smi

    image environment 
    opened by xlnwel 4
  • train.py 运行 After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0)

    train.py 运行 After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0)

    请问为什么训练模型时候的结果一直是0呢? Found log directory: douzero_checkpoints/douzero Saving arguments to douzero_checkpoints/douzero/meta.json Path to meta file already exists. Not overriding meta. Saving messages to douzero_checkpoints/douzero/out.log Path to message file already exists. New data will be appended. Saving logs data to douzero_checkpoints/douzero/logs.csv Saving logs' fields to douzero_checkpoints/douzero/fields.csv [INFO:13486 utils:118 2021-08-10 03:03:32,181] Device 0 Actor 0 started. [INFO:13498 utils:118 2021-08-10 03:03:37,035] Device 0 Actor 1 started. [INFO:13506 utils:118 2021-08-10 03:03:41,833] Device 0 Actor 2 started. [INFO:13514 utils:118 2021-08-10 03:03:54,821] Device 0 Actor 3 started. [INFO:13526 utils:118 2021-08-10 03:04:19,880] Device 0 Actor 4 started. [INFO:13468 dmc:194 2021-08-10 03:04:24,883] Saving checkpoint to douzero_checkpoints/douzero/model.tar [INFO:13468 dmc:243 2021-08-10 03:04:25,064] After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0) Stats: {'loss_landlord': 0, 'loss_landlord_down': 0, 'loss_landlord_up': 0, 'mean_episode_return_landlord': 0, 'mean_episode_return_landlord_down': 0, 'mean_episode_return_landlord_up': 0} [INFO:13468 dmc:243 2021-08-10 03:04:30,069] After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0) Stats: {'loss_landlord': 0, 'loss_landlord_down': 0, 'loss_landlord_up': 0, 'mean_episode_return_landlord': 0, 'mean_episode_return_landlord_down': 0, 'mean_episode_return_landlord_up': 0} [INFO:13468 dmc:243 2021-08-10 03:04:35,074] After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0) Stats: {'loss_landlord': 0, 'loss_landlord_down': 0, 'loss_landlord_up': 0, 'mean_episode_return_landlord': 0, 'mean_episode_return_landlord_down': 0, 'mean_episode_return_landlord_up': 0} [INFO:13468 dmc:243 2021-08-10 03:04:40,080] After 0 (L:0 U:0 D:0) frames: @ 0.0 fps (avg@ 0.0 fps) (L:0.0 U:0.0 D:0.0) Stats: {'loss_landlord': 0, 'loss_landlord_down': 0, 'loss_landlord_up': 0, 'mean_episode_return_landlord': 0, 'mean_episode_return_landlord_down': 0, 'mean_episode_return_landlord_up': 0}

    opened by Roywaller 4
  • 有没有必要设置size变量

    有没有必要设置size变量

    为什么不直接用 len(obs_z_buf) ? 而且执行 position, obs, env_output = env.step(action) 后, position 指向下一个玩家,加1加错地方了 https://github.com/kwai/DouZero/blob/17a7452333e03f4d583feeedd0c4ea3250fc493f/douzero/dmc/utils.py#L142-L143

    opened by yffbit 3
  • SL模型的训练数据是如何实现和加载的?

    SL模型的训练数据是如何实现和加载的?

    你好,我这边在做斗地主的SL模型。因为数据量特别多,达到百万局以上。我想问下你们这边关于监督学习的的数据加载方式是什么样的? 我这边主要有两个想法:

    1. 将所有数据做出numpy格式,但这样就会个上百万个numpy文件,陷入IO瓶颈。
    2. 全部读入原始数据,一边编码,一边训练。 我想问下你们哪张方法更好,或者说有哪种更好的方法?谢谢
    question 
    opened by whiplash003 3
  • why use 'Lock' in

    why use 'Lock' in "get_batch()" while using queue to communication between threads

    The codes is in douzero.dmc.utils, line 5 to 52. I am confused about the usage of lock. To my knowledge, queue module can achieve thread synchronization, then why use Lock mechanism again. I don't know whether my question is clearly. Hope it's worthy. Yours sincerely.

    opened by huzhoudaxia 3
  • 策略调整可行性

    策略调整可行性

    目前是根据当前人的手牌、历史出牌、各玩家出牌历史、剩余数量以及 炸弹组合的,非常庞大的数据量,训练需要耗费非常久的时间。

    若调整为 当前人的手牌、下家手牌、上家手牌、最近一次出牌记录 来训练 会不会更快些,也就是大家都是明牌来打斗地主。

    LSTM这里 也不要 历史出牌记录来初始化了,直接按照上面的明牌元素,Linear还是保持6层,512大小

    最终反正都是按照牌局结束来奖励。

    想知道这种策略在 地主与农民 合作与对抗上面 是否有效

    opened by SvenNJ 2
  • why use multiple threads to execute the function

    why use multiple threads to execute the function "batch_and_learn" in dmc.py.

    In douzero.dmc.dmc, about line 168 to line 174, why use multiple threads to execute the function "batch_and_learn"? Wouldn't the algorithm run slower because the GIL of python? Hope for your answer. Thank you. Best regards.

    opened by huzhoudaxia 2
  • 新版本在Windows下仍然无法Train?

    新版本在Windows下仍然无法Train?

    Log如下

    `C:\Users\A\Downloads\DouZero-main>python train.py Found log directory: douzero_checkpoints\douzero Saving arguments to douzero_checkpoints\douzero/meta.json Path to meta file already exists. Not overriding meta. Saving messages to douzero_checkpoints\douzero/out.log Path to message file already exists. New data will be appended. Saving logs data to douzero_checkpoints\douzero/logs.csv Saving logs' fields to douzero_checkpoints\douzero/fields.csv THCudaCheck FAIL file=..\torch/csrc/generic/StorageSharing.cpp line=258 error=801 : operation not supported Traceback (most recent call last): File "train.py", line 8, in train(flags) File "C:\Users\A\Downloads\DouZero-main\douzero\dmc\dmc.py", line 138, in train actor.start() File "E:\Software\miniconda3\lib\multiprocessing\process.py", line 121, in start self._popen = self._Popen(self) File "E:\Software\miniconda3\lib\multiprocessing\context.py", line 327, in _Popen return Popen(process_obj) File "E:\Software\miniconda3\lib\multiprocessing\popen_spawn_win32.py", line 93, in init reduction.dump(process_obj, to_child) File "E:\Software\miniconda3\lib\multiprocessing\reduction.py", line 60, in dump ForkingPickler(file, protocol).dump(obj) File "E:\Software\miniconda3\lib\site-packages\torch\multiprocessing\reductions.py", line 247, in reduce_tensor event_sync_required) = storage.share_cuda() RuntimeError: cuda runtime error (801) : operation not supported at ..\torch/csrc/generic/StorageSharing.cpp:258

    C:\Users\A\Downloads\DouZero-main>Traceback (most recent call last): File "", line 1, in File "E:\Software\miniconda3\lib\multiprocessing\spawn.py", line 116, in spawn_main exitcode = _main(fd, parent_sentinel) File "E:\Software\miniconda3\lib\multiprocessing\spawn.py", line 126, in _main self = reduction.pickle.load(from_parent) EOFError: Ran out of input`

    opened by Vincentzyx 1
  • Error when fighting with RLCard in landlord_up position

    Error when fighting with RLCard in landlord_up position

    I run it with python3 evaluate.py --landlord rlcard --landlord_up baselines/douzero_ADP/landlord_up.ckpt --landlord_down baselines/douzero_ADP/landlord_down.ckpt and I get this error: Traceback (most recent call last): File "/DouZero-main/douzero/evaluation/rlcard_agent.py", line 62, in act the_type = CARD_TYPE[0][last_move][0][0] KeyError: '666777BR'

    opened by orange90 1
  • 求助,用CPU无法训练问题!!!

    求助,用CPU无法训练问题!!!

    --training_device 改为 CPU后 parser.add_argument('--training_device', default='cpu', type=str, help='The index of the GPU used for training models. cpu means using cpu')

    还是出现 AssertionError: CUDA not available. If you have GPUs, please specify the ID after --gpu_devices. Otherwise, please train with CPU with python3 train.py --actor_device_cpu --training_device cpu

    opened by BESTTENG 1
  • 为什么同样是4个GPU我的训练时候的FPS很低呢,基本都在2000左右

    为什么同样是4个GPU我的训练时候的FPS很低呢,基本都在2000左右

    [INFO:1052 dmc:233 2022-07-20 17:39:38,765] After 1632000 (L:556800 U:528000 D:547200) frames: @ 1918.7 fps (avg@ 2318.1 fps) (L:0.0 U:0.0 D:1918.7) Stats: {'loss_landlord': 1.9155352115631104, 'loss_landlord_down': 2.5349276065826416, 'loss_landlord_up': 2.1095376014709473, 'mean_episode_return_landlord': 0.08421196788549423, 'mean_episode_return_landlord_down': -0.08074238896369934, 'mean_episode_return_landlord_up': -0.06534682214260101} [INFO:1052 dmc:233 2022-07-20 17:39:43,769] After 1648000 (L:563200 U:537600 D:547200) frames: @ 3197.8 fps (avg@ 2398.1 fps) (L:1279.1 U:1918.7 D:0.0) Stats: {'loss_landlord': 2.3213179111480713, 'loss_landlord_down': 2.5349276065826416, 'loss_landlord_up': 2.6052844524383545, 'mean_episode_return_landlord': 0.09171878546476364, 'mean_episode_return_landlord_down': -0.08074238896369934, 'mean_episode_return_landlord_up': -0.08009536564350128} [INFO:1052 dmc:233 2022-07-20 17:39:48,773] After 1654400 (L:569600 U:537600 D:547200) frames: @ 1279.1 fps (avg@ 2398.1 fps) (L:1279.1 U:0.0 D:0.0) Stats: {'loss_landlord': 2.185067892074585, 'loss_landlord_down': 2.5349276065826416, 'loss_landlord_up': 2.6052844524383545, 'mean_episode_return_landlord': 0.09759927541017532, 'mean_episode_return_landlord_down': -0.08074238896369934, 'mean_episode_return_landlord_up': -0.08009536564350128} [INFO:1052 dmc:233 2022-07-20 17:39:53,779] After 1673600 (L:576000 U:540800 D:556800) frames: @ 3836.1 fps (avg@ 2344.8 fps) (L:1278.7 U:639.4 D:1918.1) Stats: {'loss_landlord': 1.77787184715271, 'loss_landlord_down': 2.7444241046905518, 'loss_landlord_up': 2.508575677871704, 'mean_episode_return_landlord': 0.10005713254213333, 'mean_episode_return_landlord_down': -0.09260766953229904, 'mean_episode_return_landlord_up': -0.08521360903978348} [INFO:1052 dmc:233 2022-07-20 17:39:58,781] After 1680000 (L:576000 U:547200 D:556800) frames: @ 1279.5 fps (avg@ 2398.1 fps) (L:0.0 U:1279.5 D:0.0) Stats: {'loss_landlord': 1.77787184715271, 'loss_landlord_down': 2.7444241046905518, 'loss_landlord_up': 2.264894723892212, 'mean_episode_return_landlord': 0.10005713254213333, 'mean_episode_return_landlord_down': -0.09260766953229904, 'mean_episode_return_landlord_up': -0.08965221047401428}

    +-----------------------------------------------------------------------------+ | NVIDIA-SMI 495.29.05 Driver Version: 495.29.05 CUDA Version: 11.5 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVIDIA A100-SXM... On | 00000000:0A:00.0 Off | 0 | | N/A 31C P0 96W / 400W | 66690MiB / 81251MiB | 99% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ | 1 NVIDIA A100-SXM... On | 00000000:45:00.0 Off | 0 | | N/A 32C P0 95W / 400W | 66704MiB / 81251MiB | 98% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ | 2 NVIDIA A100-SXM... On | 00000000:4B:00.0 Off | 0 | | N/A 34C P0 95W / 400W | 66700MiB / 81251MiB | 98% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ | 3 NVIDIA A100-SXM... On | 00000000:84:00.0 Off | 0 | | N/A 39C P0 66W / 400W | 2653MiB / 81251MiB | 2% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+

    +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| +-----------------------------------------------------------------------------+

    fps一直很低,偶尔会出现FPS0的情况 偶尔也会跳到5000.请问这是正常训练的速度吗

    opened by mfxiaosheng 1
Releases(1.1.0)
FEDn is an open-source, modular and ML-framework agnostic framework for Federated Machine Learning

FEDn is an open-source, modular and ML-framework agnostic framework for Federated Machine Learning (FedML) developed and maintained by Scaleout Systems. FEDn enables highly scalable cross-silo and cr

Scaleout 75 Nov 09, 2022
Plenoxels: Radiance Fields without Neural Networks

Plenoxels: Radiance Fields without Neural Networks Alex Yu*, Sara Fridovich-Keil*, Matthew Tancik, Qinhong Chen, Benjamin Recht, Angjoo Kanazawa UC Be

Sara Fridovich-Keil 81 Dec 25, 2022
A computer vision pipeline to identify the "icons" in Christian paintings

Christian-Iconography A computer vision pipeline to identify the "icons" in Christian paintings. A bit about iconography. Iconography is related to id

Rishab Mudliar 3 Jul 30, 2022
Pywonderland - A tour in the wonderland of math with python.

A Tour in the Wonderland of Math with Python A collection of python scripts for drawing beautiful figures and animating interesting algorithms in math

Zhao Liang 4.1k Jan 03, 2023
Repository of continual learning papers

Continual learning paper repository This repository contains an incomplete (but dynamically updated) list of papers exploring continual learning in ma

29 Jan 05, 2023
Exploring whether attention is necessary for vision transformers

Do You Even Need Attention? A Stack of Feed-Forward Layers Does Surprisingly Well on ImageNet Paper/Report TL;DR We replace the attention layer in a v

Luke Melas-Kyriazi 461 Jan 07, 2023
Official implementation of the paper WAV2CLIP: LEARNING ROBUST AUDIO REPRESENTATIONS FROM CLIP

Wav2CLIP 🚧 WIP 🚧 Official implementation of the paper WAV2CLIP: LEARNING ROBUST AUDIO REPRESENTATIONS FROM CLIP 📄 🔗 Ho-Hsiang Wu, Prem Seetharaman

Descript 240 Dec 13, 2022
Image super-resolution (SR) is a fast-moving field with novel architectures attracting the spotlight

Revisiting RCAN: Improved Training for Image Super-Resolution Introduction Image super-resolution (SR) is a fast-moving field with novel architectures

Zudi Lin 76 Dec 01, 2022
🗺 General purpose U-Network implemented in Keras for image segmentation

TF-Unet General purpose U-Network implemented in Keras for image segmentation Getting started • Training • Evaluation Getting started Looking for Jupy

Or Fleisher 2 Aug 31, 2022
RoMa: A lightweight library to deal with 3D rotations in PyTorch.

RoMa: A lightweight library to deal with 3D rotations in PyTorch. RoMa (which stands for Rotation Manipulation) provides differentiable mappings betwe

NAVER 90 Dec 27, 2022
Fiddle is a Python-first configuration library particularly well suited to ML applications.

Fiddle Fiddle is a Python-first configuration library particularly well suited to ML applications. Fiddle enables deep configurability of parameters i

Google 227 Dec 26, 2022
Python library containing BART query generation and BERT-based Siamese models for neural retrieval.

Neural Retrieval Embedding-based Zero-shot Retrieval through Query Generation leverages query synthesis over large corpuses of unlabeled text (such as

Amazon Web Services - Labs 35 Apr 14, 2022
Time Delayed NN implemented in pytorch

Pytorch Time Delayed NN Time Delayed NN implemented in PyTorch. Usage kernels = [(1, 25), (2, 50), (3, 75), (4, 100), (5, 125), (6, 150)] tdnn = TDNN

Daniil Gavrilov 79 Aug 04, 2022
FactSeg: Foreground Activation Driven Small Object Semantic Segmentation in Large-Scale Remote Sensing Imagery (TGRS)

FactSeg: Foreground Activation Driven Small Object Semantic Segmentation in Large-Scale Remote Sensing Imagery by Ailong Ma, Junjue Wang*, Yanfei Zhon

Kingdrone 43 Jan 05, 2023
Code release for Convolutional Two-Stream Network Fusion for Video Action Recognition

Convolutional Two-Stream Network Fusion for Video Action Recognition

Christoph Feichtenhofer 676 Dec 31, 2022
Escaping the Gradient Vanishing: Periodic Alternatives of Softmax in Attention Mechanism

Period-alternatives-of-Softmax Experimental Demo for our paper 'Escaping the Gradient Vanishing: Periodic Alternatives of Softmax in Attention Mechani

slwang9353 0 Sep 06, 2021
NeuroFind - A solution to the to the Task given by the Oberseminar of Messtechnik Institute of TU Dresden in 2021

NeuroFind A solution to the to the Task given by the Oberseminar of Messtechnik

1 Jan 20, 2022
PyTorch Implementation of SSTNs for hyperspectral image classifications from the IEEE T-GRS paper "Spectral-Spatial Transformer Network for Hyperspectral Image Classification: A FAS Framework."

PyTorch Implementation of SSTN for Hyperspectral Image Classification Paper links: SSTN published on IEEE T-GRS. Also, you can directly find the imple

Zilong Zhong 54 Dec 19, 2022
CvT-ASSD: Convolutional vision-Transformerbased Attentive Single Shot MultiBox Detector (ICTAI 2021 CCF-C 会议)The 33rd IEEE International Conference on Tools with Artificial Intelligence

CvT-ASSD including extra CvT, CvT-SSD, VGG-ASSD models original-code-website: https://github.com/albert-jin/CvT-SSD new-code-website: https://github.c

金伟强 -上海大学人工智能小渣渣~ 5 Mar 07, 2022
Simple tools for logging and visualizing, loading and training

TNT TNT is a library providing powerful dataloading, logging and visualization utilities for Python. It is closely integrated with PyTorch and is desi

1.5k Jan 02, 2023