NVIDIA Merlin is an open source library providing end-to-end GPU-accelerated recommender systems, from feature engineering and preprocessing to training deep learning models and running inference in production.

Related tags

Deep LearningMerlin
Overview

NVIDIA Merlin

NVIDIA Merlin is an open source library designed to accelerate recommender systems on NVIDIA’s GPUs. It enables data scientists, machine learning engineers, and researchers to build high-performing recommenders at scale. Merlin includes tools to address common ETL, training, and inference challenges. Each stage of the Merlin pipeline is optimized to support hundreds of terabytes of data, which is all accessible through easy-to-use APIs. For more information, see NVIDIA Merlin.

Benefits

NVIDIA Merlin is a scalable and GPU-accelerated solution, making it easy to build recommender systems from end to end. With NVIDIA Merlin, you can:

  • transform data (ETL) for preprocessing and engineering features.
  • accelerate existing training pipelines in TensorFlow, PyTorch, or FastAI by leveraging optimized, custom-built dataloaders.
  • scale large deep learning recommender models by distributing large embedding tables that exceed available GPU and CPU memory.
  • deploy data transformations and trained models to production with only a few lines of code.

Components of NVIDIA Merlin

NVIDIA Merlin consists of the following open source libraries:

  • NVTabular
  • HugeCTR
  • Triton Inference Server

NVTabular
NVTabular is a feature engineering and preprocessing library for tabular data. NVTabular is essentially the ETL component of the Merlin ecosystem. It is designed to quickly and easily manipulate terabyte-size datasets that are used to train deep learning based recommender systems. NVTabular offers a high-level API that can be used to define complex data transformation workflows. NVTabular is also capable of transformation speedups that can be 100 times to 1,000 times faster than transformations taking place on optimized CPU clusters. With NVTabular, you can:

  • prepare datasets quickly and easily for experimentation so that more models can be trained.
  • process datasets that exceed GPU and CPU memory without having to worry about scale.
  • focus on what to do with the data and not how to do it by using abstraction at the operation level.

NVTabular DataLoaders
NVTabular provides seamless integration with common deep learning frameworks, such as TensorFlow, PyTorch, and HugeCTR. When training deep learning recommender system models, dataloading can be a bottleneck. Therefore, we’ve developed custom, highly-optimized dataloaders to accelerate existing TensorFlow and PyTorch training pipelines. The NVTabular dataloaders can lead to a speedup that is nine times faster than the same training pipeline used with the GPU. With the NVTabular dataloaders, you can:

  • remove bottlenecks from dataloading by processing large chunks of data at a time instead of item by item.
  • process datasets that don’t fit within the GPU or CPU memory by streaming from the disk.
  • prepare batches asynchronously into the GPU to avoid CPU-GPU communication.
  • integrate easily into existing TensorFlow or PyTorch training pipelines by using a similar API.

HugeCTR
HugeCTR is a GPU-accelerated framework designed to estimate click-through rates and distribute training across multiple GPUs and nodes. HugeCTR contains optimized dataloaders that can be used to prepare batches with GPU-acceleration. In addition, HugeCTR is capable of scaling large deep learning recommendation models. The neural network architectures often contain large embedding tables that represent hundreds of millions of users and items. These embedding tables can easily exceed the CPU and GPU memory. HugeCTR provides strategies for scaling large embedding tables beyond available memory. With HugeCTR, you can:

  • scale embedding tables over multiple GPUs or nodes.
  • load a subset of an embedding table into the GPU in a coarse grained, on-demand manner during the training stage.

Triton
NVTabular and HugeCTR both support the Triton Inference Server to provide GPU-accelerated inference. The Triton Inference Server is open-source inference serving software that can be used to simplify the deployment of trained AI models from any framework to production. With Triton, you can:

  • deploy NVTabular ETL workflows and trained deep learning models to production with a few lines of code.
  • deploy an ensemble of NVTabular ETL and trained deep learning models to ensure that the same data transformations are applied in production.
  • deploy models concurrently on GPUs to maximize utilization.
  • enable low latency inferencing in real time or batch inferencing to maximize GPU and CPU utilization.
  • scale the production environment with Kubernetes for orchestration, metrics, and auto-scaling using a Docker container.

Examples

A collection of end-to-end examples is available within this repository in the form of Jupyter notebooks. The example notebooks demonstrate how to:

  • download and prepare the dataset.
  • use preprocessing and engineering features.
  • train deep learning recommendation models with TensorFlow, PyTorch, FastAI, or HugeCTR.
  • deploy the models to production.

These examples are based on different datasets and provide a wide range of real-world use cases.

Resources

For more information about NVIDIA Merlin and its components, see the following:

Comments
  • Update Criteo Example with Merlin Models and Merlin Systems

    Update Criteo Example with Merlin Models and Merlin Systems

    • Added the example notebook to train DLRM Merlin Model on criteo 1TB dataset.
    • Modified README.md, examples/README.md and examples/scaling-criteo/README.md to add link and information about the example.
    • Modified examples/scaling-criteo/02-ETL-with-NVTabular.ipynb to add Target and BINARY_CLASSIFICATION tag to the label.
    documentation examples 
    opened by rvk007 31
  • [BUG] Cannot load a exported deepfm model with NGC 22.03 inference container

    [BUG] Cannot load a exported deepfm model with NGC 22.03 inference container

    run into following errors

    I0318 00:00:18.082645 172 hugectr.cc:1926] TRITONBACKEND_ModelInstanceInitialize: deepfm_0 (device 0)
    I0318 00:00:18.082694 172 hugectr.cc:1566] Triton Model Instance Initialization on device 0
    I0318 00:00:18.082792 172 hugectr.cc:1576] Dense Feature buffer allocation:
    I0318 00:00:18.083026 172 hugectr.cc:1583] Categorical Feature buffer allocation:
    I0318 00:00:18.083095 172 hugectr.cc:1601] Categorical Row Index buffer allocation:
    I0318 00:00:18.083143 172 hugectr.cc:1611] Predict result buffer allocation:
    I0318 00:00:18.083203 172 hugectr.cc:1939] ******Loading HugeCTR Model******
    I0318 00:00:18.083217 172 hugectr.cc:1631] The model origin json configuration file path is: /ensemble_models/deepfm/1/deepfm.json
    [HCTR][00:00:18][INFO][RK0][main]: Global seed is 1305961709
    [HCTR][00:00:19][WARNING][RK0][main]: Peer-to-peer access cannot be fully enabled.
    [HCTR][00:00:19][INFO][RK0][main]: Start all2all warmup
    [HCTR][00:00:19][INFO][RK0][main]: End all2all warmup
    [HCTR][00:00:19][INFO][RK0][main]: Create inference session on device: 0
    [HCTR][00:00:19][INFO][RK0][main]: Model name: deepfm
    [HCTR][00:00:19][INFO][RK0][main]: Use mixed precision: False
    [HCTR][00:00:19][INFO][RK0][main]: Use cuda graph: True
    [HCTR][00:00:19][INFO][RK0][main]: Max batchsize: 64
    [HCTR][00:00:19][INFO][RK0][main]: Use I64 input key: True
    [HCTR][00:00:19][INFO][RK0][main]: start create embedding for inference
    [HCTR][00:00:19][INFO][RK0][main]: sparse_input name data1
    [HCTR][00:00:19][INFO][RK0][main]: create embedding for inference success
    [HCTR][00:00:19][INFO][RK0][main]: Inference stage skip BinaryCrossEntropyLoss layer, replaced by Sigmoid layer
    I0318 00:00:19.826815 172 hugectr.cc:1639] ******Loading HugeCTR model successfully
    I0318 00:00:19.827763 172 model_repository_manager.cc:1149] successfully loaded 'deepfm' version 1
    E0318 00:00:19.827767 172 model_repository_manager.cc:1152] failed to load 'deepfm_nvt' version 1: Internal: TypeError: 'NoneType' object is not subscriptable
    
    At:
      /ensemble_models/deepfm_nvt/1/model.py(91): _set_output_dtype
      /ensemble_models/deepfm_nvt/1/model.py(76): initialize
    
    E0318 00:00:19.827960 172 model_repository_manager.cc:1332] Invalid argument: ensemble 'deepfm_ens' depends on 'deepfm_nvt' which has no loaded version
    I0318 00:00:19.828048 172 server.cc:522]
    +------------------+------+
    | Repository Agent | Path |
    +------------------+------+
    +------------------+------+
    
    I0318 00:00:19.828117 172 server.cc:549]
    +---------+---------------------------------------------------------+-----------------------------------------------+
    | Backend | Path                                                    | Config                                        |
    +---------+---------------------------------------------------------+-----------------------------------------------+
    | hugectr | /opt/tritonserver/backends/hugectr/libtriton_hugectr.so | {"cmdline":{"ps":"/ensemble_models/ps.json"}} |
    +---------+---------------------------------------------------------+-----------------------------------------------+
    
    I0318 00:00:19.828209 172 server.cc:592]
    +------------+---------+--------------------------------------------------------------------------+
    | Model      | Version | Status                                                                   |
    +------------+---------+--------------------------------------------------------------------------+
    | deepfm     | 1       | READY                                                                    |
    | deepfm_nvt | 1       | UNAVAILABLE: Internal: TypeError: 'NoneType' object is not subscriptable |
    |            |         |                                                                          |
    |            |         | At:                                                                      |
    |            |         |   /ensemble_models/deepfm_nvt/1/model.py(91): _set_output_dtype          |
    |            |         |   /ensemble_models/deepfm_nvt/1/model.py(76): initialize                 |
    +------------+---------+--------------------------------------------------------------------------+
    
    I0318 00:00:19.845925 172 metrics.cc:623] Collecting metrics for GPU 0: Tesla T4
    I0318 00:00:19.846404 172 tritonserver.cc:1932]
    +----------------------------------+------------------------------------------------------------------------------------------------------------------------------------+
    | Option                           | Value                                                                                                                              |
    +----------------------------------+------------------------------------------------------------------------------------------------------------------------------------+
    | server_id                        | triton                                                                                                                             |
    | server_version                   | 2.19.0                                                                                                                             |
    | server_extensions                | classification sequence model_repository model_repository(unload_dependents) schedule_policy model_configuration system_shared_mem |
    |                                  | ory cuda_shared_memory binary_tensor_data statistics trace                                                                         |
    | model_repository_path[0]         | /ensemble_models                                                                                                                   |
    | model_control_mode               | MODE_NONE                                                                                                                          |
    | strict_model_config              | 1                                                                                                                                  |
    | rate_limit                       | OFF                                                                                                                                |
    | pinned_memory_pool_byte_size     | 268435456                                                                                                                          |
    | cuda_memory_pool_byte_size{0}    | 67108864                                                                                                                           |
    | response_cache_byte_size         | 0                                                                                                                                  |
    | min_supported_compute_capability | 6.0                                                                                                                                |
    | strict_readiness                 | 1                                                                                                                                  |
    | exit_timeout                     | 30                                                                                                                                 |
    +----------------------------------+------------------------------------------------------------------------------------------------------------------------------------+
    
    
    

    Aha! Link: https://nvaiinfa.aha.io/features/MERLIN-818

    bug 
    opened by mengdong 26
  • Add example notebook for training and inference on Sagemaker

    Add example notebook for training and inference on Sagemaker

    This PR adds an example notebook that demonstrates how to use Merlin container for training and inference on the AWS Sagemaker platform. It can be roughly split into three sections:

    1. Test the training script locally to make sure that the script runs as expected,
    2. Train the model on Sagemaker by calling sagemaker.Estimator.fit(entry_point='train.py'), and
    3. Create an endpoint with Triton, and test it by sending a request.

    Note:

    • The notebook builds a custom image (built from merlin-tensorflow image) and uploads it to ECR. This step can be simplified/removed by installing the sagemaker-training library directly in the Merlin image, so I opened a separate PR: https://github.com/NVIDIA-Merlin/Merlin/pull/707.

    Steps for testing (NV-specific):

    • Ask Ben to give you access to the AWS account. After he adds you to the group, wait a couple of hours for permissions to propagate.
    • If permissions have propagated, you will see AWS NVIDIA Account on https://myapps.microsoft.com/. This can be used to access the console.
    • Install aws-azure-login and run aws-azure-login --mode=gui to authenticate. This will create the credientials in $HOME/.aws. More details here: https://gitlab-master.nvidia.com/kaizen-efficiency/wiki/-/wikis/Guides/aws-cli-sso
    • As described in README, add -v $HOME/.aws:/root/.aws to your usual docker run command.
    examples 
    opened by edknv 25
  • Upgrade upstream container references, get pyarrow from upstream

    Upgrade upstream container references, get pyarrow from upstream

    This PR upgrades the current container references in the docker container to latest available for dlfw. The dockerfile also gets the latest compatible pyarrow directly from the upstream DLFW container and gets a library newly required by tritonserver,

    chore ci 
    opened by jperez999 21
  • [RMP] Recsys Tutorial & Demo - Flesh out the multi-stage recommender example architecture

    [RMP] Recsys Tutorial & Demo - Flesh out the multi-stage recommender example architecture

    Problem:

    Customers need a clear example of a multi-stage recommender pipeline that they can follow and use to create their own versions. The upcoming Recsys tutorial will be a public sharing of this example and serves as the deadline for its completion.

    Goal:

    • Provide a clear example of a multi stage recommender pipeline that does retrieval, filtering, ranking and ordering.
    • Highlight NVTabular, our dataloader, Merlin models, and Merlin systems and how they work together.

    Constraints:

    • NVTabular doesn't support splitting of workflows for user and item features. Two separate workflows for the data are needed to process each.

    Starting Point:

    • [ ] - [ ] https://github.com/NVIDIA-Merlin/systems/issues/99 ( This has been removd from scope)
    • [ ] #458
    • [ ] #449 - This is an optional requirement (removed)
    examples 
    opened by karlhigley 18
  • Migrate the legacy examples to the Merlin repo

    Migrate the legacy examples to the Merlin repo

    We may (or may not) want to keep these examples but they've overstayed their welcome in the NVTabular repo, which is burdened with the accumulation of a lot of historical cruft. Since some of these examples use inference code that's moving to Systems, it makes more sense for them to live in the Merlin repo (if we want to keep them.)

    The PR on the other side of this migration is https://github.com/NVIDIA-Merlin/NVTabular/pull/1711

    chore 
    opened by karlhigley 16
  • Unable to export graph as ensemble when using multi-hot categorical column

    Unable to export graph as ensemble when using multi-hot categorical column

    I am using inference container 22.05. When materializing feast in cell 6 deploying multi-stage recsys notebook, I am getting exception while materializing user_features. Item features worked just fine. Please download the sample data and notebooks used for this from the link here

    Using tensorflow-inference 22.05 container. Attaching the error here Screen Shot 2022-06-02 at 6 47 45 AM

    bug P0 
    opened by mkumari-ed 16
  • Pin pyarrow version for integration tests

    Pin pyarrow version for integration tests

    Pin pyarrow version for integration tests.

    The pip install 'feast<0.20' command is installing the latest version of pyarrow (currently 10.0.1). This is incompatibile with the current version of cudf.

    Follow-up

    The container we're testing with already has pyarrow installed. So a follow-up action would be to figure out why the existing pyarrow version is not being detected and take any actions required for this to be found.

    • import pyarrow; pyarrow.__version__ reports 8.0.0.
      • However, pip show pyarrow doesn't report anything.
      • Suggesting that the pyarrow installation is incomplete (missing the dist-info directory)
    ci 
    opened by oliverholworthy 14
  • [BUG] Cannot run example - Deploying a Multi-Stage Recommender System

    [BUG] Cannot run example - Deploying a Multi-Stage Recommender System

    Hi! Thanks for providing the multi-stage recommender system example. This is really helpful to learn to use Merlin in a more realistic scenario. I'm trying to run the two examples of Building-and-deploying-multi-stage-RecSys.

    My system CUDA driver is 11.5 and I'm using nvcr.io/nvidia/merlin/merlin-tensorflow-inference:latest.

    At first I had a problem similar to #158 and I solved it by following the suggestion in #158. More specifically, I downgraded the CUDA toolkit to 11.5 in the container. After this, I could successfully run example 01-Building-Recommender-Systems-with-Merlin.ipynb but I met a problem at almost the last step while running 02-Deploying-multi-stage-RecSys-with-Merlin-Systems.ipynb. 3221653589478_ pic As shown in the above figure, I can't start a triton server while all other steps in both examples seem normal.

    Could you please help me figure out a solution? Thanks!

    opened by future-xy 13
  • [wip] Split out Systems tests from notebooks

    [wip] Split out Systems tests from notebooks

    The idea here is to split some of our integration tests away from the notebook examples. This way we can write more complete tests without worrying about the narrative that we provide to new users.

    There is currently a single subclass of unittest.TestCase with a heavy setUpClass method that:

    • Generates synthetic data
    • Defines NVT workflos
    • Trains models (TT, DLRM)
    • Ingests data into Feast
    • Ingests item vectors into FAISS

    The individual tests configure various systems ensembles and ensures, for now, that they compile.

    I am having issues with the retrieval model signature including item features, which it shouldn't. I'm opening a draft PR to see how it works on CI.

    chore 
    opened by nv-alaiacano 12
  • fixes in the PoC first notebook

    fixes in the PoC first notebook

    This PR

    • fixing the correct raw dataframe to be fed to both models so the user and item catalogs wont be confused
    • training retrieval model first and then the ranking (logical order)
    enhancement examples 
    opened by rnyak 12
  • [QST]Fail to docker build from dockerfile.ctr, pull access denied for nvcr.io/nvstaging/merlin/merlin-base

    [QST]Fail to docker build from dockerfile.ctr, pull access denied for nvcr.io/nvstaging/merlin/merlin-base

    ❓ Questions & Help

    I've suceeded to login in nvcr.io, but still cannot pull from nvcr.io/nvstaging/merlin/merlin-base. Is this NGC still available now?

    cmd: cd docker && docker build --pull -t hugectr:devel -f ./dockerfile.ctr --build-arg RELEASE=false --build-arg RMM_VER=vnightly --build-arg CUDF_VER=vnightly --build-arg NVTAB_VER=vnightly --build-arg HUGECTR_DEV_MODE=true --no-cache .

    question 
    opened by heroes999 0
  • [INF] Unresolved architectural decisions

    [INF] Unresolved architectural decisions

    Problem:

    Merlin now has a bunch of libraries that need to interoperate smoothly, but a general lack of shared abstractions, conventions, and standards that would make that possible.

    Goal:

    • Build a solid foundation for the Merlin libraries via improvements in Core

    New Functionality

    • Core:
      • Shape in column schemas (for consistent tracking across libraries)
      • Cross-framework dtype translation (e.g. via Merlin dtypes)
      • Cross-framework data transfer via zero-copy protocols (for Columns and DictArrays -> Series and Dataframes)
      • Bespoke Merlin schema file format (i.e. a Protobuf schema for Merlin schema that isn't from Tensorflow Metadata)
    • Corresponding updates in all downstream libraries

    Constraints:

    • All functionality entailed by this issue has to work in and be adoptable by all Merlin libraries

    Starting Point:

    • [ ] Proposals:
      • [ ] Shapes
      • [ ] Dtypes
      • [ ] Data transfer
      • [ ] Schema file format
    roadmap infrastructure 
    opened by karlhigley 1
  • [Task] Use pre-commit for linting in GitHub Actions Workflow

    [Task] Use pre-commit for linting in GitHub Actions Workflow

    Description

    We have been using pre-commit in projects for linting in our local commits. However, we've been running checks using a different mechanism in our GitHub actions workflows.

    Updating to use pre-commit in a GitHub Actions workflow ensures that the checks we run locally are consistent with the ones we run in CI. Reducing the risk of version discrepancies in CI and Local development.

    • [X] Models https://github.com/NVIDIA-Merlin/models/pull/106
    • [X] Dataloader https://github.com/NVIDIA-Merlin/dataloader/pull/55
    • [X] Transformers4Rec https://github.com/NVIDIA-Merlin/Transformers4Rec/pull/545
    • [X] NVTabular https://github.com/NVIDIA-Merlin/NVTabular/pull/1723
    • [x] Core https://github.com/NVIDIA-Merlin/core/pull/184
    • [x] Systems https://github.com/NVIDIA-Merlin/systems/pull/254
    • [ ] Merlin
    chore ci 
    opened by oliverholworthy 0
  • [QST] how to import and call my model in Merlin ?

    [QST] how to import and call my model in Merlin ?

    ❓ Questions & Help

    Details

    I have developed and trained a candidate retrieval model by tensorflow. How to import, call and deploy the model to production from Merlin ?

    thanks

    question 
    opened by umusa 3
Releases(v22.11.00)
  • v22.11.00(Nov 22, 2022)

    What’s Changed

    🐜 Bug Fixes

    • Update dockerfile.ci to find the NVT dev requirements file @karlhigley (#740)
    • Restrict cmake<3.25.0 to avoid an issue finding CUDA toolkit @karlhigley (#739)
    • add required dataloader dependency to run unit tests for dataloader @jperez999 (#735)
    • anchor xgboost to 1.6.2 to make tests pass @jperez999 (#726)
    • adding new metric loss_batch from merlin models @jperez999 (#727)

    📄 Documentation

    • Add example notebook for training and inference on Sagemaker @edknv (#692)

    🔧 Maintenance

    • Add Jenkinsfile @AyodeAwe (#734)
    • Update dockerfile.ci to find the NVT dev requirements file @karlhigley (#740)
    • Restrict cmake<3.25.0 to avoid an issue finding CUDA toolkit @karlhigley (#739)
    • add required dataloader dependency to run unit tests for dataloader @jperez999 (#735)
    • add dataloader unit testing to container run @jperez999 (#728)
    • Increase timeout of multi-stage notebook test from 120 to 180 @oliverholworthy (#729)
    • anchor xgboost to 1.6.2 to make tests pass @jperez999 (#726)
    • adding new metric loss_batch from merlin models @jperez999 (#727)
    Source code(tar.gz)
    Source code(zip)
  • untagged-c73c3c7a63c75e538917(Nov 9, 2022)

    What’s Changed

    • Remove nvtabular backend @jperez999 (#378)
    • Adding instructions for running the E2E example on CPU @AshishSardana (#332)
    • add fastai to torch dockerfile @jperez999 (#363)
    • Fail the build if tritonserver is missing from inference containers @benfred (#358)
    • Use python setup.py install to install SOK @jperez999 (#357)
    • hugectr dockerfile update @jperez999 (#355)
    • Add sparse operation kit and distributed embeddings to tf image @jperez999 (#354)
    • Update hugectr and torch dockerfiles @jperez999 (#353)
    • Update merlin base dockerfile for 22.05 @jperez999 (#348)
    • Update Arrow and triton versions @jperez999 (#347)
    • Sys tests fix @jperez999 (#342)
    • add Tf keras to dockerfile @jperez999 (#341)
    • Fix ci routes @jperez999 (#331)
    • Ci update @jperez999 (#337)
    • Update CI dockerfile with new cuda keyring and keys for repo @jperez999 (#334)
    • Fix CVE @jperez999 (#328)
    • Distributed embeddings failing to install because of PYTHONPATH Edit before hand @jperez999 (#326)
    • Fix for 'import error' in merlin-tensorflow-training:22.05 @EmmaQiaoCh (#317)
    • Merlin rm bad key @jperez999 (#316)
    • Merlin rm bad key @jperez999 (#314)
    • Remove bad key for Nvidia apt repositories @jperez999 (#311)
    • Add torch inf @jperez999 (#307)
    • Add torch inf @jperez999 (#305)
    • Remove merlin PYTHONPATH edits @jperez999 (#304)
    • remove horovod upgrade to fix tf perf issue @zehuanw (#302)
    • Add container infos to notebooks @bschifferer (#298)
    • Rm pip e @jperez999 (#300)

    ⚠ Breaking Changes

    • Update for change of hugectr branch name @EmmaQiaoCh (#705)
    • Fix ci order @jperez999 (#581)
    • Fix int test @jperez999 (#578)
    • Add numpy anchor version after all package building and updates @jperez999 (#566)
    • Fix unit scaling criteo inference serving @jperez999 (#559)
    • Moving loss and metrics to model.compile @oliverholworthy (#340)

    🐜 Bug Fixes

    • adding dataloader repo to dockerfile @jperez999 (#722)
    • add in reinstall for dask and distributed after feast install @jperez999 (#713)
    • create resilient directory rm @jperez999 (#718)
    • fix missing lib issue by adding from upstream @jperez999 (#689)
    • get numba from upstream container @jperez999 (#690)
    • Adding git pull command to nightly docker to ensure latest commit @jperez999 (#617)
    • Fix unit scaling criteo inference serving @jperez999 (#559)
    • Fix integration test and update NB @radekosmulski (#544)
    • Fix nightly container builds @benfred (#518)
    • Fix typo in CI script file name @karlhigley (#445)
    • remove extra CMD from containers @jperez999 (#390)
    • Add pytorch tritonserver backend to ci dockerfile @jperez999 (#441)
    • updates from entrypoint and cupy cuda116 @jperez999 (#389)
    • Add matplotlib to torch container @jperez999 (#386)
    • Add nvt backend back into containers @jperez999 (#382)
    • Remove entrypoint from inference/training images. @oliverholworthy (#336)
    • fix: "illegal instruction" error in hugectr test jobs @EmmaQiaoCh (#295)

    🚀 Features

    • Add Dropna to remove nulls in the dataset that creates error in the integration test of multi-stage deployment nb @rnyak (#629)
    • Update multi-stage deployment notebooks and integration test @rnyak (#627)
    • modify multi-stage deployment nbs wrt recent changes in systems  @rnyak (#621)
    • Changes for hugectr @EmmaQiaoCh (#502)
    • fixes in the PoC first notebook @rnyak (#487)
    • Switch over HDFS build/install scripts @bashimao (#434)
    • Add jupyter ENVs in case launching container with normal user @EmmaQiaoCh (#439)
    • fix poc nbs and move poc unit test @rnyak (#387)
    • Moving loss and metrics to model.compile @oliverholworthy (#340)

    📄 Documentation

    • docs: semver to calver banner @mikemckiernan (#715)
    • docs: Add configuration for SEO @mikemckiernan (#723)
    • docs: Remove SM from the support matrix @mikemckiernan (#721)
    • Support matrix updates for 22.10 @nvidia-merlin-bot (#719)
    • fix: Update HugeCTR version to 4.1.1 @mikemckiernan (#717)
    • Support matrix updates for 22.09 @nvidia-merlin-bot (#711)
    • Support matrix updates for 22.10 @nvidia-merlin-bot (#710)
    • fix links in Multi-stage Bulding-and-deployment nbs @rnyak (#697)
    • Support matrix updates for 22.09 @nvidia-merlin-bot (#636)
    • Address virtual dev review comments @mikemckiernan (#626)
    • modify multi-stage deployment nbs wrt recent changes in systems  @rnyak (#621)
    • Add a "New Functionality" section to the Roadmap issue template @karlhigley (#596)
    • Update Logos of all examples @bschifferer (#569)
    • Update Merlin libs graphic @mikemckiernan (#560)
    • Support matrix updates for 22.06 @nvidia-merlin-bot (#500)
    • Support matrix updates for 22.07 @nvidia-merlin-bot (#499)
    • Get TF version from Python and not Pip @mikemckiernan (#498)
    • docs: Readability improvements @ryanrussell (#440)
    • Update Criteo Example with Merlin Models and Merlin Systems @rvk007 (#380)
    • fixes in the PoC first notebook @rnyak (#487)
    • Remove unnecessary dependencies from docs builds @mikemckiernan (#466)
    • add integration tests @radekosmulski (#310)
    • Update README with installation steps @rvk007 (#430)
    • Support matrix updates for 22.06 @nvidia-merlin-bot (#435)
    • Add NGC overview descriptions for our containers @benfred (#399)
    • Add run timestamp to data @mikemckiernan (#415)
    • Update URLs for Criteo dataset @mikemckiernan (#400)
    • Document the change to three containers @mikemckiernan (#379)
    • fix poc nbs and move poc unit test @rnyak (#387)
    • Add issue templates @karlhigley (#345)
    • Hand-edit the HugeCTR versions @mikemckiernan (#329)
    • Support matrix updates for 22.05 @nvidia-merlin-bot (#321)
    • chore: Add release-drafter @mikemckiernan (#308)

    🔧 Maintenance

    • adding in instructions from the jenkins dockerfile to keep in one place @jperez999 (#725)
    • Pass mpi environment variable to tox @edknv (#724)
    • adding dataloader repo to dockerfile @jperez999 (#722)
    • docs: Add two projects to the support matrix @mikemckiernan (#720)
    • add in reinstall for dask and distributed after feast install @jperez999 (#713)
    • create resilient directory rm @jperez999 (#718)
    • Install horovod in the ci-runner image for CI testing @edknv (#712)
    • Update for change of hugectr branch name @EmmaQiaoCh (#705)
    • Add dist-info directories for packages copied in the tensorflow image. @oliverholworthy (#704)
    • revert change from 2209 requirements @jperez999 (#701)
    • revert to 22.08 @jperez999 (#698)
    • add llvmlite to base for numba dep @jperez999 (#691)
    • fix missing lib issue by adding from upstream @jperez999 (#689)
    • get numba from upstream container @jperez999 (#690)
    • update TFDE build @FDecaYed (#659)
    • Upgrade upstream container references, get pyarrow from upstream @jperez999 (#656)
    • fix: RMM and cuDF are no longer installed with pip @mikemckiernan (#637)
    • Adding hugectr to nightly build dockerfile @jperez999 (#632)
    • Increase timeout of unit test for second multi-stage notebook @oliverholworthy (#630)
    • Add Dropna to remove nulls in the dataset that creates error in the integration test of multi-stage deployment nb @rnyak (#629)
    • Enable unittest for 2stage notebooks @bschifferer (#628)
    • Update paths using BASE_DIR in multi-stage notebook to handle non-default value @oliverholworthy (#622)
    • Update multi-stage deployment notebooks and integration test @rnyak (#627)
    • anchor tf version to avoid errors in 2.10.0 libnvinfer look ups @jperez999 (#625)
    • Revert to working tritonserver call in notebook using testbooks @jperez999 (#619)
    • Remove lint env from tox config @karlhigley (#624)
    • Add a tox config file @karlhigley (#623)
    • Update ci container to use nightly base container @jperez999 (#620)
    • Adding git pull command to nightly docker to ensure latest commit @jperez999 (#617)
    • Add dock nite @jperez999 (#616)
    • Add dockerfile for nightly builds @jperez999 (#615)
    • Args mv @jperez999 (#603)
    • Args mv @jperez999 (#602)
    • anchor Tf version @jperez999 (#601)
    • Add a "New Functionality" section to the Roadmap issue template @karlhigley (#596)
    • Don't install faiss with the integration tests @benfred (#591)
    • Install numpy for building faiss @benfred (#590)
    • Fixes for faiss install @benfred (#587)
    • Include a SM80 enabled version of faiss on merlin-base container @benfred (#584)
    • Rearrange testing for faster feedback @jperez999 (#583)
    • Second fix @jperez999 (#582)
    • Update triage github actions workflow @benfred (#580)
    • Fix ci order @jperez999 (#581)
    • Reduce the size of synthetic data used in Criteo test @karlhigley (#579)
    • Fix int test @jperez999 (#578)
    • Fix int + unit tests @jperez999 (#577)
    • Add hadoop envs @EmmaQiaoCh (#565)
    • Add numpy anchor version after all package building and updates @jperez999 (#566)
    • Docker edit @jperez999 (#564)
    • Docker fix @jperez999 (#563)
    • Add skip tf crit unit @jperez999 (#561)
    • Fix unit scaling criteo inference serving @jperez999 (#559)
    • update base dockerfile @benfred (#556)
    • Update the Merlin repos in the CI image build @karlhigley (#558)
    • Remove dependencies of hugectr & install hps tf plugin to merlin-tf @EmmaQiaoCh (#549)
    • Skip the multi-stage example notebook (for now) @karlhigley (#543)
    • Skip multi-stage example integration test (for now) @karlhigley (#541)
    • Revert CMake changes @karlhigley (#538)
    • Install CMake in Merlin base image (instead of copying from build) @karlhigley (#524)
    • Mark "scaling Criteo" notebook to be skipped without TF @karlhigley (#537)
    • install tox in base image @nv-alaiacano (#532)
    • Pin fsspec==22.5.0 directly in the Merlin base image @karlhigley (#533)
    • Remove duplicate CMake installs @karlhigley (#523)
    • Make integration test script executable @karlhigley (#521)
    • Install Feast/Faiss before running Merlin integration tests @karlhigley (#520)
    • Downgrade onnxruntime due to security issue of mpmath @EmmaQiaoCh (#486)
    • Convert the default text in the roadmap issue template to comments @karlhigley (#483)
    • Add tests in the Merlin repo to the CI test scripts @karlhigley (#450)
    • Remove unnecessary dependencies from docs builds @mikemckiernan (#466)
    • Adds integration tests to Merlin Models @gabrielspmoreira (#438)
    • Add wandb @jperez999 (#470)
    • added fiddle to container for models testing @jperez999 (#469)
    • add integration tests @radekosmulski (#310)
    • Refactor the container test script to run all the SW checks, unit tests, or integration tests before failing @karlhigley (#444)
    • Switch over HDFS build/install scripts @bashimao (#434)
    • remove extra CMD from containers @jperez999 (#390)
    • Add pytorch tritonserver backend to ci dockerfile @jperez999 (#441)
    • Add jupyter ENVs in case launching container with normal user @EmmaQiaoCh (#439)
    • remove excess python path setting @jperez999 (#432)
    • Remove stale doc reviews @mikemckiernan (#417)
    • Fix failing treelite install @jperez999 (#416)
    • Add FIL support to Base container, Add e2e support in ci container @jperez999 (#414)
    • Always run NVT integration tests @benfred (#401)
    • Inline hugectr container tests @benfred (#398)
    • Update test_container.sh script @benfred (#396)
    • Restrict tritonclient to 2.22.0 @EmmaQiaoCh (#391)
    • Add sok test new @EmmaQiaoCh (#384)
    • Add skip checks to examples tests @jperez999 (#392)
    • updates from entrypoint and cupy cuda116 @jperez999 (#389)
    • Add matplotlib to torch container @jperez999 (#386)
    • Rm old docker @jperez999 (#383)
    • Add a GA workflow that requires labels on PR's @benfred (#381)
    • Support matrix updates for 22.05 @nvidia-merlin-bot (#352)
    • Support matrix updates for 21.11 @nvidia-merlin-bot (#372)
    • Support matrix updates for 21.09 @nvidia-merlin-bot (#373)
    • Support matrix updates for 22.03 @nvidia-merlin-bot (#368)
    • Support matrix updates for 22.02 @nvidia-merlin-bot (#370)
    • Support matrix updates for 22.01 @nvidia-merlin-bot (#369)
    • Support matrix updates for 21.12 @nvidia-merlin-bot (#371)
    • Refactor SMX for blossom @mikemckiernan (#351)
    • Support matrix updates for 22.04 @nvidia-merlin-bot (#367)
    • Update for hdfs @EmmaQiaoCh (#365)
    • Enable running the SMX data job in Blossom @mikemckiernan (#325)
    • Request that PRs are labeled @mikemckiernan (#327)
    • Also use horovod from DLFW for hugectr training container @benfred (#303)
    Source code(tar.gz)
    Source code(zip)
  • v22.05(Jun 9, 2022)

    What's Changed

    • Removes --user when installing NVTabular by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/64
    • Fix typos in docker files by @benfred in https://github.com/NVIDIA-Merlin/Merlin/pull/68
    • Fix link in README and update merlin examples by @benfred in https://github.com/NVIDIA-Merlin/Merlin/pull/70
    • Adds package and fixes typo by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/69
    • 21.11 DLFW by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/73
    • update unified container to support hugectr v3.3 by @zehuanw in https://github.com/NVIDIA-Merlin/Merlin/pull/76
    • fix illegal instruction by adding "PORTABLE=1" by @zehuanw in https://github.com/NVIDIA-Merlin/Merlin/pull/77
    • Fixing the issue of missing install.sh in tf container build by @zehuanw in https://github.com/NVIDIA-Merlin/Merlin/pull/79
    • Rel 21.12 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/78
    • Add new issues to the backlog project by @benfred in https://github.com/NVIDIA-Merlin/Merlin/pull/80
    • Updates dlfw 21.12 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/86
    • Cuda compat removes shell by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/89
    • [Ready to be reviewed] Support Customized HCTR Repo in the unified containers by @zehuanw in https://github.com/NVIDIA-Merlin/Merlin/pull/85
    • mask_nvt_for_tf1_image by @shijieliu in https://github.com/NVIDIA-Merlin/Merlin/pull/92
    • added hdfs and s3 support by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/94
    • Arrow s3 hdfs by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/95
    • activate orc by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/96
    • Release 22.02 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/93
    • Arrow s3 hdfs by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/98
    • Reduce containers size by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/100
    • Fix security issue jershi by @jershi425 in https://github.com/NVIDIA-Merlin/Merlin/pull/123
    • add mpi4py/onnx/onnxruntime for ctr/tf by @shijieliu in https://github.com/NVIDIA-Merlin/Merlin/pull/122
    • New containers by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/119
    • Rel 22.03 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/124
    • CI container by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/132
    • Ci container fixes by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/133
    • Merlin Release 22.03 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/147
    • Software Versions Tools by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/141
    • Fix inference container by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/135
    • Inference PoC with Merlin Systems by @rnyak in https://github.com/NVIDIA-Merlin/Merlin/pull/169
    • Update the Merlin README with additional edits from docs bash by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/167
    • Copy MovieLens and Criteo example notebooks from NVTabular by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/166
    • docs: add a docs build by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/174
    • docs: Add documentation badge by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/175
    • add basic tests by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/176
    • Initial attempt at support matrices by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/182
    • Updating Example README by @bschifferer in https://github.com/NVIDIA-Merlin/Merlin/pull/183
    • docs: Add redirect page by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/184
    • Update deploying multi-stage RecSys PoC nb by @rnyak in https://github.com/NVIDIA-Merlin/Merlin/pull/186
    • Merlin Container Release 22.04 by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/178
    • Set pandas version by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/188
    • Print more testing info by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/190
    • Test merlin-systems as part of verifying containers by @benfred in https://github.com/NVIDIA-Merlin/Merlin/pull/191
    • Update packages for scan by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/192
    • Set dask_cuda version by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/193
    • Keep HugeCTR source by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/195
    • Typo by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/198
    • removed path from feast as updated in systems by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/196
    • Fail on errors in test_container.sh by @benfred in https://github.com/NVIDIA-Merlin/Merlin/pull/194
    • docs: Add 22.04 support matrix by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/201
    • Set pandas version in CI container by @albert17 in https://github.com/NVIDIA-Merlin/Merlin/pull/200
    • fix torch horovod: explicitly does not need tensorflow by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/206
    • Torch fix by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/207
    • Torch container horovod release version by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/215
    • Add Arrow S3 support to the tensorflow-training container by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/217
    • add distributed-embeddings to merlin dockerfile by @FDecaYed in https://github.com/NVIDIA-Merlin/Merlin/pull/208
    • Revert "Add Arrow S3 support to the tensorflow-training container (… by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/220
    • Install required packages for multi-stage notebooks using %pip by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/221
    • WIP: Fix HDFS linking by @bashimao in https://github.com/NVIDIA-Merlin/Merlin/pull/151
    • Restructure container builds to use multi-stage builds and a Merlin base image by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/234
    • add args, ARG can only handle one arg at a time by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/245
    • Args fix by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/246
    • Replace run_ensemble_on_tritonserver() with send_triton_request() and do minor updates by @rnyak in https://github.com/NVIDIA-Merlin/Merlin/pull/244
    • Add a few Python and system packages for the example notebooks by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/247
    • remove horovod from torch container by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/249
    • Apply key rotation fixes to existing container Dockerfiles by @karlhigley in https://github.com/NVIDIA-Merlin/Merlin/pull/252
    • add entrypoint to all containers by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/254
    • Add hadoop scripts by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/260
    • Copy hadoop by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/261
    • Rm hadoop xtra by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/262
    • Rm e pip by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/283
    • Add Unit test for Building and deploying multi-stage Recsys nbs by @rnyak in https://github.com/NVIDIA-Merlin/Merlin/pull/281
    • docs: Use MyST and sphinx-external-toc by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/251
    • Add an ARG which could disable distributed_embeddings by @EmmaQiaoCh in https://github.com/NVIDIA-Merlin/Merlin/pull/291
    • Add Unit test for Building and deploying multi-stage Recsys nbs by @rnyak in https://github.com/NVIDIA-Merlin/Merlin/pull/288
    • docs: Tooling and automation for support matrix by @mikemckiernan in https://github.com/NVIDIA-Merlin/Merlin/pull/290
    • Add key update mechanism for ci dockerfile by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/292
    • Rm nvm by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/296
    • Revert rm by @jperez999 in https://github.com/NVIDIA-Merlin/Merlin/pull/297

    New Contributors

    • @jershi425 made their first contribution in https://github.com/NVIDIA-Merlin/Merlin/pull/123
    • @rnyak made their first contribution in https://github.com/NVIDIA-Merlin/Merlin/pull/169
    • @FDecaYed made their first contribution in https://github.com/NVIDIA-Merlin/Merlin/pull/208
    • @bashimao made their first contribution in https://github.com/NVIDIA-Merlin/Merlin/pull/151

    Full Changelog: https://github.com/NVIDIA-Merlin/Merlin/compare/v0.7.1...v22.05

    Source code(tar.gz)
    Source code(zip)
Owner
Merlin is a framework providing end-to-end GPU-accelerated recommender systems, from feature engineering to deep learning training and deploying to production
First-Order Probabilistic Programming Language

FOPPL: A First-Order Probabilistic Programming Language This is an implementation of FOPPL, an S-expression based probabilistic programming language d

Renato Costa 23 Dec 20, 2022
This repository contains the code for the CVPR 2021 paper "GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields"

GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields Project Page | Paper | Supplementary | Video | Slides | Blog | Talk If

1.1k Dec 30, 2022
Web service for facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation based on OpenFace 2.0

OpenGaze: Web Service for OpenFace Facial Behaviour Analysis Toolkit Overview OpenFace is a fantastic tool intended for computer vision and machine le

Sayom Shakib 4 Nov 03, 2022
ByteTrack(Multi-Object Tracking by Associating Every Detection Box)のPythonでのONNX推論サンプル

ByteTrack-ONNX-Sample ByteTrack(Multi-Object Tracking by Associating Every Detection Box)のPythonでのONNX推論サンプルです。 ONNXに変換したモデルも同梱しています。 変換自体を試したい方はByteT

KazuhitoTakahashi 16 Oct 26, 2022
A tensorflow=1.13 implementation of Deconvolutional Networks on Graph Data (NeurIPS 2021)

GDN A tensorflow=1.13 implementation of Deconvolutional Networks on Graph Data (NeurIPS 2021) Abstract In this paper, we consider an inverse problem i

4 Sep 13, 2022
Semi-Supervised Semantic Segmentation via Adaptive Equalization Learning, NeurIPS 2021 (Spotlight)

Semi-Supervised Semantic Segmentation via Adaptive Equalization Learning, NeurIPS 2021 (Spotlight) Abstract Due to the limited and even imbalanced dat

Hanzhe Hu 99 Dec 12, 2022
Pytorch implementation of paper "Efficient Nearest Neighbor Language Models" (EMNLP 2021)

Pytorch implementation of paper "Efficient Nearest Neighbor Language Models" (EMNLP 2021)

Junxian He 57 Jan 01, 2023
A clean and robust Pytorch implementation of PPO on continuous action space.

PPO-Continuous-Pytorch I found the current implementation of PPO on continuous action space is whether somewhat complicated or not stable. And this is

XinJingHao 56 Dec 16, 2022
This is an official implementation for "SimMIM: A Simple Framework for Masked Image Modeling".

Project This repo has been populated by an initial template to help get you started. Please make sure to update the content to build a great experienc

Microsoft 674 Dec 26, 2022
Automatically download the cwru data set, and then divide it into training data set and test data set

Automatically download the cwru data set, and then divide it into training data set and test data set.自动下载cwru数据集,然后分训练数据集和测试数据集

6 Jun 27, 2022
Repo for "Event-Stream Representation for Human Gaits Identification Using Deep Neural Networks"

Summary This is the code for the paper Event-Stream Representation for Human Gaits Identification Using Deep Neural Networks by Yanxiang Wang, Xian Zh

zhangxian 54 Jan 03, 2023
This repo is the official implementation of "L2ight: Enabling On-Chip Learning for Optical Neural Networks via Efficient in-situ Subspace Optimization".

L2ight is a closed-loop ONN on-chip learning framework to enable scalable ONN mapping and efficient in-situ learning. L2ight adopts a three-stage learning flow that first calibrates the complicated p

Jiaqi Gu 9 Jul 14, 2022
PoolFormer: MetaFormer is Actually What You Need for Vision

PoolFormer: MetaFormer is Actually What You Need for Vision (arXiv) This is a PyTorch implementation of PoolFormer proposed by our paper "MetaFormer i

Sea AI Lab 1k Dec 30, 2022
HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision

HugsVision is an open-source and easy to use all-in-one huggingface wrapper for computer vision. The goal is to create a fast, flexible and user-frien

Labrak Yanis 166 Nov 27, 2022
FAMIE is a comprehensive and efficient active learning (AL) toolkit for multilingual information extraction (IE)

FAMIE: A Fast Active Learning Framework for Multilingual Information Extraction

18 Sep 01, 2022
Evolving neural network parameters in JAX.

Evolving Neural Networks in JAX This repository holds code displaying techniques for applying evolutionary network training strategies in JAX. Each sc

Trevor Thackston 6 Feb 12, 2022
PyTorch implementation of ICLR 2022 paper PiCO: Contrastive Label Disambiguation for Partial Label Learning

PiCO: Contrastive Label Disambiguation for Partial Label Learning This is a PyTorch implementation of ICLR 2022 paper PiCO: Contrastive Label Disambig

王皓波 147 Jan 07, 2023
EMNLP 2021: Single-dataset Experts for Multi-dataset Question-Answering

MADE (Multi-Adapter Dataset Experts) This repository contains the implementation of MADE (Multi-adapter dataset experts), which is described in the pa

Princeton Natural Language Processing 68 Jul 18, 2022
It's a implement of this paper:Relation extraction via Multi-Level attention CNNs

Relation Classification via Multi-Level Attention CNNs It's a implement of this paper:Relation Classification via Multi-Level Attention CNNs. Training

Aybss 2 Nov 04, 2022
Brax is a differentiable physics engine that simulates environments made up of rigid bodies, joints, and actuators

Brax is a differentiable physics engine that simulates environments made up of rigid bodies, joints, and actuators. It's also a suite of learning algorithms to train agents to operate in these enviro

Google 1.5k Jan 02, 2023