Like Dirt-Samples, but cleaned up

Overview

Clean-Samples

Like Dirt-Samples, but cleaned up, with clear provenance and license info (generally a permissive creative commons licence but check the metadata for specifics).

The bin/meta.py python script is a reference implementation that can make a '.cleanmeta' metadata file for your own sample pack folder. See below for how to use it and contribute a sample pack of your own.

If you want to use these outside the Tidal/SuperDirt/SuperCollider ecosystem you are very welcome. You're encouraged to join discussion in the github issue tracker so that we can develop a standard way to share and index/signpost these packs.

See /tidalcycles/sounds-repetition for an example sample pack which has two sets of samples in it.

How to contribute a sample pack

Please only contribute samples if you are happy to share them under a permissive license such as CC0 or a similar creative commons license.

If you are unfamiliar with the 'git' software, please create an issue here, with a short description of your samples and a link to them and someone should be along to help shortly.

If you are familiar with git and running python scripts (or happy to learn), please follow the below instructions. This is all new - if anything is unclear please create an issue, thanks!

  1. Get your samples together in .wav format, editing them if necessary (see below for advice).

  2. Create a new repository. This isn't essential, but consider putting 'sounds-' in front of its name, e.g. 'sounds-303bass' for your 303 bass samples.

  3. Add your samples to the repository. For an example of how to organise them, see this sample pack: tidalcycles/sounds-repetition, which has two sets of samples, with a subfolder for each.

  4. Create a '.cleanmeta' metadata file for each subfolder. Again, see tidalcycles/sounds-repetition for examples. There is a python script bin/meta.py which can generate the metadata file for you, run it without parameters for help. Here is an example commandline, that was used to generate repetition.cleanmeta:

    ../Clean-Samples/bin/meta.py --maintainer alex --email [email protected] --copyright "(c) 2021 Alex McLean" --license CC0 --provenance "Various dodgy speech synths" --shortname repetition --sample-subfolder repetition/ --write .
    

    After generating the file, edit it with a text editor to fill in any missing info.

  5. When ready, add te URL of your repository to the https://github.com/tidalcycles/Clean-Samples/blob/main/Clean-Samples.quark for the Clean-Samples quark) in a pull request. You could also add it to the SuperCollider quarks database, or we can do that for you if you prefer, so that we can accept the PR to Clean-Samples once it's accepted as a quark.

Advice for preparing samples

You can use free/open source software like audacity for editing samples.

As a minimum, be sure to trim any silence from beginning/end of the samples, and that the start and end of the sample is at zero to avoid clicks (you might need to fade in / fade out by a tiny amount to achieve this).

Consider adjusting the volume/loudness too, for example normalising to -1.0db - but this is very subjective and will depend on the nature of the samples and the music they're used with. For example distorted gabba samples are intended to be very loud, and a whisper is intended to sound silent. The average non-percussive sample should be around -23dB RMS. Samples shouldn't exceed 0dB true peak. EBU recommends -1dBTP at 4x-oversampling. Samples generally shouldn't have DC offset, although e.g. some kick drum samples naturally have non-zero mean.

For more advice, you could join the discussion here.

Thanks!

Owner
TidalCycles
Live coding environment for making patterns
TidalCycles
Pytorch codes for "Self-supervised Multi-view Stereo via Effective Co-Segmentation and Data-Augmentation"

Self-Supervised-MVS This repository is the official PyTorch implementation of our AAAI 2021 paper: "Self-supervised Multi-view Stereo via Effective Co

hongbin_xu 127 Jan 04, 2023
Jingju baseline - A baseline model of our project of Beijing opera script generation

Jingju Baseline It is a baseline of our project about Beijing opera script gener

midon 1 Jan 14, 2022
《Lerning n Intrinsic Grment Spce for Interctive Authoring of Grment Animtion》

Learning an Intrinsic Garment Space for Interactive Authoring of Garment Animation Overview This is the demo code for training a motion invariant enco

YuanBo 213 Dec 14, 2022
🦙 LaMa Image Inpainting, Resolution-robust Large Mask Inpainting with Fourier Convolutions, WACV 2022

🦙 LaMa Image Inpainting, Resolution-robust Large Mask Inpainting with Fourier Convolutions, WACV 2022

Advanced Image Manipulation Lab @ Samsung AI Center Moscow 4.7k Dec 31, 2022
A Review of Deep Learning Techniques for Markerless Human Motion on Synthetic Datasets

HOW TO USE THIS PROJECT A Review of Deep Learning Techniques for Markerless Human Motion on Synthetic Datasets Based on DeepLabCut toolbox, we run wit

1 Jan 10, 2022
Implementation of gaze tracking and demo

Predicting Customer Demand by Using Gaze Detecting and Object Tracking This project is the integration of gaze detecting and object tracking. Predict

2 Oct 20, 2022
Based on Stockfish neural network(similar to LcZero)

MarcoEngine Marco Engine - interesnaya neyronnaya shakhmatnaya set', kotoraya ispol'zuyet metod samoobucheniya(dostizheniye khoroshoy igy putem proboy

Marcus Kemaul 4 Mar 12, 2022
ColossalAI-Examples - Examples of training models with hybrid parallelism using ColossalAI

ColossalAI-Examples This repository contains examples of training models with Co

HPC-AI Tech 185 Jan 09, 2023
Relative Human dataset, CVPR 2022

Relative Human (RH) contains multi-person in-the-wild RGB images with rich human annotations, including: Depth layers (DLs): relative depth relationsh

Yu Sun 112 Dec 02, 2022
Source code for "Roto-translated Local Coordinate Framesfor Interacting Dynamical Systems"

Roto-translated Local Coordinate Frames for Interacting Dynamical Systems Source code for Roto-translated Local Coordinate Frames for Interacting Dyna

Miltiadis Kofinas 19 Nov 27, 2022
PyDeepFakeDet is an integrated and scalable tool for Deepfake detection.

PyDeepFakeDet An integrated and scalable library for Deepfake detection research. Introduction PyDeepFakeDet is an integrated and scalable Deepfake de

Junke, Wang 49 Dec 11, 2022
PyContinual (An Easy and Extendible Framework for Continual Learning)

PyContinual (An Easy and Extendible Framework for Continual Learning) Easy to Use You can sumply change the baseline, backbone and task, and then read

Zixuan Ke 176 Jan 05, 2023
Accelerated SMPL operation, commonly used in generate 3D human mesh, STAR included.

SMPL2 An enchanced and accelerated SMPL operation which commonly used in 3D human mesh generation. It takes a poses, shapes, cam_trans as inputs, outp

JinTian 20 Oct 17, 2022
FG-transformer-TTS Fine-grained style control in transformer-based text-to-speech synthesis

LST-TTS Official implementation for the paper Fine-grained style control in transformer-based text-to-speech synthesis. Submitted to ICASSP 2022. Audi

Li-Wei Chen 64 Dec 30, 2022
Code for ACL 2019 Paper: "COMET: Commonsense Transformers for Automatic Knowledge Graph Construction"

To run a generation experiment (either conceptnet or atomic), follow these instructions: First Steps First clone, the repo: git clone https://github.c

Antoine Bosselut 575 Jan 01, 2023
Implementation of CVPR 2021 paper "Spatially-invariant Style-codes Controlled Makeup Transfer"

SCGAN Implementation of CVPR 2021 paper "Spatially-invariant Style-codes Controlled Makeup Transfer" Prepare The pre-trained model is avaiable at http

118 Dec 12, 2022
The official implementation of Equalization Loss for Long-Tailed Object Recognition (CVPR 2020) based on Detectron2

Equalization Loss for Long-Tailed Object Recognition Jingru Tan, Changbao Wang, Buyu Li, Quanquan Li, Wanli Ouyang, Changqing Yin, Junjie Yan ⚠️ We re

Jingru Tan 197 Dec 25, 2022
Code for our CVPR 2021 Paper "Rethinking Style Transfer: From Pixels to Parameterized Brushstrokes".

Rethinking Style Transfer: From Pixels to Parameterized Brushstrokes (CVPR 2021) Project page | Paper | Colab | Colab for Drawing App Rethinking Style

CompVis Heidelberg 153 Jan 04, 2023
Code for our paper "Graph Pre-training for AMR Parsing and Generation" in ACL2022

AMRBART An implementation for ACL2022 paper "Graph Pre-training for AMR Parsing and Generation". You may find our paper here (Arxiv). Requirements pyt

xfbai 60 Jan 03, 2023
A distributed deep learning framework that supports flexible parallelization strategies.

FlexFlow FlexFlow is a deep learning framework that accelerates distributed DNN training by automatically searching for efficient parallelization stra

528 Dec 25, 2022