Like Dirt-Samples, but cleaned up

Overview

Clean-Samples

Like Dirt-Samples, but cleaned up, with clear provenance and license info (generally a permissive creative commons licence but check the metadata for specifics).

The bin/meta.py python script is a reference implementation that can make a '.cleanmeta' metadata file for your own sample pack folder. See below for how to use it and contribute a sample pack of your own.

If you want to use these outside the Tidal/SuperDirt/SuperCollider ecosystem you are very welcome. You're encouraged to join discussion in the github issue tracker so that we can develop a standard way to share and index/signpost these packs.

See /tidalcycles/sounds-repetition for an example sample pack which has two sets of samples in it.

How to contribute a sample pack

Please only contribute samples if you are happy to share them under a permissive license such as CC0 or a similar creative commons license.

If you are unfamiliar with the 'git' software, please create an issue here, with a short description of your samples and a link to them and someone should be along to help shortly.

If you are familiar with git and running python scripts (or happy to learn), please follow the below instructions. This is all new - if anything is unclear please create an issue, thanks!

  1. Get your samples together in .wav format, editing them if necessary (see below for advice).

  2. Create a new repository. This isn't essential, but consider putting 'sounds-' in front of its name, e.g. 'sounds-303bass' for your 303 bass samples.

  3. Add your samples to the repository. For an example of how to organise them, see this sample pack: tidalcycles/sounds-repetition, which has two sets of samples, with a subfolder for each.

  4. Create a '.cleanmeta' metadata file for each subfolder. Again, see tidalcycles/sounds-repetition for examples. There is a python script bin/meta.py which can generate the metadata file for you, run it without parameters for help. Here is an example commandline, that was used to generate repetition.cleanmeta:

    ../Clean-Samples/bin/meta.py --maintainer alex --email [email protected] --copyright "(c) 2021 Alex McLean" --license CC0 --provenance "Various dodgy speech synths" --shortname repetition --sample-subfolder repetition/ --write .
    

    After generating the file, edit it with a text editor to fill in any missing info.

  5. When ready, add te URL of your repository to the https://github.com/tidalcycles/Clean-Samples/blob/main/Clean-Samples.quark for the Clean-Samples quark) in a pull request. You could also add it to the SuperCollider quarks database, or we can do that for you if you prefer, so that we can accept the PR to Clean-Samples once it's accepted as a quark.

Advice for preparing samples

You can use free/open source software like audacity for editing samples.

As a minimum, be sure to trim any silence from beginning/end of the samples, and that the start and end of the sample is at zero to avoid clicks (you might need to fade in / fade out by a tiny amount to achieve this).

Consider adjusting the volume/loudness too, for example normalising to -1.0db - but this is very subjective and will depend on the nature of the samples and the music they're used with. For example distorted gabba samples are intended to be very loud, and a whisper is intended to sound silent. The average non-percussive sample should be around -23dB RMS. Samples shouldn't exceed 0dB true peak. EBU recommends -1dBTP at 4x-oversampling. Samples generally shouldn't have DC offset, although e.g. some kick drum samples naturally have non-zero mean.

For more advice, you could join the discussion here.

Thanks!

Owner
TidalCycles
Live coding environment for making patterns
TidalCycles
A PyTorch implementation of Implicit Q-Learning

IQL-PyTorch This repository houses a minimal PyTorch implementation of Implicit Q-Learning (IQL), an offline reinforcement learning algorithm, along w

Garrett Thomas 30 Dec 12, 2022
Relaxed-machines - explorations in neuro-symbolic differentiable interpreters

Relaxed Machines Explorations in neuro-symbolic differentiable interpreters. Baby steps: inc_stop Libraries JAX Haiku Optax Resources Chapter 3 (∂4: A

Nada Amin 6 Feb 02, 2022
The pytorch implementation of the paper "text-guided neural image inpainting" at MM'2020

TDANet: Text-Guided Neural Image Inpainting, MM'2020 (Oral) MM | ArXiv This repository implements the paper "Text-Guided Neural Image Inpainting" by L

LisaiZhang 75 Dec 22, 2022
The Deep Learning with Julia book, using Flux.jl.

Deep Learning with Julia DL with Julia is a book about how to do various deep learning tasks using the Julia programming language and specifically the

Logan Kilpatrick 67 Dec 25, 2022
Implementation of "StrengthNet: Deep Learning-based Emotion Strength Assessment for Emotional Speech Synthesis"

StrengthNet Implementation of "StrengthNet: Deep Learning-based Emotion Strength Assessment for Emotional Speech Synthesis" https://arxiv.org/abs/2110

RuiLiu 65 Dec 20, 2022
Paddle pit - Rethinking Spatial Dimensions of Vision Transformers

基于Paddle实现PiT ——Rethinking Spatial Dimensions of Vision Transformers,arxiv 官方原版代

Hongtao Wen 4 Jan 15, 2022
Direct Multi-view Multi-person 3D Human Pose Estimation

Implementation of NeurIPS-2021 paper: Direct Multi-view Multi-person 3D Human Pose Estimation [paper] [video-YouTube, video-Bilibili] [slides] This is

Sea AI Lab 251 Dec 30, 2022
Alpha-IoU: A Family of Power Intersection over Union Losses for Bounding Box Regression

Alpha-IoU: A Family of Power Intersection over Union Losses for Bounding Box Regression YOLOv5 with alpha-IoU losses implemented in PyTorch. Example r

Jacobi(Jiabo He) 147 Dec 05, 2022
How to Become More Salient? Surfacing Representation Biases of the Saliency Prediction Model

How to Become More Salient? Surfacing Representation Biases of the Saliency Prediction Model

Bogdan Kulynych 49 Nov 05, 2022
Official repository for Automated Learning Rate Scheduler for Large-Batch Training (8th ICML Workshop on AutoML)

Automated Learning Rate Scheduler for Large-Batch Training The official repository for Automated Learning Rate Scheduler for Large-Batch Training (8th

Kakao Brain 35 Jan 04, 2023
Just Randoms Cats with python

Random-Cat Just Randoms Cats with python.

OriCode 2 Dec 21, 2021
Official implementation of the paper 'Details or Artifacts: A Locally Discriminative Learning Approach to Realistic Image Super-Resolution' in CVPR 2022

LDL Paper | Supplementary Material Details or Artifacts: A Locally Discriminative Learning Approach to Realistic Image Super-Resolution Jie Liang*, Hu

150 Dec 26, 2022
OntoProtein: Protein Pretraining With Ontology Embedding

OntoProtein This is the implement of the paper "OntoProtein: Protein Pretraining With Ontology Embedding". OntoProtein is an effective method that mak

ZJUNLP 80 Dec 14, 2022
Read number plates with https://platerecognizer.com/

HASS-plate-recognizer Read vehicle license plates with https://platerecognizer.com/ which offers free processing of 2500 images per month. You will ne

Robin 69 Dec 30, 2022
OBG-FCN - implementation of 'Object Boundary Guided Semantic Segmentation'

OBG-FCN This repository is to reproduce the implementation of 'Object Boundary Guided Semantic Segmentation' in http://arxiv.org/abs/1603.09742 Object

Jiu XU 3 Mar 11, 2019
Code for CVPR 2018 paper --- Texture Mapping for 3D Reconstruction with RGB-D Sensor

G2LTex This repository contains the implementation of "Texture Mapping for 3D Reconstruction with RGB-D Sensor (CVPR2018)" based on mvs-texturing. Due

Fu Yanping(付燕平) 129 Dec 30, 2022
OptNet: Differentiable Optimization as a Layer in Neural Networks

OptNet: Differentiable Optimization as a Layer in Neural Networks This repository is by Brandon Amos and J. Zico Kolter and contains the PyTorch sourc

CMU Locus Lab 428 Dec 24, 2022
Code for paper "Document-Level Argument Extraction by Conditional Generation". NAACL 21'

Argument Extraction by Generation Code for paper "Document-Level Argument Extraction by Conditional Generation". NAACL 21' Dependencies pytorch=1.6 tr

Zoey Li 87 Dec 26, 2022
[ICCV'21] NEAT: Neural Attention Fields for End-to-End Autonomous Driving

NEAT: Neural Attention Fields for End-to-End Autonomous Driving Paper | Supplementary | Video | Poster | Blog This repository is for the ICCV 2021 pap

254 Jan 02, 2023
The official implementation of CVPR 2021 Paper: Improving Weakly Supervised Visual Grounding by Contrastive Knowledge Distillation.

Improving Weakly Supervised Visual Grounding by Contrastive Knowledge Distillation This repository is the official implementation of CVPR 2021 paper:

9 Nov 14, 2022