Applause is a LuaJIT-based real-time audio programming environment based on a stream algebra. Think of it as APL on lazily evaluated streams of audio samples. In "Applause" there is no distinction between sample and control rate - every stream may provide control data at the sample rate. Also, there is no distinction between programming language and audio synthesis engine - all calculations are performed by JIT-compiled Lua code, which greatly simplifies the architecture and possibilities to extend the system by "end users". On the downside, it requires buffering between the real-time audio thread and the synthesis code which manifests in additional latency. "Applause" currently supports the following features:
- well known operations from functional and vector-based programming languages
- various oscillators for sinusoidal, sawtooth, square and triangular wave forms
- hull curve stream generators (for instance ADSR curves for instruments)
- white, pink and brown noise generators
- plotting audio samples via ASCII art and Gnuplot
- reading and writing audio files - also in "as fast as possible" (non-realtime) mode
- infinite impulse response filters (LPF, HPF, BPF, BRF)
- Fast Fourier Transform (FFT and IFFT) - also in real time (STFT)
- external audio plugins via DSSI/LADSPA
- MIDI and HID device support to generate control signals
- CLI, scripting/batch mode
- simple integration into text editors and IDEs
- Jupyter notebook support
- Linux and FreeBSD are supported and the JACK audio daemon is strictly required
See also the TODO file for a list of bugs, possible features and improvements.
The easiest way to install Applause on Linux is to install a prebuilt AppImage from a nightly build. It should run on any x86_64 Linux system that has the JACK daemon (jackd2/jackdmp) installed and running. The AppImage supports all three modes of running Applause:
./Applause-nightly-glibc2.29-x86_64.AppImage
by default launches a Jupyter notebook on HTTP port 8888. Additional parameters are passed to jupyter. Use theAPPLAUSE_OPTS
environment variable to pass commandline parameters to Applause itself../Applause-nightly-glibc2.29-x86_64.AppImage ilua
launches a Jupyter/ILua console in the terminal. Additional parameters are passed to ILua. Use theAPPLAUSE_OPTS
environment variable to pass commandline parameters to Applause itself../Applause-nightly-glibc2.29-x86_64.AppImage cli
launches a plain Applause shell (Lua prompt). Additional parameters are directly passed to Applause, but theAPPLAUSE_OPTS
environment variable can also provide parameters. This also mode also allows executing scripts, but currently you will have to pass absolute paths.
You are recommended to manually build and install LuaJIT v2.1 since distributions usually ship outdated versions:
git clone -b v2.1 https://luajit.org/git/luajit.git
cd luajit
make
sudo make install
Furthermore, install the following dependencies (Ubuntu):
sudo apt-get install build-essential libreadline-dev libjack-jackd2-dev \
libsndfile1-dev libasound2-dev feedgnuplot
On FreeBSD, you will need the following packages/ports:
pkg install gmake readline jackit evdev-proto libsndfile alsa-lib p5-feedgnuplot
To compile the project, type:
make
Up-to-date documentation is available at the website.
In case you want to build it manually, install LDoc
(for instance luarocks install ldoc
) and type:
make doc
The generated documentation will be generated in the doc/
subdirectory.
Start JACK daemon (for instance via qjackctl).
./applause -o 2
Example (one channel):
> Stream.SinOsc(440):play()
You can also run standalone scripts (batch mode), just like the standard Lua interpreter.
In order to run Jack and Applause with real-time scheduling, it should be sufficient to
add your user to the audio
group.
To give regular users access to HID devices, it should suffice to add the current user to
the input
group.
For realtime scheduling, you might have to check out the mac_priority kernel module
and add your user to the realtime
group.
Furthermore, to allow unlimited memory locking on FreeBSD for ordinary users,
you should add the following entry to /etc/login.conf
:
audio:\
:memorylocked=unlimited:\
:tc=default:
Change the login class of your user to audio
by running chpass
.
You might need to add the current user to the wheel
group and
give read acceess to evdev device nodes by creating /etc/devd.rules
:
[localrules=10]
add path 'input/*' mode 0640
echo -ne "25 \nStream.SinOsc(440):play()" | socat -,ignoreeof TCP:127.0.0.1:10000
See also client.tes for a SciTECO integration.
This is supported by EvdevStream().
Alternatively you can use aseqjoy together with a2jmidid --export-hw
to expose them as MIDI events.
This is supported by EvdevStream().
Alternatively you can use raton together with a2jmidid --export-hw
to expose them as MIDI events.
- jack_rec, jack-capture, QJackRcd or Audacity to record sessions
- jack_midi_dump or midisnoop for diplaying MIDI events
- jack-keyboard for producing MIDI note events
- midicontroller for producing MIDI CC events
- MIDI Tracker ???
- rtspeccy for a realtime spectrogram
- evtest to find and test HID devices
- listplugins and analyseplugin to inspect LADSPA plugins
- dssi_list_plugins and dssi_analyse_plugin to inspect DSSI plugins
Applause can be run in Jupyter Consoles and even Notebooks thanks to ILua. For full support of all feautures, you must currently use an unofficial ILua fork. First, install ILua into a Python environment (see also this ILua ticket):
python3 -m venv env
. env/bin/activate
pip install twisted==22.10.0 git+https://github.com/rhaberkorn/ilua.git@improvements
You can now directly run an Applause Jupyter Console session:
ilua --lua-interpreter=./applause
In order to tweak Applause command line parameters and be independant of the execution directory, use the included wrapper script. It also allows passing in additional arguments to Applause, e.g.:
APPLAUSE_OPTS="-o 2" ilua --lua-interpreter=./ilua-wrapper.sh
You can symlink this to lua
in the Python environment to make Applause the default
ILua interpreter in this Python environment:
ln -s $(pwd)/ilua-wrapper.sh env/bin/lua
If you would like to launch a Jupyter Notebook (Web UI!), first install the following Pip package:
pip install notebook
Now launch a web server and follow the onscreen instructions:
APPLAUSE_OPTS="-o 2" jupyter notebook --MultiKernelManager.default_kernel_name=lua
This works assuming that you symlinked ilua-wrapper.sh
to lua
as described above.
An alternative might be to create a custom Jupyter kernel configuration (kernel.json).
If the browser is not opened automatically on the notebook's URL, you might want to try visiting http://localhost:8888/.
Please note the following restrictions/bugs:
- You cannot publicly host the Jupyter Notebook as the sound is generated on the host machine.
- The output of some functions like Stream:toplot() is garbled.