Controlling a light show for a small solo set

I’m back on the track of my own small solo live set. The first experiment was running a video stream that would run along with the show. But now there is a new twist: The Corona virus came and there will be no live set the coming months. All public shows have been cancelled for about half a year. My first live show has been pushed to November from June. The only alternative is live streaming.

Just before the lockdown to combat the spread of the Corona virus I had bought a stage light. Just one to at least have a blue wash on stage to set a kind of moonlight mood. This was the Ayra ComPar 2. A simple LED stage light, with an IR remote and plenty of flexibility be more than just a blue stage wash.

But while staying at home and after browsing through some online articles it dawned on me: you can simply control stage lights as part of your Ableton Live set. I use Ableton Live sets to run my stage show and believe it or not I use color coding for each different song to quickly browse through all the songs without having to look up the names.

The colors match the moods of the song, so my simple idea was to use this color code to match the color of the wash on stage. A red wash for a deeply felt love song. A green wash for a song about nature. A purple wash for an up tempo hot song etc.

But why put all this effort in a stage light when there will not be a stage for months to play on? Up to then I had been a bit weary of immediately jumping to live streaming instead of playing gigs. All the bigger artists now stream live. Every night on my socials there are at least a dozen artists performing live. I’m just starting out, so what can I bring to the table?

After discussing this with a close group of musicians and my music coach it became obvious. Why not start streaming live? It’ll be fun, even if nobody watches it. I can invite friends and just have fun together. And also because I had nothing else to do I jumped in to make this stage light idea work. It would change color with the song. Not on stage, but in the attic. The attic with my home studio as my online stage.

One of the intriguing functions of the ComPar 2 is the ability to connect a XLR cable with DMX signal to control it. After diving into it and in lockdown there was a lot of time to dive into anything I found out that there are also DMX light controllers that support MIDI. From the same company I got the Ayra OSO 1612 DMX Scanmaster controller. Very friendly priced i think.

Blacked out by default
Blacked out by default

The DMX light controller simply accepts MIDI note data and maps that to programmable scenes. The controller can be connected to a chain of lights and a scene can set each light correspondingly. You can have flashing lights in a scene or movement from stage lights that can move. With 240 scenes you can probably make an interesting progression of lights for several songs, but I simply have a red, green, purple and blue scene for each song.

The controller I chose has a default setting where it blacks out all lights when starting up and that is not a bad thing at all. The only thing I must remember is to switch off the black out when playing live. That is the only attention it needs and from there everything is now running on rails. The live streaming shows allow me to test stuff out, but I’m now pretty happy with this setup.

Bad ground. When the noise is killing you…

Ok, maybe you don’t know this song, Bad Ground, from a controversial band Type O Negative. If you do know it it’ll bring a smile to your face. But if you hear it in your headphones or from your speakers, you won’t be smiling. The 50-60 Herz buzz or maybe even digital noise that ruins your listening pleasure and maybe even your recordings.

This is something I feel I need to discuss, because earlier I wrote about impedance. Like the previous article, you may be an experienced sound engineer or pro musician. In that case please skip this article. This is for the home studio creatives that just can’t keep the noise from creeping into the system. By the way you can find plenty of articles on the subject. This one just compounds it all into one.

In any case I was helping out building a home studio and when connecting a second display to the laptop there it was: bad ground. A digital fizz from the HDMI cable and a hum from the bad ground. The active speakers amplified the noise sounds coming from the signal in the unbalanced cables from the audio interface. Simple jack cables.

Ah you should say. There you have it. Unbalanced cables work by shielding a single signal wire with a mesh that wraps around it. This mesh wiring should be grounded. The shield mesh then prevents electromagnetic interference from the outside reaching the signal cable. This all goes wrong if not every device in the chain is solidly grounded along the same wire more than half a meter into the earth.

This bad ground can just be floating, catching interference from all noise sources around it. Or it can be that one part of the chain is grounded differently and the signal difference of the shielding interferes with the signal. Do you have the option to bring a solid ground in your studio setup? Please start there. In other cases there is only one other solution and that is to accept the bad ground and the interference of other signals.

Does that mean accepting the noise? No of course not. The answer to noise in the studio and on stage has always been the use of balanced cables. In balanced cables there is also the shielding mesh of a ground cable, but inside there are two twisted cables. A hot and a cold one. Usually using XLR connectors instead of the jacks. Although XLR cables can also be used unbalanced.

This time the difference between hot and cold is used as the signal. This time the effect of interference on the signal is much less, because it evens out on the twisted pair. This can only mean that you should always use balanced cables when possible.

Now if only it was that simple… Some audio interfaces only have unbalanced outputs. Instruments usually have only unbalanced outputs. what should you do with these signals to prevent bad ground? This is where you need DI’s. Direct Input devices, pick up the unbalanced signal from a jack connector and output a balanced signal to an XLR connector.

Inside the DI you will find a transformer that picks up the signal and passes it to the hot and cold wires of the XLR. But transformers are coils wrapped around the same core. the signal energy is passed from one coil to the other. This means signal loss and a coil does have a frequency response that is not 100% flat. In other words, the signal is lower in energy at the other side and can have slightly less low and high frequencies.

Passive DI’s just use the transformer and give you signal loss and a slight effect on the signal quality. Active DI’s can compensate for the signal and signal quality loss, but need a power source to make the additional electronics work. Even worse, cheaper active DI’s can add noise to the signal from the electronics. A more expensive passive DI can sound better than a cheap active DI.

In the end for my noise problem I found an affordable passive DI that sounds great. The studio setup was inexpensive and simple and there was therefore no noticeable degradation of the sound. You might also want to try this Millenium DI-E. DI’s can sound muffled, but the transformer can also add warmth to the signal. Some very expensive pre-amps for vocals and guitars use transformers to add warmth.

So there you have it. May you kill the noise…

Soundbrenner Metronome and Ableton Live Link

Since i started using the Soundbrenner Pulse and its Metronome app i’ve had serious problems connecting it to Ableton Live with Link. I read through the troubleshooting page forever. Added firewall rules everywhere. Checked the network traffic going from and to the laptop and the phone. Nothing. Almost nothing. The worst part was that somehow random suddenly the connection would work. Even more frustrating: i seemed to be the only one with these problems.

Suddenly it became obvious to me. If no one has these problems, it must be the network. Obviously the phone has to work from WiFi. My wireless network must be up to date and all should be fine, but it does work with these newfangled mesh repeaters. So my idea was: why not connect the laptop directly to the phone’s mobile hot spot and cut out the router and mesh network?

Soundbrenner Metronome App Live Link
Soundbrenner Metronome App Live Link

Suddenly everything connected flawlessly. If you ever want to use Ableton Live Link, make sure its a straight connection between devices. Any router or repeater can wreck the connection, or the reliability of the connection. Another problem finally solved.

When you need a patchbay

You might already have seen this on my socials. A nice photo of a new box stacked alongside my MIDI patchbay. Lately studio life got more complicated. I have 2 mixing tables. One for working in the studio and one for practicing live gigs. I found myself plugging instruments in and out of these mixing tables. Also, the studio mixing table, a Yamaha 01v, is getting old and some switches now already noticeably start making noise. For me this was the sign to start saving the desk and considering a patch panel.

You can spend any amount on a good one, but for my modest home studio purposes I chose the Behringer Ultrapatch Pro PX3000. With 48 channels it is well beyond my need to patch 6 channels across 12 inputs. But hey, who knows what will happen in the future. And it doesn’t break the bank at around 80 euros.

Plugging the instruments across the inputs of two tables now won’t wear down the inputs on the more expensive mixing desks any more. There is even be an option to use the patchbay in half-normal mode. In this mode I can make a setup to send the instruments to both inputs at the same time. Then you have to factor in the impedance of both mixing desks against the line outs of the instruments, but to my calculations it might just work.

Discovering Loopcloud 5.0 as a sample library manager

This maybe something that I had overlooked for too long: Loopcloud. For years the talk of the sampling library town, but I didn’t look at it until I got a demo of the new Loopmasters Loopcloud 5.0 version at the Amsterdam Dance Event this year. I also had looked at other sample managers like Algonaut Atlas, but that may be only drums oriented. Intriguing, because Atlas uses machine learning to recognize the types of samples. For me, up to now, a sample manager was simply a folder in Ableton Live to browse through. And I had always put Loopcloud away as simply a shop to buy samples with a subscription model.

How to work with the application

The Loopcloud application is a standalone application, but it integrates with your DAW through a Loopcloud plugin. You can only have it on one track in your DAW. All samples that you browse then play through that track. The idea is to start with a sample in the Loopcloud application. You can have random sorting to free your mind. Then use that to edit, slice, dice, sequence, mash up and add effects if you wish. You can drag the final result into your DAW as a sound file. Quite something different than finding a sample and then edit it in the DAW. All with the tune and tempo of your DAW. It nicely prevents you using kind of preset sounds over and over. Clever!

Loopcloud sample editor
Loopcloud sample editor

It means however, that you have to keep two applications open while working. For those of you with two monitors, maybe a no-brainer. But then again, it could just be that you already have a nice workflow with your two monitors and now you need to fit in yet another application. Anyway, there is an option to have the application dock to the sides of a window at about 20% of the width. Combined with scaling and other options, you might manage with one screen. The application sometimes forgets how you docked and scaled it.

Your library manager

Now about the library management. The moment you add your own samples to the Loopcloud application it starts scanning all the samples in it. It will try to find BPM and key information and it will try to read other information from the name of the sample or the loop. It will probably not correctly discover more complex information like the genre, loop or one-shot, or the exact instrument. All is then marked down as tags and you can start searching for things like key and BPM.

Loopcloud docked
Loopcloud docked

For this you need to click the button marked “Your Libary”. If you also want the detailed information of your scanned samples to be correct you will have to start tagging yourself. Its quite advanced, you can tag whole folders and batches of files. For a more in-depth dive into the tagging and searching you should dive into the tutorials.

Additional plugins!

But then when I found out Loopcloud as a sample manager, the tutorial also pointed me to Loopcloud Drum. A separate plugin that is actually a full sample drum instrument. It uses its own Loopcloud drumkit format and will open up a separate section in the Loopcloud manager. A strange find in a sample library manager. As a separate instrument it has its own format and its actually more of a pattern beatmaker with its own sequencer. A preset list of drum kits get activated that have been assembled from Loopcloud one shot samples of course.

Loopcloud Drum plugin
Loopcloud Drum plugin

I didn’t find any option to change the patterns in the beatmaker, other than with a mouse. You would also expect an option to edit drum kits and build your own. You can edit the mix of the kit and save that as a “user” drum kit, but I didn’t see any way to create a drum kit from your own set of one shots. Maybe this is in a future version, or in a Loopcloud subscription model that I didn’t explore. I was kind of on the lookout for tools to start making beats, other than with loops or Nerve, but this is not it yet.

Loopcloud Play plugin
Loopcloud Play plugin

And even more? The tutorial also points to the Loopcloud Play plugin. Yet another sample instrument, but this time melodic. As an instrument its quite basic, maybe so basic that you fall back into the preset trap again. There are about 7 knobs to turn and that’s it. Like the Drum instrument it has its own place in the library and again no way to choose the samples. You can save tweaks to the knob as “User” instruments. I think it needs work, as this is no match for Native Instrument’s Kontakt.

Closing out

Loopcloud has a quite intricate subscription model and not all of the features are available in all tiers. Specifically on using multiple tracks and the sample editing. However, if you just want to use it as a sample library manager you can even use the free subscription model tier. If you already own Loopmasters stuff it will automatically appear in your library. Even though it could do with more advanced detection of the samples that you load in the library, for me this was a great find and it surely beats the user folders in Ableton Live.

Its the impedance, stupid!

This is a short story about something that you take for granted in this high-tech age. That you can connect anything to anything and that it just works. This time I tripped over something that did not work and it reminded me harshly that there are classic electrical laws to take in to account: impedance matching. Even more embarrassing is that I am actually an electrical engineer that switched to computer science and music.

Zoom L-12 monitoring outputs
Zoom L-12 monitoring outputs

So these days I am working on my stage monitoring. Of course its at least my performers dream to have wireless in-ear monitoring, but then you will find that you have to invest at least hundreds of euros and you can easily go up to several thousands. This is why I started experimenting with a simple wired stereo in-ear monitoring system. The Zoom L-12 mixer/recorder that I am using has 4 mix outputs for monitoring so that is the starting point.

Lets try to set the impedance story straight without getting too technical. For that you can go to the wiki page about the subject. In short its about getting the energy from the output (a mixer) optimally to the input (headset, amplifier) of the connected device. Otherwise its kind of like fitting a wide garden hose to something that is too small. The electrical equivalent: the output impedance should be lower than the input impedance. As a rule of thumb you can expect for outputs:

  • 100 ohm to 600 ohm output impedance from line outputs
  • 0.1 ohm or less from an amplifier

And for input impedance:

  • 10K ohm input impedance or more for line inputs
  • An average of 32 ohms for headphones, but it can range from 8-600 ohm
  • Around 8 ohms for speakers

This only applies to unbalanced outputs and inputs. So that means jack plugs and speaker connections. The transformers used in balanced outputs and inputs will usually match without you having to worry about it.

Enough theory. It is always a good idea to start with the ‘zero’ option. Lets connect a simple Shure SE215 earphone to the L-12 monitoring output. It says ‘Phones’. Easy peasy. The sound comes out, but the lows are kind of missing. I just skipped over this this, because I just thought that this was the quality of the output from the L-12. Looking back this was not surprising. If you check the SE215 spec sheet you will find that with an average input impedance of 17 ohm this earphone is quite hard to drive!

A lot of energy is therefore lost, because the output impedance of the L-12 turns out to be 100 ohms. This output qualifies as a line output driver, expecting a high-impedance amplifier to pick up the signal. Actually connecting earphones to this connector is a bad idea! Listening however with a directly connected Sennheiser HD 280 Pro is a more pleasant experience. This is easily explained by its more friendly 64 ohm impedance. Energy is transferred not very efficiently (almost halved), but much more efficiently than with the Shure!

So then I first looked at the Behringer P2, a small active monitoring amplifier. It uses two AAA batteries. You can connect XLR or a stereo jack plug. Since the L-12 has stereo jack monitoring outputs, this seemed to be the way. When connecting it all and the SE215 the result was very disappointing. Like listening to overly compressed, pumping audio, with completely random frequency dips and a lot of noise. Another impedance mismatch?

I immediately blamed the Behringer P2. But when you scout for reviews, this device invariably comes out as top rated with a lot of very happy users. How is this possible? I still don’t know. Particularly vexing is that there is no specification of the input impedance of the P2. It must be that however. Because when I connect the balanced input to a balanced output, it all sounds fine. Possibly no-one uses the unbalanced jack of the P2.

This is why have fallen back to using the Thomann mini body pack 2. It allows me to use long cables and gives me volume control on the belt mounted device. The sound isn’t perfect, because the 100 ohm output still has to drive the SE215. I am still looking for that perfect wired monitoring solution. Any ideas?

A first attempt at an automatic VJ mix on stage

For some time now I am looking for a way to add video to my Ableton Live performance. In this article I am experimenting with VideoRemix Pro from MixVibes. There are many people with a similar quest so it seems and equally as many solutions. Most solutions (Resolume, Modul8) revolve around the Apple MacOS. Since I am not in the Apple ecosystem, these are not available to me. Some quite elaborate solutions use many components that all are glued together. Sometimes with MIDI, sometimes with plugins.

As a first attempt am looking for a simple single piece of software that can run inside Ableton Live for a PC. Enter VideoRemix Pro. You need to have the Pro version to run it inside Ableton Live as a plugin. When you look at the instruction video, you can see that it runs in Session mode. Which is how I use Ableton Live live. Looking at this it seems simple enough, but there is a learning curve.

This learning curve is not helped by obvious glitches and problems when using the software. I had quite a battle installing it and getting it to run as a plugin inside Live. The first problem was Live crashing when dropping the plugin on a MIDI track. Which is how you are supposed to use it. My first reaction was to ask for a refund, but after a reboot and some experimenting I got it to work. The secret for me was to make sure that VideoRemix does not use the Windows Default audio. Once I switched to the ASIO audio option that Live also uses, the plugin stopped crashing.

VideoRemix Pro runs in desktop mode as well as plugin mode, but not at the same time. The desktop mode seems solid enough, but even there I have run into glitches. This had to do mostly with customizing the Novation LaunchPad Mini that I wanted to use to control the video. The LaunchPad Mini had been just lying around as a backup for the Ableton Push that I mainly use. It is however not supported by default. The makers of the software prefer you using the full Launchpad Mk2, which has more control options of course.

This means that in order to use it, you have to define a custom control mapping for the software. This seems easy enough, since you have a MIDI learn mode in the software. It took some learning for me to use it. In short, hover over the element in VideoRemix you want to control. Then click or turn the midi knob to link it. Press it again to see if the mapping worked. After this you will see a custom mapping in the list of midi devices in the preferences, which you could then rename.

A new custom MIDI mapping in the VideoRemix Midi Preferences
A new custom MIDI mapping in the VideoRemix Midi Preferences

Then moving over to Ableton Live and running it as a plugin (remember: not at the same time), you will find this same list. Confusing enough there is a VST MIDI device there, but in my case that did not respond to any attempt to control the video. If you switch over to your custom mapping that you created in the desktop mode, things start moving. Now you can record your video sequence.

Creating or recording a video sequence is based on the 6×6 grid of buttons in VideoRemix. This means that you are limited to 36 clips that you can launch. One clip can run for 100 seconds. Hit a clip to start it. Hit it again to stop it. By default running clips is column oriented. You cannot start more clips running on the same column. One clip on the same column will stop a clip on another row. You can start an entire row with a single command. You can start an entire column, but only if you enable all clips playing in a grid of course.

If you want a more complex mix of clips with more than a few clips per song and more then a dozen of songs, you’re probably out of luck with 36 slots. It seems you have to simplify your VJ mix if you are using this software standalone. For now it will have to do for me.

The VideoRemix Clip Grid
The VideoRemix Clip Grid

The effects (FX) section is quite elaborate. You can control it as well as all the faders, through MIDI. The moment you hit full screen on the top right you will see your VJ mix full screen. Hopefully on the right video output, but I will have to look into that yet. The default set of clips also loops sound and this sound can be mixed, so you can also have sound effects playing as part of your performance.

This is my first attempt at working with video as part of a Live based performance. After quite a battle to get it working, it is now seems actually possible to have a video running as part of a Session mode sequence, like there is a real VJ at work. I am still quite worried about the overall stability of the setup and I need to get to grips with the quirks of the software.

If you have experience with this or other software setups, please comment below!

How about putting this card in your laptop?

After carrying around big and powerful laptops for years and tablets that were simply not powerful enough. I wanted to try a laptop in the ultrabook category. At that time the choice was light and powerful, but with compromises in working with graphics: the Lenovo Thinkpad T480s. These days you can buy ultrathin notebooks with additional graphics power, but that was then and this is now. Unfortunately similarly spec-ed Macbooks are out of my league.

One thing that I really checked when selecting this laptop was the support for Thunderbolt 3. This connection supports external graphics cards. Even though the onboard graphics on paper should be good enough for minimal VR support. However, I already had guessed that this would not run smoothly. Now after one year of use I finally got round to trying out the Akitio Node external housing with Thunderbolt support. Lo and behold, equipped with a ASUS GeForce GTX 1060 OC3 my laptop has now become a graphics powerhouse that easily runs VR, games and any other task smoothly as butter!

If you check the compatibility lists of Akitio, you will not find this graphics card. This list is very limited and only contains higher end cards. There is a small notice that it should work with any card, but there’s no guarantee. My idea was to try the slightly lower end, because maxing out anything like this in the end is bound to cause problems somewhere in the chain. The laptop is also a year old, so I reckoned that careful drivers go a long way. By the way, Macbooks also support Thunderbolt and external graphics cards. They don’t support the specific card I chose, because I believe you should use Radeon AMD graphics cards.

There was a struggle to get this working. When you go the Nvidia way with you graphics card, the driver installation can moan about the hardware not qualifying for installation. I got round this by checking forums and these explained that you should manually install the driver. Unpacking the software and finding and installing the driver files by hand, right clicking on .inf files did the trick.

The Akitio Node is large and slightly noisy, but you can switch it off for music work of course. Then when you want to do graphics intensive work, switch the Node on and voilà. Magic in a box! One other down side is that it does not support any other function then connecting the graphics card. There are other options with storage and other connections, like USB or network. There also is no daisy chaining of devices. The Node is a dead end.

If you too are looking for this upgrade I wanted to put this story out, because this is a working combination. There are a lot of horror stories around of combinations refusing to work. I hope this upgrade trick will work for you too!

Make your own affordable dust covers

I am a big fan of custom made covers for all studio equipment. Dust kills the quality of connectors, sliders and switches. If possible I try to use the dust covers from the instrument or equipment manufacturer. Otherwise I try to look for a Decksaver, because actually these are very clever desk space savers as well. And they fit like a glove. All too often however I find that there is no custom cover that exactly fits.

I tried to find custom covers for the Yamaha 01v and the MicroKorg, but couldn’t find any. My current solution for this is to buy flexible transparent foil and have it cut to a little more than the surface area of the device. It can attract dust, but at least the dust does not get in the equipment. It also looks quite professional and is easy to pull over the surface and slip away again.

For me this beats ill fitting cloth covers and other half baked solutions, like putting it in a box. The best alternative could be to buy thick sheets of perspex and glue a custom cover. There are shops that support you in building your own perspex cover. It will however never be as sophisticated as Decksaver covers, with extra space for knobs and bends in the device. If you’re on a budget, at least put a sheet of transparent foil over your equipment and make it last longer.

Why you should start using a 360 camera

Already four years ago I started using a 360 camera. At that time I wanted to create those videoclips where you are really in the set and I wanted viewers to experience the video. The video quality was then an issue and for me it still is, unless you have a solid budget to spend. At the 3.000 euro price point video quality is no longer a big issue. At the lower end however, things have improved slightly. I have now invested in an Insta360 ONE X at a fraction of that price, 400 euro. What has persuaded me to invest in this camera if the quality is only slightly better?

First off, it comes with software that allows you to take your full 360 degree recording and cut out a flat rectangle that looks like you recorded it with a normal camera. Where is the advantage in that? It is actually intended to allow you use it as an action camera and then in the video editing cut out, pan and zoom into any action around you. You can see samples of this in the product page. What use is that to me as a musician, you might ask. Well, how about filming a whole gig from several points and cutting, panning and zooming into all the action on stage and in the crowd? Also the software has some really captivating special effects like speeding up, turning the 360 view into a ball, fish eye etc.

Secondly, it has rock-solid stabilization, because it uses gyroscopes to record all movements. This also ensures that the recording is perfectly horizontal, even when recording at an angle. You will find that if there is too much movement in your recording, most viewers will become sea sick really fast. A smooth recording and stable recording makes the difference. I can now confidently record while walking. Also freaky is that if you use a selfie stick to hold the camera, the software will remove the stick. It will appear as if the camera is hovering above you.

Schemerstad
Schemerstad

Thirdly, it actually matters that the quality of the recordings is at least slightly better than that of the first generation of 360 cameras. The performance in low light is dramatically better and the 25% increase in pixels of camera’s in the same price range does make a difference. Am I completely happy? No of course not. I can really and wholeheartedly recommend the ONE X at the lower price tier. It has made some impossible recordings possible and I will keep using 360 as part of my video recording to capture the action and experiences.

So this is why you too (as a musician) should start using a 360 camera. Not because you want people to experience VR, but to capture everything and decide how to use the recording later. On stage and everywhere the action happens.