If you have seen my recent live streams, you will have noticed that I ‘travel around’ these days while live streaming. I’ve started to use the Green Screen effect. With OBS Studio its so dead simple that you can start using it with a few clicks in your OBS Studio scenes. Of course there are also some caveats I want to address. The main picture for this post shows you what it can look like. It may not be super realistic, but it is eye catching.
So what do you need to get this going? A Green Screen is the first item you need. It does not have to be green. It can be blue or blue-green, but it should not match skin color or something you wear. It should cover most of the background, so it will need to be at least 2 meter by 1.6 meter, which is kind of a standard size you can find in shops. It should be smooth and solid. Creases and folds can result in folds in the backdrop, but some rippling is OK.
Then you need to set up OBS Studio. Its as simple as right-clicking your camera in the scene and selecting the Filters properties. In the dialog add the Chroma Key filter and select the color of your green screen. Then slide Similarity from somewhere around 100-250 to get a good picture. Everything outside the color range will become black. Then add a backdrop image (or video!) somewhere below the camera in the the scene list and you will have your Green Screen effect.
The first caveat I bumped into was that I set it up during daytime and it kind of worked, but then I found I stream in at night time and then you need light. In fact it turned out that 2 photo studio lights came in handy. When you use at least 2 studio lights they also cancel out shadows through folds and creases in the green screen. It does however bleed a little onto you as a subject, so you will be strangely highlighted as well. This is something you can also see in my first Amsterdam subway picture. Because of the uneven lighting in subways it does not really show. Not every picture is suitable as a backdrop. Photos with people or animals don’t work, because you expect them to move.
The second effect you see is that instruments with reflective surfaces also reflect the green screen. This will result in the background shining through reflecting surfaces. My take is that its a minor distraction, so I accept some shining through of the backdrop. Its also possible that some parts of your room don’t fit well with the Green Screen, doorways or cupboards. In that case you can choose to crop the camera in the scene by dragging the sides of the camera in the scene with the Alt-key (or Apple key) down. The cropped camera borders, will be replaced by the backdrop.
In a previous post I mentioned that I use OBS Studio for my live streaming and a little bit about how. It shows that I use an ASIO plugin for audio in the OBS Studio post, but why is it needed? For me in the live stream I want to recreate the studio quality sound, but with a live touch. After all, why listen to a live stream when could just as well listen to the album or single in your favorite streaming app? Lets first see where the ASIO plugin comes into play.
For OBS Studio and the live streaming setup, I chose to use PC on the studio recording side. Its directly connected to the Internet (cabled) and can easily handle streaming when it doesn’t have to run studio work. I play the live stream on the set dedicated to playing live and i use the live side stereo PA audio out to connect it to the studio side to do the live streaming. This means the live side if the setup is exactly as I would use it live.
It all starts with the stereo output on the Zoom L12 mixing desk, that normally connects to the PA. On the mixing desk there is vocal processing and some compression on all channels to make it sound good in live situations. To get this into the live stream as audio I connect the stereo output to an input of the Yamaha mixing desk. This is then routed to a special channel in the studio side audio interface. This channel is never used in studio work.
Of course it could be that your live setup simpler then mine. Maybe only a guitar and and a microphone. But the essential part for me is this that you probably have some way to get these audio outputs to a (stereo) PA. If you don’t have a mixing panel yourself and you usually plug in to the mixing desk at the venue, this is the time to consider your own live mixing desk for streaming live. With vocal effects and the effects that you want to have on your instruments. Maybe even some compression to get more power out of the audio and make it sound more live.
But lets look at where the ASIO plugin comes into play. The ASIO plugin takes the input of the special live channel from the Yamaha mixing desk using the studio side audio interface and that becomes the audio of the stream. Because I have full control over the vocal effects on the live side, i can just use a dry mic to address the stream chat and announce songs. Then switch on delay and reverb when singing. Just like when I play live, without the need for a technician even.
Playing a live stream is different from playing live, because it has a different dynamic. In a live stream its OK to babble and chat minutes on end, this is probably not a good idea live. I find however when it comes to the audio, it helps to start out with a PA ready output signal. Similar to the audio you would send to the PA in a real live show. Also it helps to have full hands on control over your live audio mix to prevent you having to dive into hairy OBS controls while streaming live. Lastly, for me its also important that streaming live is no different from a playing live at a venue in that you can break the mix, miss notes, mix up lyrics and that you feel the same nerves while playing.
Okay, like everybody else i started streaming too. I had a planned live show, but live shows will not be possible for at least another half year. Every evening my social timelines start buzzing with live streams and all the big artists have also started to stream live. No place for me with my newly created and sometimes shaky solo live performance to make a stand? After some discussions with friends i decided to make make the jump.
But how to go about it? If you already have experience with live streaming, you can skip this entire article. This is here just for the record so to say. After some looking around I came to this setup:
OBS is surprisingly simple to set up. It has its quirks. Sometimes it does not recognize the camera, but some fiddling with settings does the trick. You define a scene by adding video and audio sources. Every time you switch from scene to scene it adds a nice cross fade to make the transition smooth. You can switch the cross fade feature off of course.
I only use one scene. The video clip is there to promote any YouTube video clip. It plays in a corner and disappears when it has played out. The logo is just “b2fab” somewhere in a corner. The HD cam is the C920 and the ASIO source is routed from my live mixer to the audio interface on the PC. I setup a limiter at -6db on the ASIO audio as a filter to make sure i don’t get distortion over any audio peaks.
I also had to choose my platform. From the start i wanted also to stream live on Facebook and Instagram. Instagram however kind of limits access to live streaming to only phones. There is software to stream from a PC, but then you have to set it up again for every session and you need to switch off two-factor authentication. For me one bridge too far for now.
I chose Restream.io as a single platform to set up for streaming from OBS. It then allows to stream to multiple platforms and bundle all the chats from the different platforms into a single session. For Facebook pages however, you need a paid subscription tier. For now I selected the free options YouTube, Twitch and Periscope. YouTube because it is easy to access for my older audience. Twitch because it seemed quite fun and i also like gaming. Periscope because it connects to Twitter.
If the live show takes shape i might step into streaming from my Facebook page. Another plan is to try the iRig Stream solution and start making separate live streams on Instagram. With high quality audio from the live mixer. I will surely blog about it if i start working with it.
For now it all works. Restream.io allows me to drop a widget on my site. Its a bit basic and only comes alive when i am live, so i have to add relevant information to it to make it interesting. If you want to drop in and join my live musings check my YouTube, Twitch and Periscope channels or my site at around 21:00 CEST.
For some time now I am looking for a way to add video to my Ableton Live performance. In this article I am experimenting with VideoRemix Pro from MixVibes. There are many people with a similar quest so it seems and equally as many solutions. Most solutions (Resolume, Modul8) revolve around the Apple MacOS. Since I am not in the Apple ecosystem, these are not available to me. Some quite elaborate solutions use many components that all are glued together. Sometimes with MIDI, sometimes with plugins.
As a first attempt am looking for a simple single piece of software that can run inside Ableton Live for a PC. Enter VideoRemix Pro. You need to have the Pro version to run it inside Ableton Live as a plugin. When you look at the instruction video, you can see that it runs in Session mode. Which is how I use Ableton Live live. Looking at this it seems simple enough, but there is a learning curve.
This learning curve is not helped by obvious glitches and problems when using the software. I had quite a battle installing it and getting it to run as a plugin inside Live. The first problem was Live crashing when dropping the plugin on a MIDI track. Which is how you are supposed to use it. My first reaction was to ask for a refund, but after a reboot and some experimenting I got it to work. The secret for me was to make sure that VideoRemix does not use the Windows Default audio. Once I switched to the ASIO audio option that Live also uses, the plugin stopped crashing.
VideoRemix Pro runs in desktop mode as well as plugin mode, but not at the same time. The desktop mode seems solid enough, but even there I have run into glitches. This had to do mostly with customizing the Novation LaunchPad Mini that I wanted to use to control the video. The LaunchPad Mini had been just lying around as a backup for the Ableton Push that I mainly use. It is however not supported by default. The makers of the software prefer you using the full Launchpad Mk2, which has more control options of course.
This means that in order to use it, you have to define a custom control mapping for the software. This seems easy enough, since you have a MIDI learn mode in the software. It took some learning for me to use it. In short, hover over the element in VideoRemix you want to control. Then click or turn the midi knob to link it. Press it again to see if the mapping worked. After this you will see a custom mapping in the list of midi devices in the preferences, which you could then rename.
Then moving over to Ableton Live and running it as a plugin (remember: not at the same time), you will find this same list. Confusing enough there is a VST MIDI device there, but in my case that did not respond to any attempt to control the video. If you switch over to your custom mapping that you created in the desktop mode, things start moving. Now you can record your video sequence.
Creating or recording a video sequence is based on the 6×6 grid of buttons in VideoRemix. This means that you are limited to 36 clips that you can launch. One clip can run for 100 seconds. Hit a clip to start it. Hit it again to stop it. By default running clips is column oriented. You cannot start more clips running on the same column. One clip on the same column will stop a clip on another row. You can start an entire row with a single command. You can start an entire column, but only if you enable all clips playing in a grid of course.
If you want a more complex mix of clips with more than a few clips per song and more then a dozen of songs, you’re probably out of luck with 36 slots. It seems you have to simplify your VJ mix if you are using this software standalone. For now it will have to do for me.
The effects (FX) section is quite elaborate. You can control it as well as all the faders, through MIDI. The moment you hit full screen on the top right you will see your VJ mix full screen. Hopefully on the right video output, but I will have to look into that yet. The default set of clips also loops sound and this sound can be mixed, so you can also have sound effects playing as part of your performance.
This is my first attempt at working with video as part of a Live based performance. After quite a battle to get it working, it is now seems actually possible to have a video running as part of a Session mode sequence, like there is a real VJ at work. I am still quite worried about the overall stability of the setup and I need to get to grips with the quirks of the software.
If you have experience with this or other software setups, please comment below!
Already four years ago I started using a 360 camera. At that time I wanted to create those videoclips where you are really in the set and I wanted viewers to experience the video. The video quality was then an issue and for me it still is, unless you have a solid budget to spend. At the 3.000 euro price point video quality is no longer a big issue. At the lower end however, things have improved slightly. I have now invested in an Insta360 ONE X at a fraction of that price, 400 euro. What has persuaded me to invest in this camera if the quality is only slightly better?
First off, it comes with software that allows you to take your full 360 degree recording and cut out a flat rectangle that looks like you recorded it with a normal camera. Where is the advantage in that? It is actually intended to allow you use it as an action camera and then in the video editing cut out, pan and zoom into any action around you. You can see samples of this in the product page. What use is that to me as a musician, you might ask. Well, how about filming a whole gig from several points and cutting, panning and zooming into all the action on stage and in the crowd? Also the software has some really captivating special effects like speeding up, turning the 360 view into a ball, fish eye etc.
Secondly, it has rock-solid stabilization, because it uses gyroscopes to record all movements. This also ensures that the recording is perfectly horizontal, even when recording at an angle. You will find that if there is too much movement in your recording, most viewers will become sea sick really fast. A smooth recording and stable recording makes the difference. I can now confidently record while walking. Also freaky is that if you use a selfie stick to hold the camera, the software will remove the stick. It will appear as if the camera is hovering above you.
Thirdly, it actually matters that the quality of the recordings is at least slightly better than that of the first generation of 360 cameras. The performance in low light is dramatically better and the 25% increase in pixels of camera’s in the same price range does make a difference. Am I completely happy? No of course not. I can really and wholeheartedly recommend the ONE X at the lower price tier. It has made some impossible recordings possible and I will keep using 360 as part of my video recording to capture the action and experiences.
So this is why you too (as a musician) should start using a 360 camera. Not because you want people to experience VR, but to capture everything and decide how to use the recording later. On stage and everywhere the action happens.
Instagram took everyone by surprise by introducing the new upright video format IGTV video channel for all users. Shooting a video was obviously a horizontally oriented wide screen experience, matching the orientation of TVs and cinema. Instagram stories however were always vertically oriented to match the way you naturally hold your phone. IGTV nicely cultivates that. Some people always record vertically and that footage is then hard to show on TV, YouTube and such. Now you have a new outlet for that, enter IGTV.
If you have your material for your music video already recorded in upright position then you are so ready to edit it for IGTV! What I can see however is that not many existing recordings were ready for IGTV, so many decided to just clip off some footage from the left and right to keep the middle bit. The worst ones cut off parts of the titling so you can clearly see that its not the right IGTV stuff. As a viewer you feel cheated, because obviously you’re missing parts of the video.
But what if you have already recorded a video clip to be shown on YouTube and its in the landscape format? How to reuse that recording to make something that looks right on IGTV? What are the technicalities of the new IGTV video format?
The first step for me with the landscape clip for the Just a Game video, was to render it without the titling. All titling that does not fit the vertical format. The format to go for is HD, but then with reversed horizontal and vertical resolutions. So 1080×1920. With lengthy music video clips, you will find that upright portrait HD results in files that are too big in size. There is a size limit for regular video uploads, a maximum 10 minutes length and 650MB. The error messages from IGTV are not at all revealing unfortunately. A clip of 4 minutes length or more however, can easily go over 650MB. Then you will have to consider HDReady 720×1280.
If you removed the titling because of the landscape format, now is the time to redo the titling for the vertical format to show the viewer that you have intended this clip to be in IGTV format. After that, all you will have to do is to use pan and zoom to cut out the upright sections of the clip that really show the user all the action in the clip. This way you don’t have to give away that the clip recording was not intended for IGTV. As always I am using Corel VideoStudio for the simple work and its capable of rendering the required output for IGTV. Now its time to upload! Tell me about your experiences!
Telling stories through songs. When there is no story I have no drive. Sometimes I actively scout for interesting stories. A sequence of words. A thought. Serendipity. Usually a picture forms in my head that shapes the story and that is when I start thinking about a clip. Last year this was very strong when Sam came to work on Memories. Its actually one of the first songs that I posted on SoundCloud, but Sam shaped it, rewrote the lyrics and sang it in the attic.
The original song was inspired by a school reunion. A magic event that reshaped my past, because most of what I remembered actually was wrong or incomplete. It seemed apt to make a video clip for it that somehow would take you back to school and down memory lane. But, how shape this image?
Usually I try to cast the original singer and maybe let him or her sing at least a few lines in the clip. I got into contact with the head of my old school, not far from Leiden, in Sassenheim. He arranged the possibility to shoot a video in the school. First everything seemed fine, but then came the terrible news that Sam could no longer take part in the project. No conflict, but personal problems. Disaster. Sam recovered, but still could not take part in the project. However, she agreed that I could use a stand in. She took part in the final result from the sideline.
A clip needs a story!
The first thing I try to do is write down the story of the clip, but in this form. A form that works for me:
The premise: the initial state of affairs that drives the plot. In this case: Coming back to your old school to look around and remember all the fun and nice memories of your school years after a school reunion.
Scenes: all the images I have in my head. The locations. The video shots. How to shoot the video, moving the camera or on a tripod.
All necessary props and assistance needed for every scene
This way I can answer all questions that people involved usually have. How long do you need to shoot in the school? Which locations? What to wear? Earlier I would just take my camera and go, but I found that when you start filming that its nice to come prepared. Also, it really helped a lot that I already knew the school and made an extra visit to fill in some details.
In this case Ingrid saved the whole project. She agreed to be the stand in for this clip and she is such a graceful and beautiful lady. Throughout shooting she remained strong and focussed. Even when we somehow seemed to be locked up in school by the cleaning ladies.
We have stuff…
A short word on equipment. I have invested in a Panasonic system camera that is capable of shooting 4K. It has image stabilization, but when shooting a shot while moving I now insist on using a gimball. When movement is not smooth its just not right, or it has to be a special effect. For shooting while moving I use the DJI Osmo that has a 4K camera attached. Earlier I found out that insufficient lighting can ruin recordings, so I also invested in special 1000W video lights. I always try to reuse the recordings in the form of making-ofs. For this shoot I also brought the VR 360 camera and an action camera to shoot the recording setups while recording. Makes sense? I hope so.
Why 4K? I have found that when the starting material is 4K, a resulting HD movie is of higher quality than a movie shot with HD equipment. Its the pixel interpolation that somehow results in a sharper image. Also, if the end result is a 4K edited movie, it is also ready for the foreseeable future of video. And if the end result somehow has imperfections, it is possible to cut out the best bits while keeping a HD quality result. Also 4K has overall better quality. No more jaggies much less moire. The only thing that I have saved on for now is framerate. My budget unfortunately does not allow me to shoot 4K at more than 25 frames a second. For now it will have to do.
…but how to use it
What I try to do is to set up all equipment manually. This is extra hassle and extra risky, but what I try to avoid is to have all equipment think for itself. These camera’s all have auto-everything settings. These settings make smart decisions to make sure the subject(s) that you film are well lighted and in focus. However, when filming I find that a camera may suddenly shift to new settings for the best lighting and focus in the middle of a recording. This will make it more difficult in the final video production to choose the correct color tones and lighting. Shifting focus in the middle of a recording will probably make the recording useless. So everything is set to manual, focus and lighting will be fixed.
In the school setting the use of my video lights with a fixed color tone makes sure that the prevailing color tone of the recordings is also fixed. For the final video production I chose to give the whole clip a warm summer tone. Even when the clip was shot on a cold and cloudy autumn afternoon.
And there is more
For this clip I also made sure that Eline from Beauty & Visagie could be there. While recording I am already juggling all the equipment and video shooting. Eline made sure that Ingrid looked her best and remained stunningly beautiful throughout the whole afternoon. She also checked clothing and made sure that colors did not clash and that contrasts were in check. Walter also assisted, supported Ingrid, guarded stuff while we were running around and made sure we had power.
During recording I continuously play the music from the clip to make sure that the ‘rhythm’ of all movements matches. It also helps making sure that the mood is correct. Even though the final production might not be very pro. This way at least it will have the right flow and mood to match the music.
Also I always make sure that everyone in the clip is aware that it will be put on YouTube and made public. For these occasions I always carry paper release forms that people can sign to agree to ‘be in the movie’. For the people directly involved I just ask, but for strangers it might be best to let them sign a form.
As always: this is just the way I try to work and this is a reminder to self. Maybe it helps you.
If you want to support me making higher quality VR 360 video clips, please click here:
In the previous post showed you how i currently currently record VR 360 video. The Gear 360 does not output large video files (typically 100 MB). On the whole these files can be processed on any laptop or PC easily. The bit rate is not as extreme as the 4K output of professional camera’s. The GB files from these camera’s can bring a lower spec PC to its knees immediately. Expect this also with balls of GoPro camera’s. You shouldn’t have problems with Gear 360 video.
You can process these files with any video editing software, as long as you only use cross fades or other basic transitions. Slow motion or any speed effects will be ok. Even some special effects will apply, like vignette effects. These will blur or darken parts of the 360 video and that can work out quite ok. Coloring effects are also fine of course. One of the effects that does not apply is anti shake, because that will snip the edges of the video and thereby breaks the stitching of full 360 video. Actually technically anti shake could be done by rolling in and out parts of the video on opposing sides. I didn’t find any effect that can do that yet.
Now, titling and logos. If you want any titling you will either have to accept that it will be curved in strange ways, or you will need software that can apply the necessary curving to mix in the titling at the right viewing distance and angle. You can try to do this using Hugin. You can find instructions on how to create images that can be blended into your VR 360 video for logo’s and titling.
So now you’ve got mixing of different shots covered and titling and thus your basic needs. But is this enough for you? The remaining problem is that you can only see the final result after finishing editing, rendering and outputting to your VR device. This makes editing a chore. Fortunately, VR 360 is catching on and there is software that allows you to edit in a real 360 way. Even on a classic flat PC monitor.
Enter Pinnacle Systems Studio 20 Ultimate. The first affordable editing software for VR 360 movies. Now you get a 360 preview window and a way to place 2D content, like logo’s and titling, in the 360 space and preview the result immediately. This also means that you can mix in flat 2D video as part of your 360 video. You still need to be aware which effects and transitions apply in the VR world, but at least you can see the results without first rendering it and move it to the viewer. I am quite sure that more video editors will support VR 360 video. For now your starting point can be Studio 20 Ultimate, or just keep it basic and simple. The end result is worth the effort!