During the mixing step, you will balance instruments, place them in the sound space, set their volumes, modify their frequencies, fix or hide the flaws. The purpose is to define each instrument's location, make them separated from one another and create a song that is pleasant to listen to.
Dynamics is the interval between the lowest and the highest volumes in a sound. OK, fine, but we are doing Rock music, who cares about low volume? We want to kick ass! Drums must beat, guitars must yell, the bass must make the walls shake and the vocals must be loud! Pump up the volume! Wooo... calm down!
If everything is loud, then nothing is
Yes, that's logical. If everything is loud all the time, nothing stands out. But it's the contrasts that bring both subtleties and raging moments to light. How do you expect to surprise your listeners with a sound hurricane if you've already given them thunder and lightning without a break? You have to spare some quiet moments before the storm. Let's compare this with drawing. A black dot on a white background will stand out more easily than the very same black dot on a dark-gray background.
Obvious? Yes, but many musicians just care for that "powerful sound" to such an extent that it eventually becomes tiring to listen to.
You can increase the contrasts through the very writing of the song (guitar/vocal bars, followed by a passage with 3 guitars, 1 distorted bass, big drums...), or through interpretation (the musicians play louder, faster...), or through the use of effects (saturation, compression...).
Contrasts don't have to be huge, it all depends on what kind of emotion you wish to convey, on the music genre, on the topic of the song, on what you are looking for...
Nowadays, too many songs just look (and sound) like this:
Do your ears a favor: don't do this
EQ stands for equalization. It's a tool that allows you to set the level of the frequencies that make the sound. Basically, it's like a volume knob, but instead of changing the volume of a whole song, you lower or raise only the frequencies of your choice. Just like when you use a stereo amplifier for instance: you can find "bass", "medium" and "treble" knobs. They are a simple EQ that help you to tune the sound.
Well, you are going to apply this same principle to each of the instruments of your song.
The purpose of equalization is simple: changing the sound, so that each instrument primarily occupies the frequency range that defines it the best, without encroaching on the other instruments.
In other words, lower the level of the frequencies that are not necessary to an instrument without distorting it, and see to it that the various instruments main frequencies do not overlap. Each instrument will thus be alone in its own sound space without encroaching on the others, ensuring all instruments are clearly distinct from the rest.
The hard part is to do this without distorting the sound of the instruments, and also to see to it that instruments which have common frequencies by nature be well-marked.
You will deduce that the more instruments there are, the more complex mixing becomes.
Music is usually mixed in stereo. Not always, but we have two ears, we use headsets or speakers that usually come in pairs, so it only makes sense to use left and right channels when mixing.
That's what panning is: acting on the width of a sound, from left to right.
On top of the EQing described above, we can also part instruments from one another by placing them in different spaces. If two guitars play simultaneously in your song, you may want to place one on the left and the other one on the right. This will help to distinguish them and it will also balance the sound.
Compression affects the dynamic range of a sound signal.
It reduces the amplitude modulations of the sound, that is to say the gap between the lowest and the highest levels.
I will use the above compressor as an example. All compressors have more or less the same basic functions, so it will be easy to reproduce the manipulations on alternative models.
Let's see how this works:
- Threshold: this defines the minimum level from which the compressor will start affecting the sound. 0 is the highest sound level.
- Ratio: this is the level of compression you are going to apply. A 1:1 ratio has no effect, whereas an infinite ratio will limit the sound to the threshold level. In this case, the compressor acts like a limiter and nothing will go higher than the threshold. The higher the ratio, the more compressed the sound will be, and the volume level will be reduced in proportion, until we use the output button.
- Knee: this knob sets how much the compression will affect the sound right before and after the threshold. The softer it is, the softer the transition will be. A hard knee will result in a more abrupt transition. It is rather difficult to hear the effect of this function.
- Attack: this knob indicates the time it takes for the compression to start acting once the threshold is reached. Set it short (0 ms) to compress the sound as soon as the threshold is reached, and set it longer to avoid compressing the attack of an instrument. For example, with a solo guitar, you may want to preserve the string attack and thus set the attack knob on 10 or 20 ms. It all depends on what you wish to do.
- Release: this knob indicates the time it takes for the compression to stop acting once the volume level goes back lower than the threshold. A long release allows for a less brutal change of gain.
- Output: it can also be called "Gain". This knob sets the output volume level after compression of the sound. It compensates for the loss of volume due to the compression.
So basically, compression takes all the sound energy there is between the threshold and the highest level of the original signal and crushes it in a smaller gap. Then you raise the whole signal with the output button, which eventually renders a denser, richer sound compared to the uncompressed signal. This can be subtle or obvious, depending on your settings and the original sound.
Reverberation brings a little air to a recording that would otherwise sound hollow or confined. It is not necessary to add much of it, or it could drown the sound and give the impression that the recording took place in a cathedral, unless this is what you are looking to obtain. A lot of reverb can be nice at some specific moments in a song, on one specific instrument, but I doubt anyone would apply this effect at a high level on a whole song.
It is recommended to put some on vocals. This will add depth. The more you add, the more the sound will seem to move away and get drowned in the mix. On the contrary, the less there is, the more vocals will seem to be close to the listener. A nice little touch of reverb makes vocals sound richer, deeper and more beautiful. Just find the right level.
Delay is a sound effect that allows to... delay an audio signal and have it repeated. It's very much like an echo: shout and you will hear your voice first, then the echo of your voice repeated one or several times.
This effect is somewhat related to the reverb effect, with one major difference: delay repeats at regular intervals without generating a continuous surrounding sound. You can use it instead of the reverb to avoid overcrowding the mix.
As a general rule, set this effect by basing it on the song tempo. Choose whichever interval you wish to hear the echoes. Here is a chart to help you:
Click on the chart to download the corresponding Excel file ("Delay by Tempo & Instrument Frequencies.xls" - 64 KB)
By changing not only the volume, but also the frequencies or some effects like reverb, you can give the impression that a sound is more or less near the listener. This adds depth to a recording.
In everyday life, how do you recognize a distant sound?
Its volume is low. At least relatively. A motorbike engine roaring in the distance will be much less loud than if it were 2 metres away from you. Thus, the first thing to do is to lower the volume of the instrument you want to place in the background.
Distant sounds don't have much bass frequencies. Bass frequencies require a lot of energy to move. Unless you hear a very powerful sound (like an explosion), the bass tone won't get to you. So you may cut low frequencies with a high-pass filter or an EQ.
On the other end of the sound spectrum, high frequencies of a distant sound are also fainted because they are absorbed by obstacles as they're moving. So you also need to lower these high frequencies.
Distant sounds are indistinct, unintelligible. Human voice, for example: if you hear someone speaking in the distance, you will of course recognize that this is a human voice (yes, your brain is that smart), you will probably know whether it belongs to a man or a woman, but you will not understand a word. You are going to have to reproduce this lack of clarity to simulate distance. A little reverb, a little chorus or some phaser touch will do. Give it a try and see what suits your music best.
You can't always tell exactly where the faint sound you're hearing is coming from. So pan the background instruments in the middle position and leave the sides to foreground instruments.
When listening to a music, a listener will mainly focus on changing elements. The use of a repeating, short musical pattern, treated as described above, will be easily forgotten and place itself in the background. The purpose of a background instrument is to add some fullness to the song, without being so interesting that it draws attention. It should be kept simple to have it blend into the song without being noticed too much.
And of course, on the contrary, foreground instruments will be the ones bringing volume, lowest and highest frequencies, melodies, and elements of surprise.
Example of a voice that gradually goes away from the listener (full screen for a better view)
When adding effects, check the result by activating and deactivating them as the song (or the instrument) plays. It will help you to see whether an effect is too present, or not enough, whether it is necessary, whether it is too agressive, etc. You may like heavy reverb, but too much of it just drowns it all and makes the song inaudible. Control is essential. Too little is useless, you might as well remove the effect completely, whereas too much will make everything weighty and will surely waste the whole piece.
An effect often turns out to be... effective if you miss it when it's gone, even though you were not certain to hear it when it was on.
Just wanted to thank you for your time organizing this. It took me days to compile all of these in the past! I am switching to a new workstation. This will be very handy.
* * * * * * * * * * * * * * * *
You're welcome. I'm sorry to say that I don't have much time anymore to update my website, and some information are getting a bit old. Some day, I'll make new tutorials about mixing a song, with better techniques than currently available at the moment, but they're still valid to some extent. Thanks for your support!
Dude, thanks a lot for this tutorial. It really helped me a lot. God bless you!
Sur le "jcm 800 mercurial", quelle différence entre la 12ax7eh et la 12ax7rsd ?
Je trouve la rsd plus "grasse" et plus disto, votre avis ?
* * * * * * * * * * * * * * * *
Effectivement, comme ça simule 2 variantes d'une lampe 12ax7, le son change un peu entre les deux, ce qui permet d'étendre les possibilités du simulateur.
La RSD est un peu plus pêchue et un peu plus brillante, alors que la EH met plus les basses en avant.
À choisir en fonction de ce qu'on cherche comme type de son.
Tout d'abord UN GRAND MERCI pour ton site et tes plugins. Sincèrement, c'est top !
Je suis sur Mac : j'ai téléchargé ignite_amps_nadir_1_0_2_mac et les amp Mac AU.
Bien mis dans le répertoire de mes plugins et components et aucun n'apparaît dans Logic Pro X ?
Quelqu'un sait pourquoi ?
Merci pour l'aide
Hi, I just stumbled on your website. Very impressed I must say. I'm planning to buy TSE X50 also. I'm currently owning a squire Strat and was wondering how I would get a warm jazz tone out of the guitar. Any idea what settings on the TSE X50 could work ? A tone like here : https://www.youtube.com/watch?v=nkW8gKsZiVk