Vienna Symphonic Library Forum
Forum Statistics

196,157 users have contributed to 43,014 threads and 258,394 posts.

In the past 24 hours, we have 0 new thread(s), 6 new post(s) and 167 new user(s).

  • I think this measures are good examples how certain instruments work in real world.

    In the higher register of an oboe, very short notes like staccato do have a higher volume range, because the tonal developement with staccatos are not as big compared to longer notes performing in pianissimo. Longer notes develop more tone therefor the volume is usually higher. Further performing vibrato needs more tonal stability, therefor medium long note with vibrato will be always much louder than a staccato in pp.

    In ff usually a staccato note will be always louder than a sustained, the whole possible impact is produced in short amount of time. Longer notes, if they are not performed marcatissimo develop the maximum within the note. Further the tone quality is more demanding on a long note compared to a staccato, the musician has to take care that the tone "works".

    Generally we take a lot of care in editing our samples, of course we can do mistakes - no question about that.

    What we have is a very good reference in our dynamic samples and the dynamic repetition patches for all the mapping values.

    If you need more control about dynamic ranges I strongly recommend VI PRO. Beside setting the individual volume of a certain patch in your template, you can adjust also the dynamic range of each patch individually. And at least you can setup volumegraphs on your keyboard to adjust volume for dedicated play ranges, even on single notes within a patch.

    best

    Herb


  • last edited
    last edited

    @herb said:

    very short notes like staccato do have a higher volume range

    That's something that I completely over looked. Don't get me wrong, I still stand by my suggestion to find ways to lessen the time it takes to fune-tune to the sound you want... but the inconsistancy between volumes makes a lot of sense now.

    With that, I would prefer to keep the range at the maximum as I think restricting the range of an instrument would then produce unnatural results. I suppose something like that has to be handled on a note by note basis, unless you want to restric the ranges or volumes in your template... which as I said might produce an instrument capable of less realistic dynamic flexibility.

    Herb (if you're still reading) or anyone else... I have an ultimatum-like question. I'd either want to strongly suggest ways of reducing the time it takes to fine-tune cross fading to get that perfect sound... (and I'm curious what attempts or concepts could possibly be implemented, if any, to help this) OR I would like some feedback on how to get those results quickly.

    ---------

    This is the only example I have already online but here goes... (Just keep in mind, I know a lot more about mastering, reverb, and mix since I made this... and it was in a massive hurry for a class I had last semester so there are plenty of problems with it.)

    http://soundcloud.com/iscorefilm/graduation-theme

    Below I posted a couple examples of the crossfading used. The french horn example includes more variation with the expression tool instead of the crossfade tool, but the point is that nothing sounded right, like what I imagine a horn player doing, until I drew it like this... and I didn't simply draw it then done, it took listening, drawing, listening, adjusting, a few times to get it the way that sounded right. - The Cello example is simply to show the kind of drawing it takes to get some good phrasing.

    I'm not saying that we shouldn't have to fine-tune... but I don't think anyone would disagree that with standard notation there isn't nearly as much detail given to the performer. Cubase is a performance tool, unlike sibelius, so I get that we need to fine-tune... but I need an easier way... and I can't use a simple mod wheel for it, I feel like 'by hand' is too sloppy when in most instances to get a sound that even sounds remotely real I have to do much 'finer' work than what doing by hand would produce.

    Any advice on how time can be saved here? Or any features possibilities that could lessen the time it takes to fine-tune? This is the whole reason I started the thread.

    Thanks,

    -Sean

    At 00:21 there is a french horn line

    http://img853.imageshack.us/img853/3040/crossfade02.jpg

    At 01:44 there is a cello line

    http://img854.imageshack.us/img854/3848/crossfade01.jpg


  • Thanks for you answer Herb.

    I must beg to differ though. My specs above reflect short notes. In the real world Stacc - Port short - Port med - Port long have a similar attack shape, they are just shorter or longer, with the portato articulations having of course decays that the stacc doesn’t.  These levels vary enoumously (Stacc -24db and Port med -11db). These portato long samples, do "develope" a bit, but I noted the loudest levels.

    You say that "medium long notes with vibrato will be always much louder than a stacc in pp" (I’m not sure that I agree). So if I play a med-long vib note, then a stacc (all in PP) it will sound musical without me editing anything? 

    Have you purposely edited stacc louder than the sus at FF? I’ll check and see if that is consistently the case.

    A performer can play a passage at FF alternating stacc, marcato, sus, legato, pizz, portato, etc, all at a very even level. Or, in another piece he can vary slightly, and in another there will be even more diferences. As you know, every phrase differs, every context differs.

    In the hands of a real musician there are literally millions of articulations because he shapes them according to the phrase, the spirit, and so on. But in the sample world, the articulations are always the morbid same old ones!. In order for me to create music with these samples I need vigorously consistent levels, so that I only have to draw in the "musical phrase and performance" with my automation. Then I can use the great humanize features to introduce "inconsistency", should that add to the context. Instead, far too often I have to spend literally hours with automation, just to bring the samples to a "neural level". 

    As I said, I found the short notes on Trumpet in C (I tested 4 different notes at both low and high vel) to being very playable and musical. But many of the WW and strings imo are much less playable.

    I may not have totally grasped all the capabilities for VIP (that I use) but I don’t see any way to "correct the levels" in any time saving & consistent fashion. And if there were settings to better the performance and these libraries, I would hope that VSL would furnish them to me.

    Can you give us your global editing criteria on what relationship the levels of the different patches  have at VSL? Are legatos always louder than sus? I haven’t detected any consistent pattern in the instruments that I’ve examined closely the last couple days. Do problems arise from the fact that there are sometimes 2, sometimes 3, sometimes 4 different velocity levels?

    To really speak well of these matters, we need lots of audio examples. I will be furnishing some in the coming days, and I invite others to do so as well.

    All best !


  • last edited
    last edited

    @Another User said:

    far too often I have to spend literally hours with automation, just to bring the samples to a "neural level".

    I agree with this 100% - Not all the time, but there are often times that to get the sound that best resembles a real player, things need to be neutral... sometimes it doesn't... But I do spend far too much time doing that, and the phrasing as I mentioned in my other post. I want that flexibility, I just don't want to spend hours per phrase to get that 'real' sound.

    I realize my comments might be redundant, I guess I just feel like this point isn't being heard very well and that it's absolutely essential to having a realistic and cost-effective workflow using samples.

    -Sean


  • I would love to get my hands on some settings that would lessen editing time. I’ve tried say lowering a stacc patch in a cell by say 3 db. That may "work" at PP velocity leve with regards to a sus patch, but at a FF level it won’t. By reducing the dynamic range of a patch, what is that going to do for me? I’ll probably need all the possible dynamic extremes.

    That’s not really what is at hand here though. We’re talking about basic stuff. At velocity 64, playing sus, stacc, & legato, and getting normal results on all instruments.

    If there are really workable settings in VIP, I’d love to have it. But if there were settings to make the basic stuff easier, why haven’t we heard about them?


  • You guys are thinking too much. I understand the wish to save time, but you've been using VSL libraries successfully for many years (I suspect), with these 'transition' faults, without batting an eye, until it became a forum topic. If it was a real issue it would have been addressed a long time ago by VSL. It's an affectation, not an issue. No one but those who listen on a microscopic level hear this, I'm sure. Who do you want to please? The people who listen, or the people who play? If listeners, it's a non-issue; if players, well, that's your ego talking. If yourself, learn to dumb it down like has always been the case.

    I am speaking grossly of course, maybe some patches need reworking, but in general, no one hears these slight differences, except the musician who wrote the piece; so move on and think in 8 bit for your listeners.

    Shawn


  • last edited
    last edited

    @jammusique said:

    I’ll probably need all the possible dynamic extremes

    That's just the point though... if an oboist can play a staccato note louder than a sustain... then having a full dynamic range would not allow for a consistant volume level between patches. - One thing I haven't played with is the x-fade on/off option. I would certainly be easier than crossfades in certain instances... but there are still plenty of times where I need cross fading on and I still run into the problem... so it would still be nice to have it addressed.

    To VSL (or anyone):

    This thread has consisted of two issues, 1) the volume consistancy bit, which has been addressed and 2) the crossfade bit about limiting the amount of work it takes to fine-tune to get a good result... which hasn't properly been addressed. I'm not trying to diminish the first point, but I created the thread entirely for the second point. Please, any answers, feature posibilities, suggestions for users?

    Thanks,

    -Sean


  • Hi all, I could'nt agree more about volume issues. I still think VSL have the best samples ever, but a lot of people are moving to other company products due to the compexity of VSL. I like this complexity when it permits me to use tons of articulations (as VSL "only" permits). But I like less the fact that I need to fine tune my volume when going from an articulation to another one. But my main concern is the phase on velocity crossfading! It has been created while ago and since it seems that time haas changed and a lot of improvement have been found. Would it be possible to change the velocity crossfading to try to avoid the most the phase and the "multi" instrument effect (ie when using solo violin and riding the xfading and hearing two violins because samples are layering). Regards

  • last edited
    last edited

    @Hicks said:

    my main concern is the phase on velocity crossfading!

    I'm glad you mentioned this. I've been writing and noticing it, then thinking to post that on this thread... but I kept forgetting. I agree completely! Phasing is definitely a concern I have and the solo violin was the exact example I was using a few hours ago.

    Could a 'crossfade humanize' feature be possible? Something to automate the imperfections that a real player has... the reason I mention this is because VSL's samples are recorded as plain and even as possible. I certainly get why, there is a need for it in many ways... but it's also in every way inhuman. I still want the ability to fine tune, but the current humanize feature for pitch imperfections allows for both fine tuning and a degree of automatic 'humanization'. Having a similar feature for crossfading, making things more human rather than very flat dynamically in my opinion would be an essential feature.

    In the meantime, I agree... VSL still has the best samples available...

    -Sean


  • last edited
    last edited

    @Hicks said:

    But my main concern is the phase on velocity crossfading! It has been created while ago and since it seems that time haas changed and a lot of improvement have been found. Would it be possible to change the velocity crossfading to try to avoid the most the phase and the "multi" instrument effect (ie when using solo violin and riding the xfading and hearing two violins because samples are layering). Regards

     

    This is a very real concern and I would hope that VSL is actively working on solving it. Phase aligning such a huge sample set is a really daunting task, so i would settle for just the legato and sustain samples. However, this would not suit the people who like to do all dynamic control with a continuous controller. This has always seemed counter-intuitive to me, as "one-shot" samples are perfectly suited to velocity control, but would those people be prepared to accept a compromise?

    AFAIK the only sample library developer that allows you to cross-fade on solo instruments without phasing is Samplemodeling, and that is not really a sample library. I would imagine that phase aligning a few MB of samples takes far less time than a few hundred GB. [;)]

    DG


  • last edited
    last edited

    @iscorefilm said:

     

    Could a 'crossfade humanize' feature be possible? Something to automate the imperfections that a real player has... the reason I mention this is because VSL's samples are recorded as plain and even as possible. I certainly get why, there is a need for it in many ways... but it's also in every way inhuman. I still want the ability to fine tune, but the current humanize feature for pitch imperfections allows for both fine tuning and a degree of automatic 'humanization'. Having a similar feature for crossfading, making things more human rather than very flat dynamically in my opinion would be an essential feature.

     

    I think that for large ensemble patches this is not really necessary, as those imperfections are ironed out when you have many players playing together in the same patch. However, it is useful to have some human element for solo instruments, and I would recommend getting a breath controller so that you can play the expression live. I know some people use a control surface, or even an expression pedal, but particularly for Woodwind and Brass instruments, I wouldn't be without my BC.

    DG


  • I definitely agree with the request for matrixes where the levels of the various articulations are matched.  I see the reasoning behind doing it the other way but that's much less useful for me - if I'm playing a phrase that mixes different lengths of notes, I want to be able to just play the phrase in and have the dynamics reflect how I'm playing it in, not jump all over the place every time I switch articulations.  Seems like the best solution would be to improve the patches as opposed to adding more interface functionality for each user to try and fix it themself.

    If there are users who like it the other way, then what about having some sort of switch in the interface?  Or providing two sets of matrixes?

    The inconsistent transitions between dynamic layers are an important issue as well, thankfully I have only noticed problems a few times with the VSL libraries.


  • last edited
    last edited

    @DG said:

    I would recommend getting a breath controller so that you can play the expression live.

    I have one last question for you then. My sole goal is to save time... I often get distracted from writing, specifically 'moving forward' with a piece cause I spend so long working on the getting the beginning sounding right. I'm OCD and if it doesn't sound right I just don't want to move on.

    would using a breath controller be faster (more accurate) than using a fader or mod wheel? Because I'm used to drawing in this expression, I do very fine-tuned 'crossfade drawings', so to speak, so I want to know basically if the BC will be better at that than faders, knobs, etc. While those seem faster, they don't allow me to fine-tune quite as accurately.. thus my question.

    Thanks,

    -Sean


  • last edited
    last edited

    @mike connelly said:

     Seems like the best solution would be to improve the patches as opposed to adding more interface functionality for each user to try and fix it themself.

    I disagree. Herb outlined the perfect reason to have different volume levels, because if you want an instrument to be as 'real' as possible, with all the dynamic range it can.. certain samples will have more range than others... making this problem inevitible. Changing the samples wouldn't be an improvement, it would take away true functionality. I think this has a very simple sollution... a button (possibly linked to a midi cc) that turns on volume leveling between patches or turns it off... just like the vel x-fade on/off feature.

    This pretty much would solve this problem for everyone, without having to make it tidious to users to make mass template changes, tedious mixer settings, etc...

    -Sean


  • last edited
    last edited

    @DG said:

    I would recommend getting a breath controller so that you can play the expression live.

    I have one last question for you then. My sole goal is to save time... I often get distracted from writing, specifically 'moving forward' with a piece cause I spend so long working on the getting the beginning sounding right. I'm OCD and if it doesn't sound right I just don't want to move on.

    would using a breath controller be faster (more accurate) than using a fader or mod wheel? Because I'm used to drawing in this expression, I do very fine-tuned 'crossfade drawings', so to speak, so I want to know basically if the BC will be better at that than faders, knobs, etc. While those seem faster, they don't allow me to fine-tune quite as accurately.. thus my question.

    Thanks,

    -Sean

     

     I have two suggestions for you:

    1. Actually write the music. Don't go straight to programming. It may sound a long way round, but I find it is quicker to do a rough pencil sketch, type a score, then program. That way the creative bit is done fast and the rest is the performance. They are two different things. If I'm writing using a lot of synths, then might do a very rough string demo before doing it properly, but for orchestral stuff it is much slower for me to program with no printed score.
    2. Depending on how you set the BC up it could well be quicker and more musical. I play all my Ww/Br legato lines in using the BC and then, if necessary, edit the BC CC data. I've never tried using BC or even xFade on short dynamics, as it doesn't seem natural to me.

    DG


  • last edited
    last edited

    @iscorefilm said:

    I disagree. Herb outlined the perfect reason to have different volume levels, because if you want an instrument to be as 'real' as possible, with all the dynamic range it can.. certain samples will have more range than others... making this problem inevitible. Changing the samples wouldn't be an improvement, it would take away true functionality. I think this has a very simple sollution... a button (possibly linked to a midi cc) that turns on volume leveling between patches or turns it off... just like the vel x-fade on/off feature.

    That's exactly what I suggested, so I guess you agree with me after all.  There's an argument to be made for both ways of doing it - since different users will want it either of the two ways, the best solution is to provide the option to have the patches behave either way.


  • lol, well maybe I do... it just seemed like you wanted to change the patches themselves, not with VI Pro but the files... which to me wouldn't have solved it for both sides. But I think the button idea is great, I don't know if VSL would consider it... but I think it wouldn't be hard to implement and would solve this issue.

    -Sean


  • Has anyone tried to set the dynamic range to 0, then link Expression and Vel XF to the same CC? I tried this briefly tonight and it seemed to smooth things out a bit-- though my ears could be deceiving me. MOH

  • I did try playing with setting expression and Vel X-fade to the same CC, and don’t remember thinking that it worked, but will try again adding also the dynamic range into the equation.

    I doubt that VSL would go back and change levels on patches, but continue to think that some behind the scene correction might be possible. Like the "button" mentioned above, maybe an "automatic" (that’s the tricky part) gain adapter with a fader to either let different levels pass (like when you want a sfz), and when you want smooth levels.

    I don’t understand anything about phasing and related subjects, but also think that the future for these VSL libraries, would be a VIP (3.0?) that would go beyond simple patch playing, but rather morphing (whatever that means) between various base patches to achieve the sound (definable though automation). I’m thinking along the lines of some advances of the libraries that I mentioned earlier.


  • I believe I suggested a morph feature toggle button for the velo X fade for VI Pro 1.0. Something similiar to Kontakt. I quite certain that something like that would improve drastically the blend between the layers.

    Also I'm still hoping for a fix for some legato instruments. I mean, imho I feel that arranging for appasionata strings takes far less time to get the legatos sounding smooth than the orchestral strings, for instance. In fact I use them much more often. I don't know, perhaps these fixes have to be done to the libraries themselves rather than the player.