Vienna Symphonic Library Forum
Forum Statistics

196,245 users have contributed to 43,015 threads and 258,398 posts.

In the past 24 hours, we have 1 new thread(s), 7 new post(s) and 152 new user(s).

  • last edited
    last edited

    Thanks for the friendly words, Jack!

    @Jack Weaver said:

    [...]Mir is ahead of other types of reverberators in displaying height (up & down) I would like this to be accentuated even more.[...]

    You are right on track! Actually it would be possible to decode _any_ listening format from MIR's native Ambisonics-recordings. For example, it would be possible to define two circles of speakers in different heights, or even speakers for a z-axis reproduction (on the ceiling above you, or even below he floor ... provided you are living on a grid 😄 ...)

    We have deliberatly restricted Vienna MIR in its present form to "conventional" (planar) reproduction formats up to 8 speakers (and for now, I can't see a huge demand for anything else). But in principle, it would be possible to add a more complex Output Format Editor to MIR Pro for exotic multi-channel setups. ... which doesn't mean that it _will_ be included ... it's just an idea we have in mind.

    ... sorry for hi-jacking the "lush strings"-thread 8-]


    /Dietz - Vienna Symphonic Library
  • Honestly, I was really only thinking about output into stereo speakers.

    VSL is so far ahead of everyone else in the audio industry regarding this type of development. When there is a work flow that meets the demands of post production engineers Mir could be become a standard.

    .


  • last edited
    last edited

    @Jack Weaver said:

    Honestly, I was really only thinking about output into stereo speakers. [...]

    LOL [<:o)] ! Seems as I was using a sledgehammer to crack a nut. Sorry for that.

    Kind regards,


    /Dietz - Vienna Symphonic Library
  • I applaud the implemanation of multi-channel surround options within MIR. I would like to see, however, a single industry standard for the speaker type and positioning for mixing and playback of multi-channel content. For a seamless 360 degree, holosonic, image specific soundfield, 5 to 7 identical (with and optional height channel), front radiating (non dipolar) full range speakers (at ear level) is the optimal approach. But if speaker positioning is one way when the music is mixed, and another upon playback, the multi channel mix be will a distortion from how it was initially concieved. That why I push to have a standard for for the playback of such recordings whether it be in film or music only.  Does Vienna have a speaker positioning standard for the mixing of 5 or 7 discrete channels? And where do you position side/ side rear channels on the circular axis for your mixes? This all may be a little premature for the current offering of MIR, but I think it's good to open up the dialog on this.

    Also, will there be more "hotspot" points within the various rooms to accomodate image specific surround envelopement options. I know this will take the music away from the "spectator sport" perspective, but I feel this coud open the floodgate to new and interesting sonic soundscapes.


  • (Note to myself: Maybe we should split this thread ...)

    Right now, the surround mixes I do are done in a standard, quite precise ITU 5.1 speaker setup. I once tried a 6.1 setup (with a rear center speaker) just for testing. Also a "classical" Quadrophonic setup worked out nicely.

    But during an earlier phase of the Vienna MIR development we had a loose collaboration with Iosono (a spin-off company of the Fraunhofer Institute in Germany). In their showroom studio in Illmenau I had the chance to do a proof-of-concept mix on a full-blown Wave Field Synthesis system. Now _that's_ the kind of surround we were dreaming of as youngsters! 8-) Simply breathtaking.

    -> [URL]http://www.iosono-sound.com/technology/[/URL]

     Looking at the screenshots of their control software, you will see that the two technologies fit like a glove:

    -> [URL]http://www.iosono-sound.com/technology/hardware-and-software/[/URL]

    What a pity that we won't see this approach in the average living room in the near future. :-/


    /Dietz - Vienna Symphonic Library
  • That's fantastic ! Or shall I say lush. [:P]


  • last edited
    last edited

    @Rob Welsh said:

    I applaud the implemanation of multi-channel surround options within MIR. I would like to see, however, a single industry standard for the speaker type and positioning for mixing and playback of multi-channel content. For a seamless 360 degree, holosonic, image specific soundfield, 5 to 7 identical (with and optional height channel), front radiating (non dipolar) full range speakers (at ear level) is the optimal approach. But if speaker positioning is one way when the music is mixed, and another upon playback, the multi channel mix be will a distortion from how it was initially concieved.

    Your suggestion is an excellent one for 7.1 systems. 5.1 already has ITU-R Recommendation BS.775-2 (07/06), which specifies speakers at 0°, ±30°, and ±110°, although it is unclear how many people actually have their speakers laid out this way. (Four of the speakers in a square seems to be more popular.) Ambisonics, however, is fundamentally different from 5.1 and 7.1, and does not need a standard speaker layout. This is one of its great advantages.

    What is encoded in Ambisonics is not speaker feeds, but direction. When mixing in Ambisonics, the positions of the speakers are unknown and are of no interest. Further, when Ambisonics is decoded to speaker feeds all of the speakers cooperate to localise a sound in its correct position so, for example, when the speakers on the left push those on the right pull. The speakers all contribute to the creation of a single coherent soundfield.

    For more information on Ambisonics, please see [url=http://www.wikipedia.org/wiki/Ambisonics]the Wikipedia page[/url].


  • Welcome and thanks for chiming in, Martin! Good to know that our little discussion attracts the interest of an Ambisonics specialist.  :-)


    /Dietz - Vienna Symphonic Library
  • I don't consider the VSL samples to be dry - there is some ambience or early reflections in the sounds which is easily heard when compared to dead dry libraries like GPO. That ambience makes the samples live and makes the stereo image natural. But I have problems to get the instruments to sounds distant when I want to, a problem sometimes with strings but especially percussion and brass. Adding convolution reverb makes a difference, but I still hear the closed-miked origin. If I could wish something, I would have all samples recorded with distant and close miking, with a blend knob. EWQLSO has it in their expensive libraries but I don't know how good it works. The second thing to wish was a function that could replace the speed control, which I don't find working so good. I want automatic chose of articulation according to the length of a played tone and playing style. Not for real time playing obviously, but for already recorded tones. Synful Orchestra has it, and I think it works pretty good (but the character of their sounds is more or less crap...)

  • I agree that there are new ways to handle samples, but I'm sure VSL will have some novelties in that field. They have all the audio they need. Maybe they will find new ways to chop it up, and manage it with their sampler.

    As for the mic distance, what reverb are you using? And are you sending pre or post fader. IMO you need at least Altiverb to work well with VSL, and all the better if you have MIR. And with the features in those verbs, you should be able to control the distince well.

    I have a feeling (I might be wrong) that close/distant micing options with samples is a feature of the past, since the advent of MIR. [Y]


  • I am using SIR1, but I belive the quality of the impulse file is more important than the actual impulse reverb make. Latency is another thing... Imagine you record a voice close and distant. Will they sound the same if the room is dry? I don't think so. The close voice will have more brightness and sound less damped, and have a different dynamics. Even if you put reverb on the close miked voice, pre or post fader, you still can tell it is close voice with reverb on it. I experience something similiar with VSL. Some instrument too closed miked to get depth between instrument groups.

  • Most orchestral recordings are actually a mixture of a main microphone system and spot microphones for added definition. This is exactly what you would achieve by using the room signal created by the MIR engine (i.e the main system) with the direct signals mixed in. Actually this concept was part of the Vienna Symphonic Library's samples from the very beginning.

    Apart from that, Vienna MIR offers a dedicated "distant" Character Preset for most Vienna Instruments to give you even more options on the direct signal itself.

    HTH,


    /Dietz - Vienna Symphonic Library
  • How close are the Vienna instruments typically mic'd?

    The issue of making a drier recording more wet isn't nearly as big a hurdle as making something mic'd close sound like its mic'd from a distance.


  • This depends. Micheal Hula, our musical director, would have all the details (which he woudn't share anyway ;-) ...), but I would say between 1 or 2 m for certain solo instruments up to 5 m and more in case of ensembles (which isn't even close to "close", in my book).


    /Dietz - Vienna Symphonic Library
  • last edited
    last edited

    @Another User said:

    This is exactly what you would achieve by using the room signal created by the MIR engine (i.e the main system) with the direct signals mixed in.
    And in that case classical recordings would be made that way - why bother setting up the extra microphones then?

  • last edited
    last edited

    @Roger Noren said:

    And in that case classical recordings would be made that way - why bother setting up the extra microphones then?

    I don't understand this question, sorry.


    /Dietz - Vienna Symphonic Library
  • 5m seems like it could be reasonable for a group, but one meter for a solo instrument seems way too close for an orchestral sound.


  • last edited
    last edited

    @Roger Noren said:

    And in that case classical recordings would be made that way - why bother setting up the extra microphones then?

    I don't understand this question, sorry.

    What you wrote seems like a contradiction to me. First you say (which I agree upon) that classical recordings are made with balancing close and distant miking, and then you say a convolution reverb can control the depth placement just as good. Then I wonder why classical recordings are not made that way since it seems much easier. My point is that close and distant miking gives different characteristics, which not entirely could be simulated by the CR. In my opinion, the best way to simulate something would be to do it as close to the real thing as possible. I understand that users don't want to have any reverb sometimes, so the close and distant miking should be done in a fairly dry room. Then, by balancing these two, adding the reverb wanted, it should give a convincing result. The demos I've heard from VSL are very impressing in every aspect expect for the depth definition.

  • last edited
    last edited

    @mike connelly said:

    5m seems like it could be reasonable for a group, but one meter for a solo instrument seems way too close for an orchestral sound.

    When you are recording real instruments in an orchestral setting it would be unlikely you'd go further away than 1m with the "spot" mic for your solo instrument. Apart from around this position sounding just fine (along with mixed in pick-up of the room and other orchestra mics) any further and the mic would tend to pick up additional instruments in preference to your solo.

    Julian


  • last edited
    last edited

    @julian said:

    When you are recording real instruments in an orchestral setting it would be unlikely you'd go further away than 1m with the "spot" mic for your solo instrument. Apart from around this position sounding just fine (along with mixed in pick-up of the room and other orchestra mics) any further and the mic would tend to pick up additional instruments in preference to your solo.

    Julian

    I'd agree that 1m may be fine for a spot mic, but in that situation it is also blended with a more distant mic.  In general, orchestral recordings sound most natural using mostly the main mics that capture the entire ensemble, and a small amount of spot mics (when they are needed).

    I'm talking about recording with only one mic and having it at 1m, which seems to be the case with some VSL solo instruments (correct me if I misunderstood).

    If you really wanted to recreate that means of recording a full orchestra, it seems like the best way to do it would be to use multiple mics (more distant mics and closer "spot mics") and let the user mix between the two.  The difference in sound between close and distant micing  is more than just verb and I doubt it can really be simulated well.