eenixon's picture

This is a follow-on from my previous query about exporting video with a soundtrack. I'm assuming this is not possible currently.

So what to do? Save Images to Movie will do the job. However, as I look inside the resulting file with, say, Media Inspector, I find the frame rate of the resulting ProRes 422 file is 41.369 fps. Assuming you'd like to use the file as a clip in a longer video or would simply like to give the file some shape with fades, etc. this frame rate seems to be problematic when it comes time to render the content in H264 for, say, Vimeo. The export preset in Resolve for Vimeo defaults to a 24fps render. This is not any kind of multiple of the clip's default of 41-and-change. Here's what it looks like as I use (with thanks) balan's simple audio reactive composition with some of my music.

To my eye, it starts pretty well but by about the 1 minute point it starts to go chunky and ends up looking almost like a stop motion project. The synch with the audio seems reasonably OK but is should... flow. Like it does in the Window in Vuo itself.

Yes, there are wheels within wheels, all sorts of factors that could be contributing to this. But, my purpose, is to try to eliminate as many of those factors as possible.

So the question: is there a way to get a movie file out of Vuo that has a frame rate that I can control or at least conforms to one of the standard rates and that will be the product of some process that takes into account as many of the other bottlenecks and limitations of any given set up as possible?

I think I'm back to a viable NRT solution but I'm really new to this software; there are hundreds of wrinkles that I'm unaware of that, if flattened out, might get me where I need to be.

Thanks for your help. ...edN


It's pretty quiet in here.

eenixon's picture
Submitted by

It's pretty quiet in here. Not to worry, I am used to talking to myself.

For those of you who have an interest in this, my state of ignorance, I continue experimenting.

I've found that, while output settings for ProRes 422 lead to frame rates of 41 - 43 fps, output using H264 creates a file of 60fps. Bingo!?

Why? And how is this remedied if it can be?

The result is a then a 100% H264 workflow in the NLE, which is doable if you're not going to mess with the output too much -- colour grading, effects, etc. The results of this change have been substituted for the file you may have looked at yesterday. I think it looks better or at least more in line with what I expected to see based on viewing the Window.

I'd like to understand what is going on with these differences because I can foresee situations where higher quality, less compressed output files will be needed. Is it as simple as setting up a periodic event set to 60fps to run the audio file and thus the rest of the composition? What about the encueing settings along the way? It would sure be nice to have some assistance with this in contrast to having to flail around on my own. Yes, experimentation is a great way to lean stuff but as a famous nineteenth-century French painter said, "He who is self-taught is learning from a very ignorant teacher."

Revised video is here:

Thanks for your help. ...edN

Hmm it is quite in here :)

Scratchpole's picture
Submitted by

Hmm it is quiet in here :) I just ran a very quick test with the example file 'SaveFramesToMovie' that Jaymie created and found it gave me unreliable results: variable frame rate and unable to write the file multiple times, even after changing the overwrite URL function (weirdly it did make me two versions on the third time I openned the comp and then not the forth). I'll try and have a play with it another day.

Why don't you do an non realtime render and then re-add the audio afterwards in your NLE of choice? Another solution would be to use a hardware video recorder such as an Atomos or a cheaper gaming recorder.

Yes, thanks. Synching audio

eenixon's picture
Submitted by

Yes, thanks. Synching audio to video in post is the only current alternative I think. The question for me is about time code. I don't understand the what and how of timecode in Vuo. It appears to be the only synch choice given there is no scratch audio in the video output against which a waveform synch might operate.

So Is this the procedure?

  • export the video NRT with timecode attached -- somehow
  • write out the audio with timecode in real time or, alternatively, as a video-less movie in NRT -- at the same time as the the video...
  • finally import to NLE and synch using timecode.

I haven't been able to find anything in the docs about what type of time code is written to these files; whether it consumes one of the audio tracks or is put into meta-data. How it is 'jammed' to the output files, etc., etc. So I'm feeling like I've been flying blind on this and have lost the appetite for just banging around until something happens.

Are you creating your audio

Scratchpole's picture
Submitted by

Are you creating your audio in Vuo? Why do you want timecode? It feels like you are overcomplicating it to me unless there is something I don't understand in what you are trying to achieve. What is wrong with rendering a movie the exact length of your audio track and re-syncing.

I'm grateful for your input

eenixon's picture
Submitted by

I'm grateful for your input and suggestions.

To answer your second question: I want timecode because it is offered as a feature in this software, it promises a possible enhancement to my workflow and, finally, as I hope I've mentioned earlier, I don't understand how it works based on my reading of the documentation and in relation to my previous experience in shooting and editing video.

If you have ever done video work using a dual system approach, i.e., video to camera, audio to external recorder, you know that synching audio to video is a rudimentary first step in the editing process. By convention, there are tow methods: a) comparing the audio wave forms of the scratch track on the video to the high quality audio from the external recorder, or b) imprinting matching timecode on both video and audio and then merging the two in the editor. Lining up video visually to an audio file, if that is what you're proposing, is fraught. Particularly in a scenario where you might want to modify frame rates and/or use a different piece of video as part of the edit.

All that aside, my reason for posting here is not to critique my way of working or my creative goals. It is to try to get some clarity around how some of the features of Vuo, a truly estimable piece of software, work. I can read what the documentation says the features do but the docs, probably because of my ignorance of the product, do not tell me what the features mean or imply in terms of the usage scenarios I've been describing.

Am I being too obtuse? Nit-picky? My apologies.