-
Notifications
You must be signed in to change notification settings - Fork 35
Description
Hi all,
I am hoping someone may be able to assist me?
I have a relatively small YouTube channel and occasionally upload frame rate analysis videos. Like many, I capture lossless or as close to lossless as I can using OBS.
90% of the time my readings (to my knowledge) have been fairly accurate. However, I am currently analyzing a 30fps game play clip, and I may be wrong but something just doesn't feel quite right and I am not prepared to put out any content that isn't as accurate as it can possibly be.
There are a few instances in the clip where the frame rate is locked at 30 and then goes in a sequence 30-29-30-29-30-29-30-29-30-29 etc.
Looking at the delta rendering when this occurs, 2 black frames are displayed with the final black frame showing that the fps has now dropped to 29 and then begins the sequence white-black-white-black-white. I believe this is the correct behaviour? But just wanted to double check. As far as I know the footage is as lossless as I can get it, bit rate of 2607 mb/s. It just seemed odd that there are a few instances in separate clips of the same game where the fps just drops from 30 and runs in a sequence 30-29-30-29-30-29-30-29-30-29. I just wasn't sure if that was actually happening in the game or something wrong with the capture.
I have tried changing the pixel difference anywhere from 0-20 but it makes absolutely no difference to the frame rate that trdrop is detecting.
Any advice/assistance would be greatly appreciated.