Hi guys
can you confirm the ordering that bytes are written out for greyscale 12-bit packed RAW images? I'm writing code to unpack them and it only works for certain resolutions, which is really weird. Not sure if bug in camera, or my mistake.
I'm reading chunks of 3 bytes at a time from my file, and converting this into a pair of pixel values. Most of the time it works, but when I change the frame rate & resolution, on occasion I'm getting complete garbage out, even though the on screen display looks good. RAW decoding works reliably at full resolution.
hexdump shows that the data in the unreadable file is somewhat correct (dark regions mostly zeros and bright regions have higher values at the correct points in the file given what I know about the lighting level) but decoding the pixel values in the same way that works at full res. just returns static. Total length of file is correct. As a wild guess, I've tried putting an offset or a bit shift into the start point for reading the file, but this doesn't seem to help. Any suggestions? Can provide sample RAW files (good & bad) which are movies of the same thing, if necessary.
Cheers,
Dan.