At a minimum, you will probably need to add the MPEG start code 0x00000001 to each frame before passing it to your decoder. See H264VideoFileSink::afterGettingFrame1() for an example of this.
You may or may not need to deal with parsing the SProp string depending on whether your H.264 stream contains SPS/PPS NAL units. The mailing list has past discussions on this topic. While configuring and using FFmpeg is outside the scope of this list, your primary decode function will probably be avcodec_decode_video2() from libavcodec. -Barry On Fri, Mar 2, 2012 at 2:10 AM, David Scravaglieri <d...@viddiga.com> wrote: > Hi, > > Well I am really stuck. > > My project is to dump Images from a live H264 stream using live555 and > ffmpeg. > > I started with the testRTSPClient prog and in the afterGettingFrame method > I want to send the fReceiveBuffer to a class that will decode each frames > with ffmpeg and dump all images to a directory. But I can't find any clear > sample code or clear HowTo method in order to do that. > > I have read some posts in the mailinglist. But it's not clear. > How to create a suitable ffmpeg codec context for consuming the > fReceiveBuffer ? > What about using parseSPropParameterSets ? > Do I have to use ffmpeg av_parser_parse2 method ? > > Any help would be appreciated. > > Thanks in advance, > David. > _______________________________________________ > live-devel mailing list > live-devel@lists.live555.com > http://lists.live555.com/mailman/listinfo/live-devel >
_______________________________________________ live-devel mailing list live-devel@lists.live555.com http://lists.live555.com/mailman/listinfo/live-devel