<!-- Some styling for better description lists --><style type='text/css'>dt { font-weight: bold;float: left;display:inline;margin-right: 1em} dd { display:block; margin-left: 2em}</style>

   stdint: As I know, the CTU nor slice layer doesn't support field picture
   <br> paulk-leonov, do you have a sample for hevc field video?
   paulk-leonov: stdint, no, we didn't have time to work on interlaced hevc sadly
   stdint: paulk-leonov, from what I know, the hevc won't be coded as field
   <br> that is why only SEI care about that, not in the slice header
   paulk-leonov: ah? that seems surprising to me
   <br> maybe it's not very common but as far as I know, the spec allows field-coded hevc
   stdint: the problem, if is coded as a frame in CTU, but it will be displayed as field, then decoder should not care about that
   paulk-leonov: jernej, did you have an hevc sample?
   stdint: paulk-leonov, that is why that patch make me confused
   <br> and about the v4l2_h264_dpb_entry which is deleted at that patch
   <br> I think the decoder only need the pic_num for poc
   <br> frame_num is not necessary then when there is a poc
   <br> for a avc frame, frame_num = poc * 2
   <br> and poc * 2 + 1  or poc * 2 for fileds
   <br> I forget which one is for bottom or top
   paulk-leonov: so our decoder hardware definitely has a field for specifying the field when decoding
   <br> so it might be deinterlacing also
   <br> stdint, there is a dedicated series about h.264, feel free to comment there about frame_num/poc :)
   <br> it is most likely redundant indeed
   stdint: paulk-leonov, I doubt on that, maybe it is just the reuse of the h.264
   <br> paulk-leonov, which one is for h.264, I don't watch the mail list for a little long time
   paulk-leonov: stdint, it's that one: https://patchwork.kernel.org/cover/10684569/
   <br> https://patchwork.kernel.org/patch/10684567/ for the controls
   stdint: paulk-leonov, thank you
   tagr: paulk-leonov: hey, I've been porting tegra-vde to V4L2 with some success but hit a bit of a snag
   <br> paulk-leonov: maybe this is obvious and I just missed it, but how do I go from DPB entries to V4L2 buffers?
   paulk-leonov: tagr, hey, well that particular topic is still being discussed actually :)
   <br> tagr, for now, it's based on the timestamp of the capture buffer
   <br> (but we haven't updated the h264 and h265 series to reflect that yet)
   tagr: oh... I see
   paulk-leonov: so the dpb entry will have a u64 "ts" field that matches the internal kernel representation of struct timeval
   tagr: libva-v4l2-request seems to have dpb.buf_index for that, but that doesn't seem to exist anymore in newer patches
   paulk-leonov: tagr, indeed we used to rely on buf index directly
   <br> I have made https://github.com/bootlin/v4l2-request-test/commits/ts-ref for timestamp-based references
   <br> for h265
   <br> we will properly update userspace when the decision is somewhat final and our patch series are updated
   tagr: okay, I'll work with a local hack until then
   <br> paulk-leonov: how do you make sure that buffers stick around long enough?
   <br> I had naively assume some sort of reference counting, but if all you carry around is a timestamp that's rather hard to do
   paulk-leonov: tagr, it's basically up to userspace to not reuse buffers that will be used as reference
   tagr: I see...
   paulk-leonov: tagr, WIP spec is at https://patchwork.kernel.org/patch/10713675/
   tagr: so basically you'd create a bunch of buffers and once decoded you don't reuse buffers until they're no longer used as reference?
   paulk-leonov: tagr, right
   <br> tagr, usually players/ffmpeg will allocate enough buffer so that no problematic situation happens
   tagr: okay, good
   hverkuil: <u>tagr</u>: paulk-leonov: we need better support for making sure the buffers containing the required reference frame are actually still valid. It's one of the reasons cedrus is still in staging.
   <br> paulk-leonov: do you understand Randy Li's email? To be honest, I can't follow his email why using timestamps is bad.
   tagr: this is perhaps naive, but couldn't we queue reference frames as part of a request?
   paulk-leonov: hverkuil, yes totally, but can we do better than increasing the refcount on buffers processed as reference for another frame?
   <br> hverkuil, I don't follow what the issue with timestamps is either
   <br> hverkuil, Randy Li is stdint btw :)
   hverkuil: <u>tagr</u>: There have been a lot of discussions about this and we ended up with the current approach. Anyway, you can't queue a buffer containing a reference frame, you just want to reference a buffer. Actually queuing a capture buffer means that it will be written to, but that's not what you want.
   tagr: mhm... I see
   hverkuil: paulk-leonov: I think there are two approaches: one is that the refcount is incremented when you start a job to decode a frame that depends on a reference buffer, and decremented when the job is done. The other is to increment the refcount when a buffer containing the reference frame is returned to userspace, and decrement it when the queue is canceled or that buffer is queued to the driver again.
   <br> The latter approach assumes that the driver knows that it decoded a reference frame. I'm not 100% certain that that's always known.
   paulk-leonov: hverkuil, that's the issue: the driver can't predict in advance what will be reused as reference later
   hverkuil: So the first approach would be best, but that requires some help from vb2.
   <br> lunch, back later
   paulk-leonov: hverkuil, yes I like the first approach
   <br> ok
   tagr: I obviously missed the relevant discussions, but why was the timestamp chosen instead of the buffer index as a way to reference frames? buffer index seems like a more natural choice to me and sounds like it would be more readily available
   LazyGrizzly: <u>pinchartl</u>: Hi there, concerning patch v4l: Add simple packed Bayer raw12 pixel formats (https://www.mail-archive.com/linux-media@vger.kernel.org/msg134523.html) I thought I'd talk to you before resubmitting.
   <br> All Usb 3.0 cameras from The Imaging Source have either 10 bit bayer packed or 12 bit bayer packed. These formats are basically originating in the UsbVision protocol definition. Format definitions are here:
   <br> https://www.visiononline.org/userAssets/aiaUploads/file/USB3_Vision_Specification_v1-0-1.pdf#G11.1076182
   <br> Will it suffice to address the points brought up by you and sakari or should there be additional things done?
   pinchartl: in addition to addressing the points raised by Sakari and I, could you also link to that document from the patch ?
   <br> that should be all that is needed
   <br> (I didn't know about that specification, thank you for pointing it out)
   LazyGrizzly: sure. no problem
   hverkuil: <u>tagr</u>: the big problem with using a buffer index is that that means that userspace has to predict in which capture buffer a decoded frame will end up. But that's a driver implementation and not something that can be known in general. There are other reasons as well, but that's the main one.
   <br> we had lots of discussions about this, and this is what we ended up with.
   <br> I'm still not entirely happy about using the time stamp to tag buffers (my preference was a separate tag field), but this will do the job.
   tagr: <u>hverkuil</u>: won't userspace know which buffer the decoded frame ended up with after dequeueing it?
   hverkuil: No. Not in general. Drivers can discard capture buffers if they want to, or even reorder capture buffers. It's up to the driver implementation. Another issue is if the codec decodes each slice to a separate capture buffer (doesn't happen know, but might happen in the future).
   <br> In that case a reference frame is actually stored in multiple capture buffers. In this case the driver will give all capture buffers the same timestamp.
   tagr: ugh... that sounds like something that codecs would have a hard time supporting, reference frames made from multiple buffers
   hverkuil: The problem is that this is a public API, so we needed something that was as future proof as we could make it.
   svarbanov: <u>hverkuil</u>: do we have archive for #v4l channel ? I want to check what was the conclusion about addition of vb2::queue_release
   hverkuil: <u>svarbanov</u>: https://linuxtv.org/irc/irclogger_logs/v4l
   ***: benjiG has left
   tagr: paulk-leonov: do you happen to have any code for ffmpeg? or any other players for that matter?
   paulk-leonov: tagr, there's this tree with WIP support directly in ffmpeg: https://github.com/Kwiboo/FFmpeg
   <br> otherwise we use our libva
   tagr: paulk-leonov: I've never used libva before, and it looks like libva-v4l2-request only supports H264 in slices, VDE supports H264 full frame decoding
   <br> but I suppose that should be relatively easy to add
   paulk-leonov: tagr, right, I guess I should give you some context there
   <br> we initially wanted to always have "1 frame worth of slices" as input and no less
   <br> and that's what our libva does, it congregates each slice
   <br> and only sends 1 request with the aggregate
   <br> so I think that's what you want in our case
   <br> however, some bitstream parameters only apply to single slices, so we need to dissociate between slices
   <br> the current idea is to have an array of slice parameters for that, with dedicated offfsets in the same output buffer (where all the slices are still aggregated)
   <br> so each slice param indicates an offset in the buffer, where the slice begins
   <br> but that's not implemented for now
   tagr: paulk-leonov: hmm... VDE also has a mode that allows decoding on slice granularity, I was kind of assuming that that mode would be for V4L2_PIX_FMT_H264_SLICE
   <br> I currently support V4L2_PIX_FMT_H264
   <br> which I assumed would be for decoding from the bitstream of a full frame
   paulk-leonov: yes, the non-_SLICE fashions are for stateful cases where the "hardware" parses the bitstream itself
   tagr: paulk-leonov: does stateless hardware that parses the bitstream itself also qualify?
   <br> VDE basically takes a bitstream of a full frame along with DPB frames and gives you back the decoded frame (or an error)
   <br> so it doesn't keep the state, but it has bitstream and syntax engines that deal with most of the details
   <br> so basically we just need the SPS, PPS, DPB frames and the bitstream
   paulk-leonov: tagr, so that bitstream also contains all the metadata then?
   tagr: no, not the meta data, we still need to pass some parameters from the SPS and PPS
   paulk-leonov: tagr, mhh then I would think that what is left to provide the hardware is only the slices
   <br> but maybe it also needs the NAL header or something?
   tagr: yeah, I think so
   <br> it needs annex-b format
   <br> basically my test program is currently a tiny ffmpeg program that demuxes a file, converts the video stream to annex-b and passes that to the V4L2 decoder
   paulk-leonov: tagr, it looks like it's the same case as rockchip then
   <br> I thought it was an exception, but it looks like it's not
   <br> so we should certainly consider making it mandatory for userspace to provide the full data NALU including the header
   tagr: my understanding is that most userspace will already do that anyway
   <br> the existing VDE interface has a VDPAU driver that works fine
   <br> so that does get the correct data
   <br> and ffmpeg also provides the correct data, though you need to enable the annex-b filter, which ffmpeg does about 50% of the time anyway
   paulk-leonov: at https://github.com/FFmpeg/FFmpeg/blob/master/libavcodec/vdpau_h264.c#L182
   <br> it's adding the start code prefix
   <br> which it doesn't do in VAAPI https://github.com/FFmpeg/FFmpeg/blob/master/libavcodec/vaapi_h264.c#L331
   <br> tagr, do you also need to send the start code?
   tagr: paulk-leonov: hmm... I thought that was what annex-b was doing already
   <br> paulk-leonov: I do have { 0x00, 0x00, 0x00, 0x01 } as the first four bytes of the bitstream that I pass to VDE
   paulk-leonov: tagr, ok great
   <br> tagr, so I think we need to require userspace to pass annex-b format for slices in the spec
   tagr: paulk-leonov: one other thing I'm currently hardcoding for v4l2 is framebuffer modifiers
   <br> VDE on Tegra124 supports the same block-linear formats as the GPU and display pipelines
   paulk-leonov: tagr, the way we went for that is to have a dedicated v4l2 pixfmt that matches a { DRM format, DRM modifier } couple
   <br> our VPU also outputs tiled frames in a custom format
   tagr: ugh... isn't that going to explode combinatorially?
   <br> we have a block-height parameter that can vary from 0-5, so we'd need at least 6 new pixel formats
   <br> actually 7 because we also have "legacy" tiling on older generations
   ndufresne: yes, it will on some platform, someone would need to plumb modifier support, but it's not an easy task
   tagr: could the modifier be passed as a v4l2_ctrl?
   <br> technically the block-height parameter could vary from frame to frame, so it would have to be per-frame, so a control isn't going to work in that case
   Kwiboo: <u>tagr</u>: just a heads up: if you want to try and use my ffmpeg v4l2 request hwaccel that paulk-leonov linked to for h264/hevc you will also need some patches from jernej at https://github.com/jernejsk/LibreELEC.tv/tree/hw_dec_ffmpeg/projects/Allwinner/patches/linux to add some missing fields/ctrl that ffmpeg hwaccel expects
   tagr: it could perhaps in practice, because the block-height would in practice be fixed with constant resolution
   <br> <u>Kwiboo</u>: okay, sounds good, thanks
   Kwiboo: the mpeg2 hwaccel has been tested with cedrus and a rockchip vpu mpeg2 decoder at https://github.com/Kwiboo/linux-rockchip/commit/fc6697bd862ba27dda22c30cf5417d6e68619501
   <br> for h264/hevc it currently send one slice per v4l2 request and that seems to be working with the cedrus driver (with jernej patches)
   <br> if you want to use mpv with my ffmpeg hwaccel you need something like https://github.com/Kwiboo/LibreELEC.tv/blob/rockchip-next-h3/projects/Allwinner/packages/mpv-drmprime/patches/mpv-drmprime-1000-wip.patch and run mpv with --vo=null --hwdec=mpeg2video-drm get mpv to start decoding, I have only used mpv to print out ffmpeg log messages and use kodi-gbm to test decoded frame output on screen
   tfiga: Tagr
   <br> Oops
   <br> <u>tagr</u>: I'd imagine the modifier being added to v4l2_format
   <br> Until now we didn't really have any good case for separate modifiers though
   paulk-leonov: Kwiboo, jernej: I'm about to send out a patch for allowing references to the current capture buffer (for interlaced), would you like some form of credit on the patch (you brought it up initially)?
   Kwiboo: paulk-leonov: great, credit is not needed for my part :)
   jernej: paulk-leonov: Kwiboo actually solved that mistery, so no need from my part
   paulk-leonov: okay, noted :)
   cybrNaut: i'm struggling to find out if linux can drive the Velleman CAMCOLI4 (https://www.velleman.eu/products/view/?lang=en&amp;id=410028)