[01:29] <CarlFK> sudo modprobe vivid, chromium-browser, https://webrtc.github.io/samples/src/content/devices/input-output/   shouldn't I see vivi as a camera choice?
[07:03] <kbingham> CarlFK, On firefox, I see my UVC webcam, VIMC and VIVID as camera options. On Chrome I only see my UVC and VIMC. So it could be some browser difference in their layers for handling V4L2/Video inputs.
[07:06] <CarlFK> kbingham: thanks.  just tried FF, and yes, I see ... vivid, webcam, vivid ...
[07:07] <CarlFK> not sure I care about the 2x vivid.  really trying to track down why my thinkpad cam only works 1/2 the time.
[07:08] <kbingham> rcoote, do you mean display it with a test application like qv4l2?
[07:09] <kbingham> or you want to make your own application which displays video?
[07:11] <kbingham> CarlFK, I see, then you probably need to isolate if it's the camera not working, (i.e. check with an application like qv4l2 / guvcview / cheese) and check the kernel logs if you see failures in those apps.
[07:50] <rcoote> kbingham: I need to create an own application.
[08:04] <bobsaccamano> Hi..is it possible to have a "Virtual" camera device that contains (fixed) projections of multiple physical v4l cameras?
[08:05] <kbingham> rcoote, For QT support you'll need a qt channel ... but libcamera does have a very basic test application for rendering video in a qt window, which might be useful as an example https://git.linuxtv.org/libcamera.git/tree/src/qcam
[08:05] <kbingham> (but it doesn't use the v4l2 api ... only 'libcamera' api of course)
[08:06] <kbingham> bobsaccamano, Is it possible? yes - but that sounds like the sort of thing done by specialised hardware to do the mapping.
[08:07] <kbingham> bobsaccamano, you could do simple compositing of mulitple cameras with gstreamer for example.
[08:08] <kbingham> or use hardware suitable for compositing multiple images onto a destination frame as is common in display hardware. Or on automotive SoCs there are hardware components that can map images to a 3d surface to handle things like the 360 / 3d camera surround views
[08:09] <bobsaccamano> kbingham: thanks! The latter is exactly what I'm looking for
[08:10] <bobsaccamano> Is the mapping of multiple cameras to a 3D surface also possible in software? (perhaps GPU-accelerated/CUDA)?
[08:11] <kbingham> bobsaccamano, if you want to do more than one, then you certainly need hardware accelleration. And sure, it can probably be done with openGL perhaps or something like that - but it's way out of scope for this channel I'm afraid (and beyond my knowledge base)
[08:11] <kbingham> (more than one single image ;D)
[08:12] <bobsaccamano> well, in this case it is 3-4 video streams from an equal number of V4L devices ;)
[08:12] <bobsaccamano> cool, that's helpful. Thanks again
[09:17] <rcoote> kbingham: Thanks. I am forced to use V4L. This is the /dev/video0 and /dev/video1 . qv4l2 looks good. I should try to port it to embedded linux. That is the device I am working on.
[09:18] <kbingham> rcoote, what device? Embedded linux is a fairly broad coverage for anything that is not a laptop/PC
[09:36] <rcoote> kbingham, basically an IMX6 board running a yocto linux distribution.
[09:44] <kbingham> rcoote, I think qv4l2 is part of the v4l-utils package.
[09:45] <kbingham> rcoote, Except openembedded (the yocto base) disables qv4l2 - so you'd have to manually add it back and fix whatever is the reason they disable it ;-) https://github.com/openembedded/meta-openembedded/blob/master/meta-oe/recipes-multimedia/v4l2apps/v4l-utils_1.18.1.bb#L30
[10:15] <tfiga> rcoote: why are you forced to use V4L2?
[10:15] <tfiga> libcamera still uses V4L2 as its backend to talk to the video devices, but it just exposes a higher level API to the applications
[10:20] <kbingham> rcoote, And infact libcamera is in the openembedded/yocto recipes too :)
[14:21] <ndufresne> ezequielg: so, if we move out the invariant, we no longer need slice params for the Hantro case
[14:22] <ndufresne> and for cedrus, we only ever pass 1 slice params at a time, so my questions is if there is a way to pass all slices buffers/params at once ?
[14:22] <ndufresne> hence if slice_params should still be an array
[14:23] <ezequielg> ndufresne: right, basically allowing applications to choose between models: 1-slice per request, or N-slice per request. Now, why would anyone choose the former?
[14:23] <ndufresne> Or perhaps cedrus supports having all slices in 1 buffer, using the offset in the slice params ?
[14:24] <ndufresne> I'm asking since we have no userspace for the former, unless chromium ? I haven't check the code there
[14:24] <CarlFK> kbingham: thank you for mentioning qv4l2 - I have been wanting something like that for years.
[14:30] <ezequielg> ndufresne: iirc, jernej and Kwiboo have implemented this using ffmpeg/librelec for cedrus. i can be wrong though.
[14:31] <ezequielg> once again, iirc, the V4L2_BUF_FLAG_M2M_HOLD_CAPTURE_BUF flag was introduced to support cedrus this way.
[14:56] <ndufresne> ezequielg: ffmpeg is single slice params, using HOLD_CAPTURE_bUF
[14:56] <ndufresne> with a fallback to requeuing the capture buffer
[14:57] <ndufresne> While it does fill the slice array for the other cases, it will stop filling passed 16 entry, and it does not matter since Hantro only reads the first one
[14:57] <ezequielg> hverkuil: i don't think we can easily support passing an array of slice params controls, do we?
[14:58] <ezequielg> i think for that to work sanely, we'd have to support variable-length control arrays.
[15:05] <jernej> yeah, main issue is that variable-length control arrays are not supported
[15:06] <jernej> that would be very useful feature for hevc too
[15:20] <hverkuil> hverkuil: yeah, variable length array support is the missing feature here.
[15:21] <tfiga> Why do we need variable length?
[15:21] <tfiga> The array can be fixed length, we just need num_slices somewhere
[15:23] <tfiga> ezequielg: jernej: hverkuil: ^^
[15:28] <jernej> do you know max. number of slices?
[15:28] <ezequielg> tfiga: iirc, v4l2 needs to know how many elements in a control array at control creation time, in the driver open/probe.
[15:28] <ezequielg> so you would have to pass max number of slices, and that's not ideal.
[15:28] <jernej> iirc it can be absurd number
[15:29] <ezequielg> jernej: yup, exactly.
[15:29] <jernej> similarly in hevc, there is one field, which can be extremely big in theory
[15:30] <tfiga> Well, one can still set it to 16 or so and probably wouldn't ever hit a case where it's not enough
[15:30] <jernej> s/field/array/
[15:31] <hverkuil> I think I will have to give this another go. It's a lot more difficult to get right than it looks at first glance (which is why I didn't implement this yet).
[15:31] <hverkuil> But for these stateless codecs it is a desirable feature.
[15:32] <tfiga>  AFAIR libav just assumed a max of 32
[15:34] <ezequielg> hverkuil: indeed, it would be nice to have.
[15:36] <ezequielg> i think we can still move forward with the uapi, following the changes i recently posted -- wether you are able to pass an array of slice controls, or not, the uapi should be the same. or is this too naive?
[15:39] <hverkuil> Well, it would mean that the stateless codec API will be dependent on a new feature (variable length controls). The struct representing an array element will be the same in either case though.
[18:42] <Kwiboo> hverkuil: variable-length array is very desirable feature :-), for rkvdec hevc I am currently forced to use a hard coded .cfg.dims = { 16 } and had to add a num_slices field to a separate ctrl in order to signal how many ctrl elements contains valid slice data
[18:44] <Kwiboo> the rkvdec hw can be configured for up to 600 slices in hevc mode
[19:03] <ndufresne> tfiga: there is no upper bound for slices
[19:04] <ndufresne> I think libav will break sooner then later with 4K and 8k content
[19:05] <ndufresne> but the point I was raising, is that with the move of the invariant to the decode params, and cedrus begin trully slice based, we never use more then one anymore in any implementation
[19:05] <ndufresne> tfiga: or perhaps chromium ?
[20:47] <tfiga> ndufresne: one slice per buffer is not practical unless one bumps the maximum number of buffers
[20:48] <tfiga> Chromium packs all the slices into one buffer
[20:49] <tfiga> Generally I haven't seen multislice videos in real life too much, at least H.264
[21:01] <ndufresne> tfiga: well, on hantro/rkvdec, only the slice[0] is ever used
[21:02] <ndufresne> and for this reason we are moving all the slice invariants into the decoder params, so the hantro/rkvdec (what chromium works for) to no requires any slice_params
[21:02] <ndufresne> as for cedrus driver, it's only tested with 1 slice/buffer, so you make a request, with 1 buffer 1 slice_params and queue
[21:02] <ndufresne> so all userspace that works on cedrus strictly use 1 slice params
[21:03] <ndufresne> tfiga: btw, try Apple trailers ;-P
[21:03] <ndufresne> or make a WebRTC call with Safari
[21:04] <ndufresne> tfiga: packing all slice in one buffer + with starts codes is of course what all userspace do for hantro/rkvdec, but the subject was about the slice parameters
[21:08] <tfiga> ndufresne: we also have the stateless mediatek, but I don't remember if it supported per slice decoding
[21:09] <tfiga> Still, as I said above, with current maximum number of buffers in v4l2, per-slice decoding is not practical...
[21:09] <tfiga> I suppose the userspace you mentioned isn't really used anywhere in production
[21:10] <ndufresne> depends what you define, khodi and gstreamer (but that's early)
[21:10] <ndufresne> Note, I still don't yet agree, in the sense that I do double slice decoding to maintain good parallelism
[21:11] <ndufresne> so I have at most two sliced queued per frame, and then the final bit is poped later, after reordering
[21:11] <ndufresne> and the yield pretty decent performance here, even at 16 slices
[21:12] <ndufresne> tfiga: but you are right that we cannot do like VAAPI, in vaapi you can wrap all sliced into buffers, and same for slice_params, everything is "buffer", and just send that in one go to your GPU
[21:13] <ndufresne> but that method is also unique to desktop, as they use a lot more memory
[21:13] <tfiga> It's really poor from power point of view, because you need to do frequent synchronization.
[21:14] <tfiga> Rather than just asynchronously queuing the slices as they come to you and getting back free buffers together with a ready frame
[21:14] <ndufresne> if we really wanted something that is nice for power, we'd move away from v4l2 and build a new sub-system
[21:14] <ndufresne> (personal opinion)
[21:14] <tfiga> For video codecs, I don't think v4l2 is missing that much
[21:15] <tfiga> Higher maximum number of buffers and better handling of data offsets
[21:15] <ndufresne> note that I do asynchronous queuing of slices in gst, just that I picked a depth of 2, but it would be trivial to pick a larger depth
[21:15] <tfiga> And then one can just use a big ring buffer for queuing the bitstream
[21:15] <ndufresne> but I cannot do full frame, cause 32 is too low
[21:16] <ndufresne> for the power, the way we pass the params seems pretty inefficient
[21:18] <ndufresne> tfiga: do you think group N slices into 1 v4l2_buffers, with N < num_slice/frame would be fine ?
[21:18] <ndufresne> it's would be a small gain over using N request + N buffers
[21:18] <ndufresne> even though, I don't poll all these request, the order is guarantied, I only need to poll the last queued, and then I can assume all other request are done
[21:22] <tfiga> I think it would definitely be better than 1 slice per buffer, but that's not a problem on its own.
[21:29] <ndufresne> tfiga: anyway, let's fix what's broken first, right now we can send all slices in 1 buffer, but we may not have enough slices in the control array (cause a control array has static size)
[21:29] <ndufresne> I think leaving it with arbitrary 16 slices is very poor, and bad for HEVC
[21:30] <ndufresne> and that's only needed for Cedrus, slice parameters will soon be optional
[21:30] <jernej> tfiga: currently Cedrus is not designed to process more than one slice per request
[21:30] <ndufresne> tfiga: btw, if there is such a thing as MTK stateless, it would be good nice to provide info
[21:30] <tfiga> jernej: that's okay, as long as it's asynchronous
[21:31] <tfiga> ndufresne: patches are on the list
[21:31] <tfiga> Have been for quite a while.
[21:31] <jernej> well, current ffmpeg request api implementation is such that it waits on each slice being processed
[21:32] <jernej> (out of tree, of course)
[21:32] <jernej> and yes, it's used in production
[21:32] <jernej> but not in any stable release, yet
[21:33] <jernej> neither me nor kwiboo found time to rewrite this logic to wait only on last slice
[21:42] <tfiga> In any case, per slice operation could be practical from chromium point of view if we could allocate one big ring buffer and queue the same buffer multiple times with different offset and bytes used
[21:43] <tfiga> And if there is enough buffer indexes to start queuing second frame without waiting for the first one to be decoded
[21:44] <tfiga> But we'll likely just stick to the full frame model to avoid maintaining two different code paths
[21:44] <tfiga> Because we need to support hantro
[21:51] <Kwiboo> ndufresne: have you seen any performance gain from having multiple requests in-flight? performance wise ffmpeg and rkvdec is fast enough for 4k@50hz 10-bit hevc broadcast samples with waiting using a single request
[22:02] <tfiga> Have you done any testing in a battery powered or thermally constrained system in a regular environment, not just isolated playback?
[23:08] <Kwiboo> no I have not, I have mainly tested on my rock pi 4 hanging in open-air from tv :-), I do know that 4k@30fps 10-bit hevc has been tested on rk322x devices showing similar performance/cpu usage as bsp kernel+mpp
[23:51] <CarlFK> 2.6 seconds after boot and I get  [    2.641593] ------------[ cut here ]------------  which maybe I should do something about before trying to fix  [   81.029575] usb 2-1.6: USB disconnect, device number 4  -- dmesg: http://paste.ubuntu.com/p/Cs2j3wdcF8/
[23:52] <CarlFK> the disconnect might be the cause of my random webcam problems