Ray-Patch: An Efficient Querying for Light Field Transformers

Tomás Berriel Martins, Javier Civera; Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, 2023, pp. 1158-1163

Abstract


In this paper we propose the Ray-Patch querying, a novel model to efficiently query transformers to decode implicit representations into target views. Our Ray-Patch decoding reduces the computational footprint and increases inference speed up to one order of magnitude compared to previous models, without losing global attention, and hence maintaining specific task metrics. The key idea of our novel querying is to split the target image into a set of patches, then querying the transformer for each patch to extract a set of feature vectors, which are finally decoded into the target image using convolutional layers. Our experimental results quantify the effectiveness of our method, specifically the notable boost in rendering speed for the same task metrics.

Related Material


[pdf]
[bibtex]
@InProceedings{Martins_2023_ICCV, author = {Martins, Tom\'as Berriel and Civera, Javier}, title = {Ray-Patch: An Efficient Querying for Light Field Transformers}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = {October}, year = {2023}, pages = {1158-1163} }