Attention And Positional Encoding Are (Almost) All You Need For Shape Matching

dc.contributor.authorRaganato, Alessandroen_US
dc.contributor.authorPasi, Gabriellaen_US
dc.contributor.authorMelzi, Simoneen_US
dc.contributor.editorMemari, Pooranen_US
dc.contributor.editorSolomon, Justinen_US
dc.date.accessioned2023-06-30T06:19:14Z
dc.date.available2023-06-30T06:19:14Z
dc.date.issued2023
dc.description.abstractThe fast development of novel approaches derived from the Transformers architecture has led to outstanding performance in different scenarios, from Natural Language Processing to Computer Vision. Recently, they achieved impressive results even in the challenging task of non-rigid shape matching. However, little is known about the capability of the Transformer-encoder architecture for the shape matching task, and its performances still remained largely unexplored. In this paper, we step back and investigate the contribution made by the Transformer-encoder architecture compared to its more recent alternatives, focusing on why and how it works on this specific task. Thanks to the versatility of our implementation, we can harness the bi-directional structure of the correspondence problem, making it more interpretable. Furthermore, we prove that positional encodings are essential for processing unordered point clouds. Through a comprehensive set of experiments, we find that attention and positional encoding are (almost) all you need for shape matching. The simple Transformer-encoder architecture, coupled with relative position encoding in the attention mechanism, is able to obtain strong improvements, reaching the current state-of-the-art.en_US
dc.description.number5
dc.description.sectionheadersShape Correspondence
dc.description.seriesinformationComputer Graphics Forum
dc.description.volume42
dc.identifier.doi10.1111/cgf.14912
dc.identifier.issn1467-8659
dc.identifier.pages12 pages
dc.identifier.urihttps://doi.org/10.1111/cgf.14912
dc.identifier.urihttps://diglib.eg.org:443/handle/10.1111/cgf14912
dc.publisherThe Eurographics Association and John Wiley & Sons Ltd.en_US
dc.subjectCCS Concepts: Computing methodologies -> Shape analysis; Theory of computation -> Computational geometry
dc.subjectComputing methodologies
dc.subjectShape analysis
dc.subjectTheory of computation
dc.subjectComputational geometry
dc.titleAttention And Positional Encoding Are (Almost) All You Need For Shape Matchingen_US
Files
Original bundle
Now showing 1 - 2 of 2
Loading...
Thumbnail Image
Name:
v42i5_16_14912.pdf
Size:
14.8 MB
Format:
Adobe Portable Document Format
Loading...
Thumbnail Image
Name:
paper1012_mm1.pdf
Size:
11.31 MB
Format:
Adobe Portable Document Format
Collections