Matches in SemOpenAlex for { <https://semopenalex.org/work/W4386072015> ?p ?o ?g. }
- W4386072015 abstract "In this paper, we efficiently transfer the surpassing representation power of the vision foundation models, such as ViT and Swin, for video understanding with only a few trainable parameters. Previous adaptation methods have simultaneously considered spatial and temporal modeling with a unified learnable module but still suffered from fully leveraging the representative capabilities of image transformers. We argue that the popular dual-path (two-stream) architecture in video models can mitigate this problem. We propose a novel DualPath adaptation separated into spatial and temporal adaptation paths, where a lightweight bottleneck adapter is employed in each transformer block. Especially for temporal dynamic modeling, we incorporate consecutive frames into a grid-like frameset to precisely imitate vision transformers' capability that extrapolates relationships between tokens. In addition, we extensively investigate the multiple baselines from a unified perspective in video understanding and compare them with DualPath. Experimental results on four action recognition benchmarks prove that pretrained image transformers with DualPath can be effectively generalized beyond the data domain." @default.
- W4386072015 created "2023-08-23" @default.
- W4386072015 creator A5067272515 @default.
- W4386072015 creator A5080891156 @default.
- W4386072015 creator A5083639385 @default.
- W4386072015 date "2023-06-01" @default.
- W4386072015 modified "2023-10-17" @default.
- W4386072015 title "Dual-Path Adaptation from Image to Video Transformers" @default.
- W4386072015 cites W1983364832 @default.
- W4386072015 cites W2108598243 @default.
- W4386072015 cites W2126579184 @default.
- W4386072015 cites W2531409750 @default.
- W4386072015 cites W2625366777 @default.
- W4386072015 cites W2962843773 @default.
- W4386072015 cites W2963155035 @default.
- W4386072015 cites W2963524571 @default.
- W4386072015 cites W2981851019 @default.
- W4386072015 cites W2990152177 @default.
- W4386072015 cites W2990503944 @default.
- W4386072015 cites W3035524453 @default.
- W4386072015 cites W3118473641 @default.
- W4386072015 cites W3121523901 @default.
- W4386072015 cites W3138516171 @default.
- W4386072015 cites W3145450063 @default.
- W4386072015 cites W3171516518 @default.
- W4386072015 cites W3172942063 @default.
- W4386072015 cites W3174702398 @default.
- W4386072015 cites W3174770825 @default.
- W4386072015 cites W3176780013 @default.
- W4386072015 cites W3176828726 @default.
- W4386072015 cites W3203711169 @default.
- W4386072015 cites W3206930349 @default.
- W4386072015 cites W4214493665 @default.
- W4386072015 cites W4214516465 @default.
- W4386072015 cites W4214612132 @default.
- W4386072015 cites W4214614183 @default.
- W4386072015 cites W4287891024 @default.
- W4386072015 cites W4312372834 @default.
- W4386072015 cites W4312560592 @default.
- W4386072015 cites W4312658081 @default.
- W4386072015 cites W4312769131 @default.
- W4386072015 cites W4312804044 @default.
- W4386072015 cites W4312884055 @default.
- W4386072015 cites W4313007634 @default.
- W4386072015 cites W4313007769 @default.
- W4386072015 cites W4313071966 @default.
- W4386072015 cites W4313156423 @default.
- W4386072015 doi "https://doi.org/10.1109/cvpr52729.2023.00219" @default.
- W4386072015 hasPublicationYear "2023" @default.
- W4386072015 type Work @default.
- W4386072015 citedByCount "0" @default.
- W4386072015 crossrefType "proceedings-article" @default.
- W4386072015 hasAuthorship W4386072015A5067272515 @default.
- W4386072015 hasAuthorship W4386072015A5080891156 @default.
- W4386072015 hasAuthorship W4386072015A5083639385 @default.
- W4386072015 hasConcept C119599485 @default.
- W4386072015 hasConcept C123657996 @default.
- W4386072015 hasConcept C127413603 @default.
- W4386072015 hasConcept C142362112 @default.
- W4386072015 hasConcept C149635348 @default.
- W4386072015 hasConcept C153349607 @default.
- W4386072015 hasConcept C154945302 @default.
- W4386072015 hasConcept C165801399 @default.
- W4386072015 hasConcept C187691185 @default.
- W4386072015 hasConcept C2524010 @default.
- W4386072015 hasConcept C2776434776 @default.
- W4386072015 hasConcept C2780513914 @default.
- W4386072015 hasConcept C31972630 @default.
- W4386072015 hasConcept C33923547 @default.
- W4386072015 hasConcept C41008148 @default.
- W4386072015 hasConcept C66322947 @default.
- W4386072015 hasConcept C95623464 @default.
- W4386072015 hasConceptScore W4386072015C119599485 @default.
- W4386072015 hasConceptScore W4386072015C123657996 @default.
- W4386072015 hasConceptScore W4386072015C127413603 @default.
- W4386072015 hasConceptScore W4386072015C142362112 @default.
- W4386072015 hasConceptScore W4386072015C149635348 @default.
- W4386072015 hasConceptScore W4386072015C153349607 @default.
- W4386072015 hasConceptScore W4386072015C154945302 @default.
- W4386072015 hasConceptScore W4386072015C165801399 @default.
- W4386072015 hasConceptScore W4386072015C187691185 @default.
- W4386072015 hasConceptScore W4386072015C2524010 @default.
- W4386072015 hasConceptScore W4386072015C2776434776 @default.
- W4386072015 hasConceptScore W4386072015C2780513914 @default.
- W4386072015 hasConceptScore W4386072015C31972630 @default.
- W4386072015 hasConceptScore W4386072015C33923547 @default.
- W4386072015 hasConceptScore W4386072015C41008148 @default.
- W4386072015 hasConceptScore W4386072015C66322947 @default.
- W4386072015 hasConceptScore W4386072015C95623464 @default.
- W4386072015 hasFunder F4320322120 @default.
- W4386072015 hasLocation W43860720151 @default.
- W4386072015 hasOpenAccess W4386072015 @default.
- W4386072015 hasPrimaryLocation W43860720151 @default.
- W4386072015 hasRelatedWork W1891287906 @default.
- W4386072015 hasRelatedWork W2036807459 @default.
- W4386072015 hasRelatedWork W2087937280 @default.
- W4386072015 hasRelatedWork W2353647904 @default.
- W4386072015 hasRelatedWork W2354251581 @default.
- W4386072015 hasRelatedWork W2357461155 @default.
- W4386072015 hasRelatedWork W2384129116 @default.