Matches in SemOpenAlex for { <https://semopenalex.org/work/W3176799298> ?p ?o ?g. }
- W3176799298 abstract "We present CLIP2Video network to transfer the image-language pre-training model to video-text retrieval in an end-to-end manner. Leading approaches in the domain of video-and-language learning try to distill the spatio-temporal video features and multi-modal interaction between videos and languages from a large-scale video-text dataset. Different from them, we leverage pretrained image-language model, simplify it as a two-stage framework with co-learning of image-text and enhancing temporal relations between video frames and video-text respectively, make it able to train on comparatively small datasets. Specifically, based on the spatial semantics captured by Contrastive Language-Image Pretraining (CLIP) model, our model involves a Temporal Difference Block to capture motions at fine temporal video frames, and a Temporal Alignment Block to re-align the tokens of video clips and phrases and enhance the multi-modal correlation. We conduct thorough ablation studies, and achieve state-of-the-art performance on major text-to-video and video-to-text retrieval benchmarks, including new records of retrieval accuracy on MSR-VTT, MSVD and VATEX." @default.
- W3176799298 created "2021-07-05" @default.
- W3176799298 creator A5020428892 @default.
- W3176799298 creator A5047568087 @default.
- W3176799298 creator A5063529024 @default.
- W3176799298 creator A5065035829 @default.
- W3176799298 date "2021-06-21" @default.
- W3176799298 modified "2023-09-27" @default.
- W3176799298 title "CLIP2Video: Mastering Video-Text Retrieval via Image CLIP." @default.
- W3176799298 cites W1527575280 @default.
- W3176799298 cites W1893116441 @default.
- W3176799298 cites W1927052826 @default.
- W3176799298 cites W2164290393 @default.
- W3176799298 cites W2425121537 @default.
- W3176799298 cites W2626778328 @default.
- W3176799298 cites W2742343242 @default.
- W3176799298 cites W2796207103 @default.
- W3176799298 cites W2808399042 @default.
- W3176799298 cites W2883831329 @default.
- W3176799298 cites W2948859046 @default.
- W3176799298 cites W2950082793 @default.
- W3176799298 cites W2952138345 @default.
- W3176799298 cites W2962784628 @default.
- W3176799298 cites W2965458216 @default.
- W3176799298 cites W2975813532 @default.
- W3176799298 cites W2980037812 @default.
- W3176799298 cites W2990503944 @default.
- W3176799298 cites W2999004587 @default.
- W3176799298 cites W3006320872 @default.
- W3176799298 cites W3007938835 @default.
- W3176799298 cites W3009380496 @default.
- W3176799298 cites W3035265375 @default.
- W3176799298 cites W3035635319 @default.
- W3176799298 cites W3043840704 @default.
- W3176799298 cites W3045687178 @default.
- W3176799298 cites W3092470998 @default.
- W3176799298 cites W3094502228 @default.
- W3176799298 cites W3126721948 @default.
- W3176799298 cites W3127384563 @default.
- W3176799298 cites W3129719298 @default.
- W3176799298 cites W3135367836 @default.
- W3176799298 cites W3137120824 @default.
- W3176799298 cites W3139129996 @default.
- W3176799298 cites W3145807616 @default.
- W3176799298 cites W3147387781 @default.
- W3176799298 cites W3152798676 @default.
- W3176799298 cites W3204588463 @default.
- W3176799298 hasPublicationYear "2021" @default.
- W3176799298 type Work @default.
- W3176799298 sameAs 3176799298 @default.
- W3176799298 citedByCount "4" @default.
- W3176799298 countsByYear W31767992982021 @default.
- W3176799298 crossrefType "posted-content" @default.
- W3176799298 hasAuthorship W3176799298A5020428892 @default.
- W3176799298 hasAuthorship W3176799298A5047568087 @default.
- W3176799298 hasAuthorship W3176799298A5063529024 @default.
- W3176799298 hasAuthorship W3176799298A5065035829 @default.
- W3176799298 hasConcept C153083717 @default.
- W3176799298 hasConcept C154945302 @default.
- W3176799298 hasConcept C184337299 @default.
- W3176799298 hasConcept C185592680 @default.
- W3176799298 hasConcept C188027245 @default.
- W3176799298 hasConcept C199360897 @default.
- W3176799298 hasConcept C204321447 @default.
- W3176799298 hasConcept C2524010 @default.
- W3176799298 hasConcept C2777210771 @default.
- W3176799298 hasConcept C31972630 @default.
- W3176799298 hasConcept C33923547 @default.
- W3176799298 hasConcept C41008148 @default.
- W3176799298 hasConcept C71139939 @default.
- W3176799298 hasConceptScore W3176799298C153083717 @default.
- W3176799298 hasConceptScore W3176799298C154945302 @default.
- W3176799298 hasConceptScore W3176799298C184337299 @default.
- W3176799298 hasConceptScore W3176799298C185592680 @default.
- W3176799298 hasConceptScore W3176799298C188027245 @default.
- W3176799298 hasConceptScore W3176799298C199360897 @default.
- W3176799298 hasConceptScore W3176799298C204321447 @default.
- W3176799298 hasConceptScore W3176799298C2524010 @default.
- W3176799298 hasConceptScore W3176799298C2777210771 @default.
- W3176799298 hasConceptScore W3176799298C31972630 @default.
- W3176799298 hasConceptScore W3176799298C33923547 @default.
- W3176799298 hasConceptScore W3176799298C41008148 @default.
- W3176799298 hasConceptScore W3176799298C71139939 @default.
- W3176799298 hasLocation W31767992981 @default.
- W3176799298 hasOpenAccess W3176799298 @default.
- W3176799298 hasPrimaryLocation W31767992981 @default.
- W3176799298 hasRelatedWork W1480032551 @default.
- W3176799298 hasRelatedWork W1532551495 @default.
- W3176799298 hasRelatedWork W1980967897 @default.
- W3176799298 hasRelatedWork W2064700975 @default.
- W3176799298 hasRelatedWork W2081706464 @default.
- W3176799298 hasRelatedWork W2098498971 @default.
- W3176799298 hasRelatedWork W2098666081 @default.
- W3176799298 hasRelatedWork W2100947710 @default.
- W3176799298 hasRelatedWork W2118638711 @default.
- W3176799298 hasRelatedWork W2119100458 @default.
- W3176799298 hasRelatedWork W2134629245 @default.
- W3176799298 hasRelatedWork W2277801889 @default.
- W3176799298 hasRelatedWork W2988806583 @default.
- W3176799298 hasRelatedWork W3000449477 @default.