在线时间:8:00-16:00
迪恩网络APP
随时随地掌握行业动态
扫描二维码
关注迪恩网络微信公众号
开源软件名称(OpenSource Name):frostinassiky/gtad开源软件地址(OpenSource Url):https://github.com/frostinassiky/gtad开源编程语言(OpenSource Language):Python 86.2%开源软件介绍(OpenSource Introduction):G-TADThis repo holds the codes of paper: "G-TAD: Sub-Graph Localization for Temporal Action Detection", accepted in CVPR 2020. Update15 Dec 2020: The configuration for HACS Segment dataset is in the 24 Nov 2020: to celebrate my 2nd anniversary with Sally,
I released the code for ActivityNet. :P Please checkout the branch 30 Mar 2020: THUMOS14 feature is available! GooogleDrive, OneDrive 15 Apr 2020: THUMOS14 code is published! I update the post processing code so the experimental result is slightly better than the orignal paper! 29 Apr 2020: We updated our code based on @Phoenix1327's comment. The experimental result is slightly better. Please see details in this issue. OverviewTemporal action detection is a fundamental yet challenging task in video understanding. Video context is a critical cue to effectively detect actions, but current works mainly focus on temporal context, while neglecting semantic context as well as other important context properties. In this work, we propose a graph convolutional network (GCN) model to adaptively incorporate multi-level semantic context into video features and cast temporal action detection as a sub-graph localization problem. Specifically, we formulate video snippets as graph nodes, snippet-snippet correlations as edges, and actions associated with context as target sub-graphs. With graph convolution as the basic operation, we design a GCN block called GCNeXt, which learns the features of each node by aggregating its context and dynamically updates the edges in the graph. To localize each sub-graph, we also design a SGAlign layer to embed each sub-graph into the Euclidean space. Extensive experiments show that G-TAD is capable of finding effective video context without extra supervision and achieves state-of-the-art performance on two detection benchmarks. On ActityNet-1.3, we obtain an average mAP of 34.09%; on THUMOS14, we obtain 40.16% in [email protected], beating all the other one-stage methods. Dependencies
InstallationBased on the idea of ROI Alignment from Mask-RCNN, we devoloped SGAlign layer in our implementation. You have to compile a short cuda code to run Algorithm 1 in our paper.
Data setupTo reproduce the results in THUMOS14 without further changes:
Code Architecture
Train and evaluationAfter downloading the dataset and setting up the envirionment, you can start from the following script. python gtad_train.py
python gtad_inference.py
python gtad_postprocessing.py or bash gtad_thumos.sh | tee log.txt If everything goes well, you can get the following result:
BibtexCVPR Version.
ReferenceThose are very helpful and promising implementations for the temporal action localization task. My implementations borrow ideas from them.
Contactmengmeng.xu[at]kaust.edu.sa |
2023-10-27
2022-08-15
2022-08-17
2022-09-23
2022-08-13
请发表评论