How to extract perframe features using mmaction2?
See original GitHub issueHi,
I’m a beginner of video understanding. I have installed mmaction2 and denseflow, and it works well. But I’m not sure how to extract perframe features using it.
- Is there any ready-made apis to solve the above problem? I’ve taken much time to search the solutions, but I didn’t find.
- If no ready-made apis, could you plz share related codes with us? It’s important for beginners.
- What is the difference between the config files eg.
tsn_r50_320p_1x1x8_50e_activitynet_clip_rgb.py,tsn_r50_320p_1x1x8_50e_activitynet_video_rgb.py, specifically between video and clip?
Thank you for your awesome repo. And hope your reply.
Issue Analytics
- State:
- Created a year ago
- Comments:6 (1 by maintainers)
Top Results From Across the Web
Data Preparation — MMAction2 0.24.1 documentation
Extract frames. To extract both frames and optical flow, you can use the tool denseflow we wrote. Since different frame extraction tools produce...
Read more >Feature extraction of BMN using TSN · Issue #158 - GitHub
In the BMN Model Zoo there are results of feature extracted by MMAction but I found on details in Data Preparation about how...
Read more >Weakly Supervised Violence Detection in Surveillance Video
First, they extracted video features using a VGG-f [35] model pretrained in ... [69] can extract bounding boxes within 0.0331 s per frame....
Read more >(PDF) Spatio-Temporal Action Detection Under Large Motion
The TOI-Align operation extracts per-track features from the entire video sequence, ... per-frame features with a RoI-Align operation [16].
Read more >Temporal Pyramid Network for Action Recognition
By sampling frames at different rates as input, backbone networks in [5, 35] are able to extract features of different receptive fields and...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found

By stating extrat perframe features, I am not sure what is the real case. But you may refer to this python file. https://github.com/open-mmlab/mmaction2/blob/fda023269e0fd9ed7ac36d48c67d2b86d422d602/tools/data/activitynet/tsn_feature_extraction.py#L1
Yes, you can refer to mmaction2/tools/data/activitynet/tsn_feature_extraction.py for frame-level feature extraction. The difference between two configs is:
clip_rgbuses clip-level annotations to train the recognizer (only use the groundtruth segments), whilevideo_rgbuses video-level annotations (the whole untrimmed video).