question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

How to extract perframe features using mmaction2?

See original GitHub issue

Hi,

I’m a beginner of video understanding. I have installed mmaction2 and denseflow, and it works well. But I’m not sure how to extract perframe features using it.

  1. Is there any ready-made apis to solve the above problem? I’ve taken much time to search the solutions, but I didn’t find.
  2. If no ready-made apis, could you plz share related codes with us? It’s important for beginners.
  3. What is the difference between the config files eg. tsn_r50_320p_1x1x8_50e_activitynet_clip_rgb.py, tsn_r50_320p_1x1x8_50e_activitynet_video_rgb.py , specifically between video and clip?

Thank you for your awesome repo. And hope your reply.

Issue Analytics

  • State:closed
  • Created a year ago
  • Comments:6 (1 by maintainers)

github_iconTop GitHub Comments

2reactions
makecentcommented, Jun 1, 2022

By stating extrat perframe features, I am not sure what is the real case. But you may refer to this python file. https://github.com/open-mmlab/mmaction2/blob/fda023269e0fd9ed7ac36d48c67d2b86d422d602/tools/data/activitynet/tsn_feature_extraction.py#L1

1reaction
kennymckormickcommented, Jun 2, 2022

Yes, you can refer to mmaction2/tools/data/activitynet/tsn_feature_extraction.py for frame-level feature extraction. The difference between two configs is: clip_rgb uses clip-level annotations to train the recognizer (only use the groundtruth segments), while video_rgb uses video-level annotations (the whole untrimmed video).

Read more comments on GitHub >

github_iconTop Results From Across the Web

Data Preparation — MMAction2 0.24.1 documentation
Extract frames. To extract both frames and optical flow, you can use the tool denseflow we wrote. Since different frame extraction tools produce...
Read more >
Feature extraction of BMN using TSN · Issue #158 - GitHub
In the BMN Model Zoo there are results of feature extracted by MMAction but I found on details in Data Preparation about how...
Read more >
Weakly Supervised Violence Detection in Surveillance Video
First, they extracted video features using a VGG-f [35] model pretrained in ... [69] can extract bounding boxes within 0.0331 s per frame....
Read more >
(PDF) Spatio-Temporal Action Detection Under Large Motion
The TOI-Align operation extracts per-track features from the entire video sequence, ... per-frame features with a RoI-Align operation [16].
Read more >
Temporal Pyramid Network for Action Recognition
By sampling frames at different rates as input, backbone networks in [5, 35] are able to extract features of different receptive fields and...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found