公开数据集统一放在 root/datasets 路径下,大家可以通过复制/软链接进行调用哦!

数据集详情页涵盖所属标签以及数据集的基本说明。数据集格式等具体信息可点击数据集来源链接查看哦~

A Video Dataset of Atomic Visual Actions

数据集大小437.0 MB

更新时间2021-08-02 16:39:47

数据集标签
动作识别人体物体识别分类视频

数据集路径

数据集无需解压和下载,可直接在代码中更改数据集路径使用

/datasets/ava/

数据集简介

AVA is a project that provides audiovisual annotations of video for improving our understanding of human activity. AVA is a project that provides audiovisual annotations of video for improving our understanding of human activity. Each of the video clips has been exhaustively annotated by human annotators, and together they represent a rich variety of scenes, recording conditions, and expressions of human activity.

数据集说明

There are annotations for: Kinetics (AVA-Kinetics) - a crossover between AVA and Kinetics. In order to provide localized action labels on a wider variety of visual scenes, authors provide AVA action labels on videos from Kinetics-700, nearly doubling the number of total annotations, and increasing the number of unique videos by over 500x. Actions (AvA Actions) - the AVA dataset densely annotates 80 atomic visual actions in 430 15-minute movie clips, where actions are localized in space and time, resulting in 1.62M action labels with multiple labels per human occurring frequently. Spoken Activity (AVA ActiveSpeaker, AVA Speech). AVA ActiveSpeaker: associates speaking activity with a visible face, on the AVA v1.0 videos, resulting in 3.65 million frames labeled across ~39K face tracks. AVA Speech densely annotates audio-based speech activity in AVA v1.0 videos, and explicitly labels 3 background noise conditions, resulting in ~46K labeled segments spanning 45 hours of data.