or4cl3ai/Aiden_t5
Text Generation • Updated
• 692 • 20
image imagewidth (px) 84 478 | labels class label 15 classes |
|---|---|
11sitting | |
14using_laptop | |
7hugging | |
12sleeping | |
14using_laptop | |
12sleeping | |
4drinking | |
7hugging | |
1clapping | |
3dancing | |
2cycling | |
4drinking | |
1clapping | |
0calling | |
12sleeping | |
4drinking | |
0calling | |
8laughing | |
14using_laptop | |
14using_laptop | |
1clapping | |
5eating | |
6fighting | |
9listening_to_music | |
3dancing | |
4drinking | |
2cycling | |
8laughing | |
4drinking | |
3dancing | |
9listening_to_music | |
2cycling | |
11sitting | |
11sitting | |
12sleeping | |
5eating | |
10running | |
10running | |
7hugging | |
7hugging | |
0calling | |
14using_laptop | |
9listening_to_music | |
2cycling | |
12sleeping | |
7hugging | |
13texting | |
12sleeping | |
14using_laptop | |
10running | |
7hugging | |
0calling | |
14using_laptop | |
4drinking | |
7hugging | |
2cycling | |
8laughing | |
0calling | |
1clapping | |
8laughing | |
13texting | |
11sitting | |
8laughing | |
14using_laptop | |
12sleeping | |
9listening_to_music | |
2cycling | |
1clapping | |
13texting | |
13texting | |
5eating | |
2cycling | |
4drinking | |
0calling | |
10running | |
4drinking | |
13texting | |
11sitting | |
0calling | |
7hugging | |
5eating | |
12sleeping | |
14using_laptop | |
14using_laptop | |
10running | |
0calling | |
14using_laptop | |
14using_laptop | |
9listening_to_music | |
0calling | |
3dancing | |
8laughing | |
2cycling | |
1clapping | |
14using_laptop | |
0calling | |
3dancing | |
4drinking | |
13texting | |
6fighting |
A dataset from kaggle. origin: https://dphi.tech/challenges/data-sprint-76-human-activity-recognition/233/data
The data instances have the following fields:
image: A PIL.Image.Image object containing the image. Note that when accessing the image column: dataset[0]["image"] the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the "image" column, i.e. dataset[0]["image"] should always be preferred over dataset["image"][0].labels: an int classification label. All test data is labeled 0.{
'calling': 0,
'clapping': 1,
'cycling': 2,
'dancing': 3,
'drinking': 4,
'eating': 5,
'fighting': 6,
'hugging': 7,
'laughing': 8,
'listening_to_music': 9,
'running': 10,
'sitting': 11,
'sleeping': 12,
'texting': 13,
'using_laptop': 14
}
| train | test | |
|---|---|---|
| # of examples | 12600 | 5400 |
>>> from datasets import load_dataset
>>> ds = load_dataset("Bingsu/Human_Action_Recognition")
>>> ds
DatasetDict({
test: Dataset({
features: ['image', 'labels'],
num_rows: 5400
})
train: Dataset({
features: ['image', 'labels'],
num_rows: 12600
})
})
>>> ds["train"].features
{'image': Image(decode=True, id=None),
'labels': ClassLabel(num_classes=15, names=['calling', 'clapping', 'cycling', 'dancing', 'drinking', 'eating', 'fighting', 'hugging', 'laughing', 'listening_to_music', 'running', 'sitting', 'sleeping', 'texting', 'using_laptop'], id=None)}
>>> ds["train"][0]
{'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=240x160>,
'labels': 11}