parquet-converter commited on
Commit
2bd6018
1 Parent(s): c59dbb8

Update parquet files

Browse files
README.dataset.txt DELETED
@@ -1,12 +0,0 @@
1
- # undefined > raw-images
2
- https://public.roboflow.ai/object-detection/undefined
3
-
4
- Provided by undefined
5
- License: CC BY 4.0
6
-
7
- ## About this Dataset
8
- This dataset was created by exporting images from [images.cv](https://images.cv/dataset/forklift-image-classification-dataset) and labeling them as an object detection dataset. **The dataset contains 421 raw images (v1 - raw-images) and labeled classes include:**
9
- * forklift
10
- * person
11
-
12
- ![Example annotated image from the dataset from the dataset](https://i.imgur.com/a6hWEG4.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,49 +0,0 @@
1
- ---
2
- task_categories:
3
- - object-detection
4
- tags:
5
- - roboflow
6
- ---
7
-
8
- ### Roboflow Dataset Page
9
- https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv/dataset/1
10
-
11
- ### Dataset Labels
12
-
13
- ```
14
- ['forklift', 'person']
15
- ```
16
-
17
- ### Citation
18
-
19
- ```
20
- @misc{ forklift-dsitv_dataset,
21
- title = { Forklift Dataset },
22
- type = { Open Source Dataset },
23
- author = { Mohamed Traore },
24
- howpublished = { \\url{ https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv } },
25
- url = { https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv },
26
- journal = { Roboflow Universe },
27
- publisher = { Roboflow },
28
- year = { 2022 },
29
- month = { mar },
30
- note = { visited on 2023-01-01 },
31
- }
32
- ```
33
-
34
- ### License
35
- CC BY 4.0
36
-
37
- ### Dataset Summary
38
- This dataset was exported via roboflow.ai on April 3, 2022 at 9:01 PM GMT
39
-
40
- It includes 421 images.
41
- Forklift are annotated in COCO format.
42
-
43
- The following pre-processing was applied to each image:
44
- * Auto-orientation of pixel data (with EXIF-orientation stripping)
45
-
46
- No image augmentation techniques were applied.
47
-
48
-
49
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.roboflow.txt DELETED
@@ -1,15 +0,0 @@
1
-
2
- Forklift - v1 raw-images
3
- ==============================
4
-
5
- This dataset was exported via roboflow.ai on April 3, 2022 at 9:01 PM GMT
6
-
7
- It includes 421 images.
8
- Forklift are annotated in COCO format.
9
-
10
- The following pre-processing was applied to each image:
11
- * Auto-orientation of pixel data (with EXIF-orientation stripping)
12
-
13
- No image augmentation techniques were applied.
14
-
15
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/valid.zip → default/forklift-object-detection-test.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:881c60bacab0a2eed1add7e9dfa12eeb31a3afcbeca4dc15096e35e3ac1e4a86
3
- size 3774165
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d3f4a1e260d035988b7e845e0166c470f36a5bab8adb37299b11ad783c215b
3
+ size 2912705
data/train.zip → default/forklift-object-detection-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d59c69ecf95a6520485444a2a61c349f24618ffa0b54ffee5f9422e08ac0f30d
3
- size 13533922
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eb3141cd7d571aeb5cb644aaa33b4dc79ca4bde6394495f3f7da17b0edeed89
3
+ size 14172165
data/test.zip → default/forklift-object-detection-validation.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1a52eeef7da48be78556a1dd9424460c51cb8737176b31c01b56bf92b1c7395
3
- size 2771676
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d3ec4e3297afff899e497a775d6413b163eda412382835eb54af4665d95dbf0
3
+ size 3950957
forklift-object-detection.py DELETED
@@ -1,121 +0,0 @@
1
- import collections
2
- import json
3
- import os
4
-
5
- import datasets
6
-
7
-
8
- _HOMEPAGE = "https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv/dataset/1"
9
- _LICENSE = "CC BY 4.0"
10
- _CITATION = """\
11
- @misc{ forklift-dsitv_dataset,
12
- title = { Forklift Dataset },
13
- type = { Open Source Dataset },
14
- author = { Mohamed Traore },
15
- howpublished = { \\url{ https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv } },
16
- url = { https://universe.roboflow.com/mohamed-traore-2ekkp/forklift-dsitv },
17
- journal = { Roboflow Universe },
18
- publisher = { Roboflow },
19
- year = { 2022 },
20
- month = { mar },
21
- note = { visited on 2023-01-01 },
22
- }
23
- """
24
- _URLS = {
25
- "train": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/train.zip",
26
- "validation": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/valid.zip",
27
- "test": "https://huggingface.co/datasets/keremberke/forklift-object-detection/resolve/main/data/test.zip",
28
- }
29
-
30
- _CATEGORIES = ['forklift', 'person']
31
- _ANNOTATION_FILENAME = "_annotations.coco.json"
32
-
33
-
34
- class FORKLIFTOBJECTDETECTION(datasets.GeneratorBasedBuilder):
35
- VERSION = datasets.Version("1.0.0")
36
-
37
- def _info(self):
38
- features = datasets.Features(
39
- {
40
- "image_id": datasets.Value("int64"),
41
- "image": datasets.Image(),
42
- "width": datasets.Value("int32"),
43
- "height": datasets.Value("int32"),
44
- "objects": datasets.Sequence(
45
- {
46
- "id": datasets.Value("int64"),
47
- "area": datasets.Value("int64"),
48
- "bbox": datasets.Sequence(datasets.Value("float32"), length=4),
49
- "category": datasets.ClassLabel(names=_CATEGORIES),
50
- }
51
- ),
52
- }
53
- )
54
- return datasets.DatasetInfo(
55
- features=features,
56
- homepage=_HOMEPAGE,
57
- citation=_CITATION,
58
- license=_LICENSE,
59
- )
60
-
61
- def _split_generators(self, dl_manager):
62
- data_files = dl_manager.download_and_extract(_URLS)
63
- return [
64
- datasets.SplitGenerator(
65
- name=datasets.Split.TRAIN,
66
- gen_kwargs={
67
- "folder_dir": data_files["train"],
68
- },
69
- ),
70
- datasets.SplitGenerator(
71
- name=datasets.Split.VALIDATION,
72
- gen_kwargs={
73
- "folder_dir": data_files["validation"],
74
- },
75
- ),
76
- datasets.SplitGenerator(
77
- name=datasets.Split.TEST,
78
- gen_kwargs={
79
- "folder_dir": data_files["test"],
80
- },
81
- ),
82
- ]
83
-
84
- def _generate_examples(self, folder_dir):
85
- def process_annot(annot, category_id_to_category):
86
- return {
87
- "id": annot["id"],
88
- "area": annot["area"],
89
- "bbox": annot["bbox"],
90
- "category": category_id_to_category[annot["category_id"]],
91
- }
92
-
93
- image_id_to_image = {}
94
- idx = 0
95
-
96
- annotation_filepath = os.path.join(folder_dir, _ANNOTATION_FILENAME)
97
- with open(annotation_filepath, "r") as f:
98
- annotations = json.load(f)
99
- category_id_to_category = {category["id"]: category["name"] for category in annotations["categories"]}
100
- image_id_to_annotations = collections.defaultdict(list)
101
- for annot in annotations["annotations"]:
102
- image_id_to_annotations[annot["image_id"]].append(annot)
103
- image_id_to_image = {annot["file_name"]: annot for annot in annotations["images"]}
104
-
105
- for filename in os.listdir(folder_dir):
106
- filepath = os.path.join(folder_dir, filename)
107
- if filename in image_id_to_image:
108
- image = image_id_to_image[filename]
109
- objects = [
110
- process_annot(annot, category_id_to_category) for annot in image_id_to_annotations[image["id"]]
111
- ]
112
- with open(filepath, "rb") as f:
113
- image_bytes = f.read()
114
- yield idx, {
115
- "image_id": image["id"],
116
- "image": {"path": filepath, "bytes": image_bytes},
117
- "width": image["width"],
118
- "height": image["height"],
119
- "objects": objects,
120
- }
121
- idx += 1