FedVison dataset is created jointly by WeBank and ExtremeVision to facilitate the advancement of academic research and industrial applications of federated learning.
- Provides images data sets with standardized annotation for federated object detection.
- Provides key statistics and systems metrics of the data sets.
- Provides a set of implementations of baseline for further research.
We introduce two realistic federated datasets.
- Federated Street, a real-world object detection dataset that annotates images captured by a set of street cameras based on object present in them, including 7 classes. In this dataset, each or every few cameras serve as a device.
| Dataset | Number of devices | Total samples | Number of class |
|---|---|---|---|
| Federated Street | 5, 20 | 956 | 7 |
- Street_Dataset.tar contains the image data and ground truth for the train and test set of the street data set.
- Images: The directory which contains the train and test image data.
- train_label.json: The annotations file is saved in json format. train_label.json is a
list, which contains the annotation information of the Images set. The length oflistis the same as the number of image and each value in thelistrepresents one image_info. Eachimage_infois in format ofdictionarywith keys and values. The keys ofimage_infoareimage_id,device1_id,device2_idanditems. We split the street data set in two ways. For the first, we split the data into 5 parts according to the geographic information. Besides, we turn 5 into 20. Therefore we havedevice1_idanddevice2_id. It means that we have 5 or 20 devices.itemsis a list, which may contain multiple objects.
[
{
"image_id": the id of the train image, for example 009579.
"device1_id": the id of device1 ,specifies which device the image is on.
"device2_id": the id of device2.
"items": [
{
"class": the class of one object,
"bbox": ["xmin", "ymin", "xmax", "ymax"], the coordinates of a bounding box
},
...
]
},
...
] - test_label.json: The annotations of test data are almost the same as of the train_label.json. The only difference between them is that
the
image_infoof test data does not have the keydevice_id.
We use he standard PASCAL VOC 2010 mean Average Precision (mAP) for evaluation (mean is taken over per-class APs).
To be considered a correct detection, the overlap ratio 


when denotes the intersection of the predicted and ground truth bounding boxes and
their union.
Average Precision is calculated for each class respectively.
where n is the number of total object in given class.
For