Fashionpedia is a new dataset which consists of two parts: (1) an ontology built by fashion experts containing 27 main apparel categories, 19 apparel parts, 294 fine-grained attributes and their relationships; (2) a dataset with 48k everyday and celebrity event fashion images annotated with segmentation masks and their associated per-mask fine-grained attributes, built upon the Fashionpedia ontology.
Check out our project page here and download the data at this link.
An illustration of the Fashionpedia dataset and ontology: (a) main garment masks; (b) garment part masks; (c) both main garment and garment part masks; (d) fine-grained apparel attributes; (e) an exploded view of the annotation diagram: the image is annotated with both instance segmentation masks (white boxes) and per-mask fine-grained attributes (black boxes); (f) visualization of the Fashionpedia ontology: we created Fashionpedia ontology and separate the concept of categories (yellow nodes) and attributes (blue nodes) in fashion. It covers pre-defined garment categories used by both Deepfashion2 and ModaNet. Mapping with DeepFashion2 also shows the versatility of using attributes and categories. We are able to present all 13 garment classes in DeepFashion2 with 11 main garment categories, 1 garment part, and 7 attributes.
With the introduction of the dataset, we explore the new task of instance segmentation with attribute localization. The proposed task requires both localizing an object and describing its properties, which unifies instance segmentation (detect and segment each object instance) and fine-grained visual attribute categorization (recognize one or multiple attributes).
This Fashionpedia API enables reading, and visualizing annotations, and evaluating results. Check out our two demos for more details.
# create a new environment:
python3 -m venv env # Create a virtual environment
source env/bin/activate # Activate virtual environment
# step 1: install COCO API:
# Note: COCO API requires numpy to install. Ensure that you have numpy installed.
# e.g. pip install numpy
pip install 'git+https://github.com/cocodataset/cocoapi.git#subdirectory=PythonAPI'
# step 2: install Fashionpedia API via pip
pip install fashionpedia
# step 3: test. You can test that you have correctly installed the fashionpedia api
# by running the following command inside the repo.
python test.py
# do your work ...
deactivate # exit
Clone the repo first and then do the following steps inside the repo:
# create a new environment:
python3 -m venv env # Create a virtual environment
source env/bin/activate # Activate virtual environment
# step 1: install COCO API:
# Note: COCO API requires numpy to install. Ensure that you have numpy installed.
# e.g. pip install numpy
pip install 'git+https://github.com/cocodataset/cocoapi.git#subdirectory=PythonAPI'
# step 2: install required packages
pip install -r requirements.txt
# step 3: test. You can test that you have correctly installed the fashionpedia api
# by running the following command inside the repo.
python test.py
# do your work here...
deactivate # exit
The evaluation code provide here can be used to measure results on the publicly available validation set. It computes multiple metrics describe below. We consider both localization constraint (IoU) and attributes prediction constraint (F1 score) when computing
Average Precision (AP)
AP@[ IoU=0.50:0.95 | F1=0.50:0.95 |area=all | maxDets=100] # Primary metric
AP@[ IoU=0.50 | F1=0.50:0.95 |area=all | maxDets=100]
AP@[ IoU=0.75 | F1=0.50:0.95 |area=all | maxDets=100]
AP@[ IoU=0.50:0.95 | F1=0.50 |area=all | maxDets=100]
AP@[ IoU=0.50:0.95 | F1=0.75 |area=all | maxDets=100]
AP@[ IoU=0.50 | F1=0.50 |area=all | maxDets=100]
AP@[ IoU=0.50 | F1=0.75 |area=all | maxDets=100]
AP@[ IoU=0.75 | F1=0.50 |area=all | maxDets=100]
AP@[ IoU=0.75 | F1=0.75 |area=all | maxDets=100]
AP@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= s | maxDets=100]
AP@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= m | maxDets=100]
AP@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= l | maxDets=100]
Average Recall (AR)
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area=all | maxDets= 1]
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area=all | maxDets= 10]
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area=all | maxDets=100]
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= s | maxDets=100]
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= m | maxDets=100]
AR@[ IoU=0.50:0.95 | F1=0.50:0.95 |area= l | maxDets=100]
Additionally, we also provide metrics with only IoU constraint and only F1 thresholds constraint, for better understanding of the algorithm. See evaluation demo for more details.
The results format is similar to COCO format for object detection with additional attribute_ids
filed. See evaluation demo and also loadRes()
in Fashionpedia API.
Detection with bounding boxes and localized attributes prediction:
[{
"image_id": int,
"category_id": int,
"attribute_ids": [int],
"bbox": [x,y,width,height],
"score": float,
}]
Detection with segmentation masks and localized attributes prediction:
[{
"image_id": int,
"category_id": int,
"attribute_ids": [int],
"segmentation": RLE,
"score": float,
}]
For the task of instance segmentation with attribute localization, we present a strong baseline model named Attribute-Mask R-CNN that is built upon Mask R-CNN for Fashionpedia. Check out our predictior demo and paper for more details.
The 1). pretrained model ckpts and 2). training code including data conversion, model training and inference here.
@inproceedings{jia2020fashionpedia,
title={Fashionpedia: Ontology, Segmentation, and an Attribute Localization Dataset},
author={Jia, Menglin and Shi, Mengyun and Sirotenko, Mikhail and Cui, Yin and Cardie, Claire and Hariharan, Bharath and Adam, Hartwig and Belongie, Serge}
booktitle={European Conference on Computer Vision (ECCV)},
year={2020}
}
Both fashionpedia and fashionpediaEval are sub-classes of COCO and COCOeval from PythonAPI for COCO. LVIS-API was also referenced during the development.