300W-Style and AFLW-Style are created based on 300W and AFLW, respectively. They are introduced in Style Aggregated Network for Facial Landmark Detection. In a nutshell, I applied three different style changes (light, gray, sketch) to the original facial images of 300W and AFLW. Together with the original images, 300W-Style and AFLW-Style have four variants of each image in the original 300W and AFLW. In the same time, each variant share the same annotation for facial landmarks. They are used to facilitate the analysis for facial landmark detection problem.
- Download 300W-Style and AFLW-Style from Google Drive or Baidu Yun, and extract the downloaded files into
~/datasets/
. Since the tgz files for 300W-Style and AFLW-Style are too large to download, I split them into several small parts in the300W-Style-Splits
folder and theAFLW-Style-Splits
folder on Google Drive. You can download all parts and usecat 300W-Style.tgz.parta* > 300W-Style.tgz
andcat AFLW-Style.tgz.parta* > AFLW-Style.tgz
to get the original tgz file. - In 300W-Style and AFLW-Style directories, the
Original
sub-directory contains the original images from 300-W and AFLW - The sketch, light, and gray style images are used to analyze the image style variance in facial landmark detection.
- For simplification, we change some file names, such as removing the space or unifying the file extension.
In cluster.py
, we use ResNet-152 to classify the 4 style class of 300W-Style and AFLW-Style, it obtains 93.5% accuracy (training on 300W-Style and evaluating on AFLW-Style), and 96.6% accuracy (training on AFLW-Style and evaluating on 300W-Style).
If you can't download the datasets via the above link, please try:
300W-Style : https://drive.google.com/open?id=1wy9ZUSWE4V2WdMbXS3Jkq7kQEYHSpLk1
AFLW-Style : https://drive.google.com/open?id=1y5JrOd86NGHTPZLYLgeqRJzT2T4ACWJe
300W-Style.tgz
should be extracted into ~/datasets/300W-Style
by typing tar xzvf 300W-Style.tgz; mv 300W-Convert 300W-Style
.
It has the following structure:
--300W-Gray
--300W ; afw ; helen ; ibug ; lfpw
--300W-Light
--300W ; afw ; helen ; ibug ; lfpw
--300W-Sketch
--300W ; afw ; helen ; ibug ; lfpw
--300W-Original
--300W ; afw ; helen ; ibug ; lfpw
--Bounding_Boxes
--*.mat
AFLW-Style.tgz
should be extracted into ~/datasets/AFLW-Style
by typing tar xzvf AFLW-Style.tgz; mv AFLW-Convert AFLW-Style
.
It has the following structure (annotation
is generated by python aflw_from_mat.py
):
--aflw-Gray
--0 2 3
--aflw-Light
--0 2 3
--aflw-Sketch
--0 2 3
--aflw-Original
--0 2 3
--annotation
--0 2 3
If these two datasets help your research, please cite the following papers:
@inproceedings{dong2018san,
title={Style Aggregated Network for Facial Landmark Detection},
author={Dong, Xuanyi and Yan, Yan and Ouyang, Wanli and Yang, Yi},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
pages={379--388},
doi={10.1109/CVPR.2018.00047},
year={2018}
}