Sacha Jungerman1, Atul Ingle2, Sotiris Nousias3, Mian Wei4, Mel White5, Mohit Gupta1
1University of Wisconsin - Madison, 2Portland State University, 3Purdue University, 4Carnegie Mellon University, 5US Naval Research Lab
The reconstruction dataset (~425G training set + ~42G test set, ~133G and ~13G compressed respectively) consists of photoncube/image pairs from 50 unique simulated scenes, plus another 5 scenes for the test set (for which ground truths are not made public). Each photoncube consists of 1024 bitplanes, and the associated ground truth reconstruction corresponds to the last bitplane. A sample of this dataset can be downloaded here (~3.5GB).
The full dataset is hosted on a publicly accessible S3 bucket and compressed into ~8.5GB chunks. To download it, you'll need to use the aws-cli, and an LZMA-capable unzipping utility such as the 7zip cli to extract each zip file into its respective directory. Once installed, you can list out the dataset's components using:
$ aws s3 ls --summarize --human-readable --recursive s3://public-datasets/challenges/reconstruction --endpoint=https://web.s3.wisc.edu --no-sign-request
To download all the dataset archives to a predefined $DOWNLOAD_DIR in one shot, you can use the following command:
$ aws s3 sync s3://public-datasets/challenges/reconstruction $DOWNLOAD_DIR --endpoint=https://web.s3.wisc.edu --no-sign-request
You then need to extract all zips into their respective directories, which can be done with the following command in
bash:
$ for zip in $(find $DOWNLOAD_DIR -type f -name *.zip); do 7z x $zip -o$(dirname $zip) && rm -f $zip; done
For more single photon datasets, including real captures, see here.