This repository contains all the files and code necessary to reproduce my winning solution to the pothole detection problem given at the MIIA - Deep Learning Hackathon. I hope this can be helpful to the participants new to the field of Machine/Deep Learning (or anyone else).
Note, since this was a hackathon, we had <10 hours to solve the problem. I'm sure there is still plenty of room for improvement and wouldn't be surprised if the code contains some bugs. I'll leave that to you to discover. Feel free to contact me if you have any questions and/or suggestions ([email protected]).
I strongly recommend the cloud service, Crestle. It takes 3 minutes to create an account and two clicks to open up a Jupyter Notebook with all the necessary libraries pre-installed and access to a powerful GPU.
Once your server is set up, clone this repo with the terminal command
git clone https://github.com/jandremarais/PotholeDetection.git
Since the data is too big to save in a Github repo, you need to download it from this Google Drive link. If using a cloud server, you can download the data with this command (thanks to this SO answer)
python gddown.py 0B1IZ6xxwxyvTdkV0dm9hWm9pX0k train.zip
and then extract files with jar -xf train.zip
. You can do the same for the test files, python gddown.py 0B1IZ6xxwxyvTSEQzaVVkYXNKdzQ test.zip
Note the directory structure:
data
├── test
│ └── unknown
│ └── QCiNkpJJqhUrhJw.JPG
├── train
│ ├── negative
│ │ └── QBvfRmpwrRbuhxE.JPG
│ └── positive
│ └── QBJSjQDEIQPuaxb.JPG
└── valid
├── negative
│ └── QbKxDKiHtsYBtZq.JPG
└── positive
└── QPJVBLDFqTEnvrc.JPG
The images inside the folders of this repo are only samples and should be deleted before downloading the data.
data_crop
follows the same structure. Its contents will be created by a code cell in the solution.ipynb
notebook.
Edit: Since the labels of the test images are now made available, the test folder from the Google Drive link also have 'positive' and 'negative' subdirectories.
The folder custom_layers
and the files densenet121.py
and resnet_101.py
are from this awesome repo. It contains code to create common ConvNet architectures in Keras, including pretrained ImageNet weights. See the README.md
of the original repo to get access to the pretrained weights and save them in the models
directory (again using the gddown.py
script). I edited the model scripts for compatibility with Keras 2.0.
Now you are ready to run the code in the solution.ipynb
notebook. I've included some explanations and comments in-between the code cells to increase readability.