You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Nov 9, 2022. It is now read-only.
# optional: log in to W&B to see your training metrics,
17
-
# track your experiments, and submit your models to the benchmark
18
-
wandb login
19
-
# train Fastai model
20
-
python3 main.py
21
-
# train Keras model
22
-
python3 main_keras.py
23
16
24
-
```
25
-
26
-
### Training
17
+
# optional: log in to W&B to track your experiements
18
+
wandb login
27
19
28
-
To start training a model on a small sample dataset run the following, once working you should use the *full dataset* by changing `main.py`
20
+
# train a Keras model
21
+
python3 main_keras.py
29
22
30
-
```
31
-
pip3 install -r requirements.txt
23
+
# train a Fastai model
32
24
python3 main.py
33
25
```
34
26
35
-
This will download the sample dataset and begin training a model. You can monitor training performance on [Weights and Biases](https://www.wandb.com/). Once training is complete inference will be performed on all test scenes and a number of prediction images with names like `123123_ABCABC-prediction.png` will be created. After the the images are created they will be scored. Here's what the prediction looks like, not bad for 50 lines of code but there is a lot of room for improvement:
27
+
This will download the sample dataset and begin training a model. You can monitor training performance on [Weights and Biases](https://www.wandb.com/). Once training is complete inference will be performed on all test scenes and a number of prediction images with names like `123123_ABCABC-prediction.png` will be created in the `wandb` directory. After the the images are created they will be scored. Here's what a prediction looks like, not bad for 50 lines of code but there is a lot of room for improvement:
The *full dataset* can be downloaded by changing a line in `main.py` this is the dataset that should be used for benchmarking. The dataset comprises a number of aerial scenes captured from drones. Each scene has a ground resolution of 10 cm per pixel. For each scene there is a corresponding "image", "elevation" and "label". The image is an RGB tif, the elevation is a single channel floating point .tif and the label is a PNG with 7 colors representing the 7 classes. Please see `index.csv` - inside the downloaded dataset - for a description of the quality of each labelled image and the distribution of the labels. To use the dataset you can split it into smaller chips (see `images2chips.py`). Here is an example of one of the labelled scenes:
33
+
The dataset comprises a number of aerial scenes captured from drones. Each scene has a ground resolution of 10 cm per pixel. For each scene there is a corresponding "image", "elevation" and "label". These are located in the `images`, `elevation` and `labels` directory.
34
+
35
+
The images an RGB tifs, the elevations are single channel floating point tifs (where each pixel value represents elevation in meters) and finally the labels are PNGs with 7 colors representing the 7 classes (documented below)
36
+
37
+
In addition please see `index.csv` - inside the downloaded dataset folder - for a description of the quality of each labelled image and the distribution of the labels.
38
+
39
+
To use a dataset for training it need to first be converted to chips (see `images2chips.py`). This will created an `images-chips` and `label-chips` directory which contains a number of `300x300` (by default) RGB images. The `label-chips` are also RGB but will be very low pixel intensities `[0 .. 7]` so will appear black as first glance. You can use the `color2class` and `category2mask` function to switch between the two label representations.
0 commit comments