$29
In HW2 problem 1, you will need to implement two semantic segmentation models and answer some questions in the report.
For more details, please click [this link](https://drive.google.com/file/d/1NLhvpKT8LrF-oOuFlz4sSuiX5AmUDw1Y/view?usp=sharing) to view the slides of HW2 and click [this link](https://drive.google.com/file/d/1W29SSyjGmUBmoq0dSW8qbHZ8xpdgLL6j/view?usp=sharing) to view the slides of HW2 problem 1.
Usage
To start working on this assignment, you should clone this repository into your local machine by using the following command.
git clone https://github.com/DLCV-Fall-2019/hw2-<username>.git
Note that you should replace `<username>` with your own GitHub username.
Dataset
In the starter code of this repository, we have provided a shell script for downloading and extracting the dataset for this assignment. For Linux users, simply use the following command.
bash ./get_dataset.sh
The shell script will automatically download the dataset and store the data in a folder called `hw2_data`. Note that this command by default only works on Linux. If you are using other operating systems or you can not download the dataset by running the command above, you should download the dataset from [this link](https://drive.google.com/file/d/1Lp3KS9Gh1LZx6_WVQsSd5H0iHmFAsmFn/view?usp=sharing) and unzip the compressed file manually.
> ⚠️ ***IMPORTANT NOTE*** ⚠️
> You should keep a copy of the dataset only in your local machine. **DO NOT** upload the dataset to this remote repository. If you extract the dataset manually, be sure to put them in a folder called `hw2_data` under the root directory of your local repository so that it will be included in the default `.gitignore` file.
Evaluation
To evaluate your model, you can run the provided evaluation script provided in the starter code by using the following command.
python3 mean_iou_evaluate.py <--pred PredictionDir> <--labels GroundTruthDir>
- `<PredictionDir>` should be the directory to your predicted semantic segmentation map (e.g. `hw2_data/prediction/`)
- `<GroundTruthDir>` should be the directory of ground truth (e.g. `hw2_data/val/seg/`)
Note that your predicted segmentation semantic map file should have the same filename as that of its corresponding ground truth label file (both of extension ``.png``).
Visualization
To visualization the ground truth or predicted semantic segmentation map in an image, you can run the provided visualization script provided in the starter code by using the following command.
python3 viz_mask.py <--img_path xxxx.png> <--seg_path xxxx.png>
You have a three-day delay quota for the whole semester. Once you have exceeded your quota, the credit of any late submission will be deducted by 30% each day.
Note that while it is possible to continue your work in this repository after the deadline, **we will by default grade your last commit before the deadline** specified above. If you wish to use your quota or submit an earlier version of your repository, please contact the TAs and let them know which commit to grade.
Academic Honesty
- Taking any unfair advantages over other class members (or letting anyone do so) is strictly prohibited. Violating university policy would result in an **F** grade for this course (**NOT** negotiable).
- If you refer to some parts of the public code, you are required to specify the references in your report (e.g. URL to GitHub repositories).
- You are encouraged to discuss homework assignments with your fellow class members, but you must complete the assignment by yourself. TAs will compare the similarity of everyone’s submission. Any form of cheating or plagiarism will not be tolerated and will also result in an **F** grade for students with such misconduct.
Submission Format
Aside from your own Python scripts and model files, you should make sure that your submission includes *at least* the following files in the root directory of this repository:
1. `hw2_<StudentID>.pdf`
The report of your homework assignment. Refer to the "*Submission*" section in the slides for what you should include in the report. Note that you should replace `<StudentID>` with your student ID, **NOT** your GitHub username.
2. `hw2.sh`
The shell script file for running your baseline model.
3. `hw2_best.sh`
The shell script file for running your improved model.
We will run your code in the following manner:
CUDA_VISIBLE_DEVICES=GPU_NUMBER bash hw2.sh $1 $2
CUDA_VISIBLE_DEVICES=GPU_NUMBER bash hw2_best.sh $1 $2
where `$1` is the testing images directory (e.g. `test/images/`), and `$2` is the output prediction directory (e.g. `test/labelTxt_hbb_pred/` ). Please do not create the output prediction directory in your bash script or python codes.
Packages
This homework should be done using python3.6. For a list of packages you are allowed to import in this assignment, please refer to the requirments.txt for more details.
You can run the following command to install all the packages listed in the requirements.txt:
pip3 install -r requirements.txt
Note that using packages with different versions will very likely lead to compatibility issues, so make sure that you install the correct version if one is specified above. E-mail or ask the TAs first if you want to import other packages.
Remarks
- If your model is larger than GitHub’s maximum capacity (100MB), you can upload your model to another cloud service (e.g. Dropbox). However, your shell script files should be able to download the model automatically. For a tutorial on how to do this using Dropbox, please click [this link](https://goo.gl/XvCaLR).
- **DO NOT** hard code any path in your file or script, and the execution time of your testing code should not exceed an allowed maximum of **10 minutes**.
- **Please refer to HW2 problem 1 slides for details about the penalty that may incur if we fail to run your code or reproduce your results.**