Skip to content

Latest commit

 

History

History
 
 

Instance_Matching

Instance Matching Module

This directory hosts the code and dataset for Instance Matching Module in the SketchyScene Colorization system.

Requirements

  • Python 3
  • Tensorflow (>= 1.3.0)
  • scipy
  • PIL

Preparations

  • Please follow the instructions here for MATCHING dataset preparation.
  • If you want to train the network, please follow the the instructions here for pre-trained model preparation.

Training

After the preparations, run:

python3 matching_main.py --mode 'train' --data_base_dir 'path/to/SketchyScene-dataset'
  • Set --model to 'deeplab_v3plus', 'fcn_8s' or 'segnet' for other alternatives. The default is deeplab (Deeplab-v2 model).

Evaluation

Please follow these steps for the evaluation:

  1. Please make sure the trained model has been placed under outputs/snapshots. We have also provided our trained model, which can be downloaded here.

  2. Our Instance Matching Module consists of Mask R-CNN for instance segmentation. Since the evaluation will be performed on a sketch image for multiple times (several testing sentences), it is recommended to generate the segmented data first and reuse it to avoid the redundant time caused by the repeated forwarding procedure. To do so, please go to the SketchyScene repository for some preparations:

    1. Strictly follow the instructions in the Instance-Segmentation part to setup the Mask R-CNN model.

    2. After setuping the Instance Segmentation model, place the two json files data/sentence_instance_val.json and data/sentence_instance_test.json of MATCHING dataset to SketchyScene/Instance_Segmentation/data/, and then run the following command under SketchyScene/Instance_Segmentation to generate the segmented data:

      python3 segment_data_generation.py --data_basedir 'path/to/SketchyScene-dataset'
      
      • Set --use_edgelist to 0 if you don't want to use edgelist. The default is 1 and you must strictly follow the instructions to prepare for the edgelist data.
    3. Finally you will find the segmented data under outputs/inst_segm_output_data. Then place them to SketchySceneColorization/Instance_Matching/outputs/inst_segm_output_data/test(val)/seg_data.

  3. After these, run:

    python3 matching_main.py --mode 'eval' --dataset 'val' --data_base_dir 'path/to/SketchyScene-dataset'
    
    • Set --dataset to val/test.
    • Set --visualized to 1 if you want to see visual results during evaluation. The default is 0.

Inference

Here you can select an sketch image and input any instructions to see the visual results. Please firstly make sure the trained model and the segmented data have been placed correctly as in the Evaluation part. Then run:

python3 matching_main.py --mode 'inference' --data_base_dir 'path/to/SketchyScene-dataset' \
                         --dataset 'val' --image_id 2 --instruction 'your instruction'
  • Set --dataset to val/test.
  • Set --image_id to the image you want.
  • Input your instruction after --instruction.

🔥 We have provided a test case for example: a sketch image and its output segmentation data are placed under examples directory. You can try it:

python3 matching_main.py --mode 'inference' --data_base_dir 'examples' --seg_data_dir 'examples' \
                         --dataset 'test' --image_id 30 \
                         --instruction 'all the grasses are dark green'
  • Try other instructions for multiple times by changing the text all the grasses are dark green :)

If everything goes well, you can obtain the result like this: example

Citation

Please cite the corresponding paper if you found the datasets or code useful:

@article{zouSA2019sketchcolorization,
  title = {Language-based Colorization of Scene Sketches},
  author = {Zou, Changqing and Mo, Haoran and Gao, Chengying and Du, Ruofei and Fu, Hongbo},
  journal = {ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH Asia 2019)},
  year = {2019},
  volume = 38,
  number = 6,
  pages = {233:1--233:16}
}