Professional Documents
Culture Documents
Where's Wally
Where's Wally
1
5. Finally, we train a linear regression model to generate B. Generative Adversarial Networks(GAN) model
more accurate bounding boxes for each recognized
object. In this project, we used the GAN model to generate a new
Wally image. Generative adversarial networks (GANs) are
an emerging technique for both semi-supervised and
unsupervised learning was proposed by Ian Goodfellow in
2014.[1]
They can be characterized by training a pair of networks
in competition with each other. The generator G creates
forgeries, and the discriminator D receives both forgeries and
real images and aims to tell them apart (see Fig. 9). Both are
trained simultaneously and in competition with each other.
Fig. 4. Input image
Fig. 5. Selectively search for regions of interest, input these regions into the
CNN, and go through the convolutional network
Fig. 6. CNN extracts features for each region and uses SVM to classify
these regions into different categories
The evaluation criterion is that D(x) is as large as possible,
the closer to 1, the better, and D(G(z)) is as small as
possible. [5]
2
B. Model design, preparation and training V. RESULT
We used the training data with the label for model After 3000 epochs, the result of the loss function is
training and combined the object recognition and test data for reasonable (see Fig. 12), and we tested some images and got
the model of finding Wally. Our model architecture(see Fig. a significant identification. (see Fig. 13) In the demo image
11) as follows: of finding Wally, we got 98~100% accuracy, which is a good
result for the characters identified in the illustrations using
RCNN model.
Object Recognition
Classes number:90
Kernel size: 3
Regularization weight: 0.00004
Feature extractor
depth: 16
activation: RELU
decay: 0.9997
Hard example miner
Num hard examples: 3000
Loss type: CLASSIFICATION
Max negatives per positive: 3
Min negatives per image: 0
Training
Batch size: 24
Optimizer:RMSprop
Momentum optimizer value: 0.9
decay: 0.9
Exponential decay learning rate
Initial learning rate: 0.004
Decay steps: 800720
Decay factor: 0.95
Fig. 13. Finding Wally
3
[3] Q. Xu, G. Huang, Y. Yuan, C. Guo, Y. Sun, F. Wu, K. Weinberger.”
An empirical study on evaluation metrics of generative adversarial
We used the dataset and GAN model to generate a new wally networks”.arXiv preprint arXiv:1806.07755
image, but the result is not qualified without enough dataset. [4] M.T. Rosenstein, Z. Marx, L.P. Kaelbling, and T.G. Dietterich, “To
(see Fig. 14) However, we can find that the GAN-generated transfer or not to transfer,” In NIPS05 Workshop, Inductive Transfer:
10 Years Later, 2005.
images are close to the original image in terms of
composition and color. [5] S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE
Transactions on knowledge and data engineering, vol. 22, no. 10, pp.
1345–1359, 2010.
[6] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I.
Goodfellow, and R. Fergus, “Intriguing properties of neural
networks,” arXiv preprint arXiv:1312.6199, 2013.
Fig. 14. GAN result
[7] Hossein Hosseini, Baicen Xiao, Mayoore Jaiswal, and Radha
Poovendran. “On the Limitation of Convolutional Neural Networks in
Recognizing Negative Images”. In: Machine Learning and
VI. CONCLUSION Applications (ICMLA), 2017 16th IEEE International Conference on.
IEEE. 2017, pp. 352–358.
Revised from
https://diglib.eg.org/bitstream/handle/10.2312/cgvc20211313/027-
031.pdf?sequence=1&isAllowed=y
Label src:
https://github.com/tadejmagajna/HereIsWally/blob/master/trained_mo
del/labels.txt
Graph src: https://arxiv.org/pdf/1311.2524.pdf
https://github.com/tensorflow/models/edit/master/research/
object_detection/samples/configs/ssd_inception_v2_coco.config
REFERENCES
[1] I. Goodfellow et al. "Generative adversarial nets." Advances in neural
information processing systems. 2014.
[2] A. Creswell, T. White, V. Dumoulin, K. Arulkumaran, B. Sengupta et
al.” Generative adversarial networks: An overview”.IEEE Signal
Processing Magazine 35 (1), 53-65.