36
Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015) Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015) 味曽野雅史 (weblab B4) 2015/10/15 2015/10/15 味曽野雅史 (weblab B4) 1

論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

  • Upload
    mmisono

  • View
    2.221

  • Download
    1

Embed Size (px)

Citation preview

Page 1: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Deep Neural Networks are Easily Fooled: High ConfidencePredictions for Unrecognizable Images (Anh Nguyen, Jason

Yosinski, Jeff Clune; CVPR 2015)

味曽野雅史 (weblab B4)

2015/10/15

2015/10/15 味曽野雅史 (weblab B4) 1

Page 2: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

紹介する論文

• Deep Neural Networks are Easily Fooled: High Confidence Predictions forUnrecognizable Images• Anh Nguyen, Jason Yosinski, Jeff Clune• University of Wyoming, Cornell University• Computer Vision and Pattern Recognition (CVPR ’15), IEEE, 2015.

• http://www.evolvingai.org/fooling• 被引用: 39件 (Google Scholar)• 画像識別において,人間には理解できなくても DNNが 99.99%の確信度を持つような画像を作成することができる; 進化的アルゴリズムにより,画像を生成

2015/10/15 味曽野雅史 (weblab B4) 2

Page 3: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

画像例

Figure 1: ImageNet で訓練された DNN で確信度 99.6%以上となる画像 (本論文より)

2015/10/15 味曽野雅史 (weblab B4) 3

Page 4: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

背景

• DNNの中身はブラックボックス的に扱われることが多い• DNNを可視化することにより,何かしらの知見を得たい

2015/10/15 味曽野雅史 (weblab B4) 4

Page 5: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

関連研究 : (C)NNの可視化

• 第一層のフィルタを可視化• 高次元の表現を可視化• 各層ごとにどういう計算がおこなわれるのかを可視化• J.Yosinski et al. (2014)• Mahendran & Vadaldi (2014)

• deconvolutionによる可視化• Zeiler & Fergus (2013)

• maximize activation• Erhan et al. (2009)• K.Simonyan et al. (2013)• C.Szegedy et al. (2013)• I.J.Goodfellow et al. (2014)• Deep Dream (2015)• J.Yosinski et al. (2015)

2015/10/15 味曽野雅史 (weblab B4) 5

Page 6: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

第一層のフィルタを可視化

• 例えばガボールフィルタに近いものが学習される例が確認されている• Conv -> Pool -> FC -> FC -> Softmax の NN (MNIST) の第一層のフィルタ (http://eric-yuan.me/cnn/)

2015/10/15 味曽野雅史 (weblab B4) 6

Page 7: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

高次元の表現を可視化

• http://colah.github.io/posts/2015-01-Visualizing-Representations/

2015/10/15 味曽野雅史 (weblab B4) 7

Page 8: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Deconvolutionによる可視化• ある入力画像において,どの部分が活性化に寄与したかを可視化する

(Zeiler & Fergus)

2015/10/15 味曽野雅史 (weblab B4) 8

Page 9: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Deconvolutionによる可視化

Figure 2: Zeiler & Fergus (2013) より2015/10/15 味曽野雅史 (weblab B4) 9

Page 10: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Maximize Activation

• 各層において,各ユニットの出力の和が最大になるような画像を生成する• ランダムなノイズパターン x = x0 からスタート

• 活性化関数 ai(x)に対して x← x + η ∂ai (x)∂x により最適化

• このとき,適当な正則化項を用いないと,人間が見てよく分からない画像が生成される

2015/10/15 味曽野雅史 (weblab B4) 10

Page 11: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

本論文の例

• 正則化項なし (99.99%の確信度)

2015/10/15 味曽野雅史 (weblab B4) 11

Page 12: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

K.Simonayn et al. (2013)

• L2正則化を利用

2015/10/15 味曽野雅史 (weblab B4) 12

Page 13: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Deep Dream

• 隣接するピクセル同士が相関するような制約を入れている

2015/10/15 味曽野雅史 (weblab B4) 13

Page 14: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

J.Yosinski et al. (2015)

• 4つの異なる正則化方法を提案

2015/10/15 味曽野雅史 (weblab B4) 14

Page 15: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

関連研究 : DNNと人間の視覚の違い

• 正しく分類されている画像を,少し変更するだけで誤認識するように変更できる

• C.Szegedy et al. (2013)• I.J.Goodfellow et al. (2014)

2015/10/15 味曽野雅史 (weblab B4) 15

Page 16: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

C.Szegedy et al. (2013)

• 正しく識別出来てる画像を少し歪ませることで,全く別のものとして識別させる

2015/10/15 味曽野雅史 (weblab B4) 16

Page 17: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Goodfellow et al. (2014)

Figure 3:

2015/10/15 味曽野雅史 (weblab B4) 17

Page 18: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

どう画像を変化させているか?

• 通常の学習の場合 → 誤差逆伝播,つまり誤識別した場合パラメータを調整する

• この時,パラメータを調整せず,逆に画像に少し変更を加えたら..?

2015/10/15 味曽野雅史 (weblab B4) 18

Page 19: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

この論文の内容

-EAで画像を生成したところ,よく分からないのができた,というのが実際の経緯っぽい

2015/10/15 味曽野雅史 (weblab B4) 19

Page 20: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

この論文で使われるネットワークモデル

• AlexNet (A.Krizhevsky et al. 2012)• LeNet (Y.LeCun et al. 1998)• 既に訓練済みのものを利用 (caffe版)

2015/10/15 味曽野雅史 (weblab B4) 20

Page 21: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

画像の生成手法

• 進化的アルゴリズム (Evolutionary Algorithm)• MAP-Elites (multi-dimensional archive of phenotypic elites)• A.Cully et al. 2015• ベストな個体を記憶• 各繰り返しステップにおいて,• 適当に個体を選び,適当に突然変異• もし現在のベストな個体よりも良ければ,それと交換

• 今回の場合,ベストな個体 = 確信度が高い• direct encoding• 遺伝子として画像をそのまま表現• 例) MNISTなら 28x28のピクセル• 各ピクセルは一様乱数で初期化

• indirect encoding• Compositional Pattern-Producing Network (CPPN)• K.O.Stanley (2007)• ピクセル位置 (x,y)を受け取り,色値を出力する NNのようなもの

2015/10/15 味曽野雅史 (weblab B4) 21

Page 22: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

LeNet (MNIST)

Figure 4: direct encoding: 確信度 99.99%

2015/10/15 味曽野雅史 (weblab B4) 22

Page 23: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

LeNet (MNIST)

Figure 5: indirect encoding: 確信度 99.99%

2015/10/15 味曽野雅史 (weblab B4) 23

Page 24: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

AlexNet (ImageNet)

• direct encodingにおける世代数と確信度の変化• direct encodingはあまりうまくいかず

2015/10/15 味曽野雅史 (weblab B4) 24

Page 25: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

AlexNet (ImageNet)

• indirect encodingにおける世代数と確信度の変化

2015/10/15 味曽野雅史 (weblab B4) 25

Page 26: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

indirect encodingによって生成された画像

Figure 6: 平均確信度 99.12%2015/10/15 味曽野雅史 (weblab B4) 26

Page 27: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

fooling imageを入れた学習

• DNN1 で学習 (nクラス)• DNN1 を使ってfooling imageを作成• DNN2 で fooling imageを入れた (n+1)クラスの分類問題を学習• これを繰り返す

2015/10/15 味曽野雅史 (weblab B4) 27

Page 28: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

fooling imageを入れた学習: ImageNetでの結果• DNN2 で作成したfooling imageの確信度は DNN1 のときと比べ,大きく下

がった (平均確信度 88.1%から 11.7%)• ただし,MNISTの場合は確信度はそれほど下がらなかった

2015/10/15 味曽野雅史 (weblab B4) 28

Page 29: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

生成された画像に対する考察

• ユニークな特徴を強調するように進化すれば良い• (人間には知覚できない)少しの変化で識別するクラスが変えられることを考えると,多様性が生まれているに驚きがある

2015/10/15 味曽野雅史 (weblab B4) 29

Page 30: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

系統的な画像の生成

• 同じ種類 (犬など)は,同じようなものが進化で得られる

2015/10/15 味曽野雅史 (weblab B4) 30

Page 31: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

繰り返し模様を除去した場合

• indirect encodingで得られる結果には,繰り返し模様が多い• 繰り返しを除去した場合 (下図): 繰り返しが確信度に寄与することが確認できる

2015/10/15 味曽野雅史 (weblab B4) 31

Page 32: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

仮説: 生成モデルと識別モデル• 今回の DNNのような識別モデル p(y |X)の場合,入力 X に対して,クラス y を決める境界面を学習する• 入力空間が高次元の場合,あるクラスと認識する空間は訓練データの規模よりも大きくなる

• (仮説) 進化的計算で生成された画像は,境界面よりも遠く離れていた• 生成モデル p(y , X)であれば,p(y |X)の他に p(X)も必要.今回のような問題はおきないかもしれない

2015/10/15 味曽野雅史 (weblab B4) 32

Page 33: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

まとめ

• 進化的計算によりあるクラスと 99%の確信度で識別される画像を生成すると,人間には判断がつかないものが生成される

• 進化的計算の結果得られる画像として考えられるもの• どのクラスに対応する画像も似たようなものが生成される

(C.Szegedyらの実験結果より)• 人間に認識できる画像が生成される

• 実験してみると,上の二つのどちらでもなかった• もちろん筆者が意図的に認識できないような画像を生成したわけではない

• DNNの実世界への応用を考えると,セキュリティ的な問題がある• 防犯カメラ,自動運転,音声認識,画像検索エンジン,…

• CPPN EAによる画像生成は,DNN可視化の新規手法ともいえる

2015/10/15 味曽野雅史 (weblab B4) 33

Page 34: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Deep Visualization Toolbox• http://yosinski.com/deepvis• https://github.com/yosinski/deep-visualization-toolbox

2015/10/15 味曽野雅史 (weblab B4) 34

Page 35: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Reference

• Yosinski, J., Clune, J., Bengio, Y., and Lipson, H. How transferable arefeatures in deep neural networks? In Ghahramani, Z., Welling, M., Cortes,C., Lawrence, N.D., and Weinberger, K.Q. (eds.), Advances in NeuralInformation Processing Systems 27, pp. 3320–3328. Curran Associates, Inc.,Decembe 2014.• Mahendran, A. and Vedaldi, A. Understanding Deep Image Representations

by Inverting Them. ArXiv e-prints, November 2014.• Zeiler, Matthew D and Fergus, Rob. Visualizing and understanding

convolutional neural networks. arXiv preprint arXiv:1311.2901, 2013.• Erhan, Dumitru, Bengio, Yoshua, Courville, Aaron, and Vincent, Pascal.

Visualizing higher-layer features of a deep network. Technical report,Technical report, University of Montreal, 2009.• K. Simonyan, A. Vedaldi, and A. Zisserman. Deep inside convolutional

networks: Visualising image classification models and saliency maps. arXivpreprint arXiv:1312.6034, 2013

2015/10/15 味曽野雅史 (weblab B4) 35

Page 36: 論文輪読: Deep neural networks are easily fooled: High confidence predictions for unrecognizable images

Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images (Anh Nguyen, Jason Yosinski, Jeff Clune; CVPR 2015)

Reference

• K. O. Stanley and R. Miikkulainen. A taxonomy for artificial embryogeny.Artificial Life, 9(2):93–130, 2003• A. Cully, J. Clune, and J.-B. Mouret. Robots that can adapt like natural

animals. arXiv preprint arXiv:1407.3501, 2014.• Szegedy, Christian, Zaremba, Wojciech, Sutskever, Ilya, Bruna, Joan, Erhan,

Dumitru, Goodfellow, Ian J., and Fergus, Rob. Intriguing properties ofneural networks. CoRR, abs/1312.6199, 2013.• Goodfellow, Ian J, Shlens, Jonathon, and Szegedy, Christian. Explaining

and Harnessing Adversarial Examples. ArXiv e-prints, December 2014.• Jason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, and Hod Lipson.

Understanding Neural Networks Through Deep Visualization, ICML DLWorkshop 2015.• Alexander Mordvintsev, Christopher Olah, Mike Tyka. Inceptionism: Going

Deeper into Neural Networks http://googlere-search.blogspot.jp/2015/06/inceptionism-going-deeper-into-neural.html

2015/10/15 味曽野雅史 (weblab B4) 36