跳到主要內容

簡易檢索 / 詳目顯示

研究生: 鄭可昕
Cheng, Ko-Hsin
論文名稱: 基於深度學習框架之夜晚霧霾圖像模擬與復原方法評估
Nighttime Haze Images Simulation and Restoration Using Deep Learning Frameworks
指導教授: 廖文宏
Liao, Wen-Hung
口試委員: 陳駿丞
Chen, Jun-Cheng
彭彥璁
Peng, Yan-Tsung
學位類別: 碩士
Master
系所名稱: 理學院 - 資訊科學系碩士在職專班
Excutive Master Program of Computer Science
論文出版年: 2021
畢業學年度: 109
語文別: 中文
論文頁數: 57
中文關鍵詞: 深度學習夜晚圖像霧霾模擬圖像去霧圖像復原
外文關鍵詞: Deep learning, Nighttime images, Fog simulation, Haze removal, Image restoration
DOI URL: http://doi.org/10.6814/NCCU202100281
相關次數: 點閱:98下載:28
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 近年來氣候異常、空氣污染問題日漸嚴重,使得日常中發生霧霾現象的次數越來越多,在霧霾環境中拍攝的圖像,會使得圖像的清晰度與對比度大幅降低,當霧霾現象發生在夜晚,伴隨燈光的干擾,其圖像品質更差。隨著深度學習在圖像領域研究成果的突破,如何將深度學習方法應用於霧霾圖像的復原與去霧,逐漸成為研究者感興趣的主題之一。
    本研究以霧霾圖像形成原理與圖像深度為概念,結合生成對抗網路、大氣散射模型與圖像深度估計等方法,在清晰的夜晚圖像上,疊加霧霾效果,模擬出夜晚霧霾圖像,並透過深度學習方法,將模擬的圖像作為訓練資料,訓練一組模型能夠應用在復原模擬的夜晚霧霾圖像。
    為了對模型進一步評估與分析,本研究亦使用真實夜晚霧霾圖像做測試,檢驗模型的泛化能力。此外,為能更客觀地確認去霧成效,我們計算並比較復原前與復原後圖像之圖像品質指標,以及使用YOLOv5目標偵測方法,以所得之mAP作為衡量基準,均可觀察到處理前後的明顯差異。


    In recent years, extreme weather and air pollution problems have become serious, causing the frequent formation of haze in our daily life. Images taken in a haze environment will significantly lose sharpness and contrast. When haze occurs at nighttime, the image quality worsens due to the interference of light. With the rapid progress of deep learning in the field of computer vision, applying deep neural networks to the restoration of images degraded by haze has become one of the topics of interest to researchers.
    This research employs the concept of fog and haze image formation and image depth, in combination with generative adversarial network, atmospheric scattering model, and image depth estimation, to simulate nighttime haze images by superimposing the haze on clear nighttime images. Based on deep learning methods, the simulated images are used as training data to build a model that can successfully restore the simulated nighttime haze images.
    To evaluate the effectiveness of the proposed approach, we also use real nighttime haze images to observe the generalization ability of the model. To examine the effect of dehazing in an objective manner, several image quality indices have been computed and compared. Additionally, the YOLOv5 object detection method has been utilized to calculate the mAP of the detection before and after restoration. All results indicate improved performance after image dehazing.

    謝辭 i
    摘要 ii
    Abstract iii
    目錄 iv
    圖目錄 vii
    表目錄 ix
    第一章 緒論 1
    1.1 研究背景與動機 1
    1.2 研究目的 2
    1.3 論文架構 3
    第二章 技術背景與相關研究 4
    2.1 深度學習的發展背景 4
    2.2 霧霾圖像形成原理 6
    2.3 生成對抗網路 7
    2.3.1 pix2pix 8
    2.3.2 CycleGAN 9
    2.3.3 ToDayGAN 9
    2.4 基於深度學習的圖像深度資訊估計 10
    2.5 合成霧霾方法 11
    2.6 圖像去霧方法 12
    2.6.1 基於傳統圖像處理 12
    2.6.2 基於深度神經網路 13
    2.7 圖像品質評估指標 13
    2.7.1 峰值訊噪比(PSNR) 14
    2.7.2 結構相似性(SSIM) 14
    2.7.3 均方誤差(MSE) 15
    2.7.4 學習感知圖像塊相似度(LPIPS) 15
    2.7.5 無參考圖像空間品質評估(BRISQUE) 15
    2.7.6 自然圖像品質評估(NIQE) 15
    2.8 目標偵測評估指標 16
    2.8.1 精確率與召回率 16
    2.8.2 IoU 與 mAP 17
    第三章 研究方法 18
    3.1 基本構想 18
    3.2 前期研究 19
    3.2.1 夜晚霧霾圖像一階段生成 20
    3.2.2 夜晚霧霾圖像二階段生成 22
    3.2.3 資料標註 24
    3.2.4 夜晚霧霾圖像去霧 25
    3.2.5 圖像去霧結果評估 26
    3.3 研究架構設計 26
    3.3.1 問題陳述 26
    3.3.2 研究架構 26
    3.4 目標設定 29
    第四章 研究過程與結果分析 30
    4.1 實驗環境 30
    4.2 研究過程 31
    4.2.1 夜晚無霧圖像資料集 31
    4.2.2 白天圖像生成 32
    4.2.3 深度資訊估計 34
    4.2.4 夜晚霧霾圖像模擬 35
    4.2.5 去霧預訓練模型評估 37
    4.2.6 去霧模型重新訓練 38
    4.2.7 去霧效果評估 42
    4.3 結果分析 43
    4.3.1 圖像品質評估分析與討論 43
    4.3.2 霧霾圖像與去霧圖像之目標偵測結果分析 44
    4.3.3 真實夜晚霧霾圖像套用去霧模型結果分析 50
    4.4 研究結果之應用 51
    第五章 結論與未來研究方向 52
    5.1 結論 52
    5.2 未來研究方向 53
    參考文獻 54

    [1] 中央氣象局數位科普網:迷茫之城-是霧,或是霾? https://pweb.cwb.gov.tw/PopularScience/index.php/weather/277
    [2] ImageNet Large Scale Visual Recognition Competition (ILSVRC). http://www.image-net.org/challenges/LSVRC/
    [3] 國立台灣大學計算機及資訊網路中心電子報,第 0038 期,ISSN:2077- 8813, http://www.cc.ntu.edu.tw/chinese/epaper/0038/20160920_3805.html
    [4] Michael A. Nielsen, "Neural Networks and Deep Learning", Determination Press, 2015. http://neuralnetworksanddeeplearning.com/index.html
    [5] ImageNet Winning CNN Architectures (ILSVRC). https://www.kaggle.com/getting-started/149448
    [6] McCartney, E. J. (1976). Optics of the atmosphere: scattering by molecules and particles. nyjw.
    [7] He, K., Sun, J., & Tang, X. (2010). Single image haze removal using dark channel prior. IEEE transactions on pattern analysis and machine intelligence, 33(12), 2341-2353.
    [8] Zhang, T., Shao, C., & Wang, X. (2011, October). Atmospheric scattering- based multiple images fog removal. In 2011 4th International Congress on Image and Signal Processing (Vol. 1, pp. 108-112). IEEE.
    [9] Zhu, J. X., Meng, L. L., Wu, W. X., Choi, D., & Ni, J. J. (2020). Generative adversarial network-based atmospheric scattering model for image dehazing. Digital Communications and Networks.
    [10] Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., ... & Bengio, Y. (2014). Generative adversarial nets. In Advances in neural information processing systems (pp. 2672-2680).
    [11] Isola, P., Zhu, J. Y., Zhou, T., & Efros, A. A. (2017). Image-to-image translation with conditional adversarial networks. InProceedings of the IEEE conference on computer vision and pattern recognition (pp. 1125-1134).
    [12] Zhu, J. Y., Park, T., Isola, P., & Efros, A. A. (2017). Unpaired image-to- image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 2223-2232).
    [13] Anoosheh, A., Sattler, T., Timofte, R., Pollefeys, M., & Van Gool, L. (2019, May). Night-to-day image translation for retrieval-based localization. In 2019 International Conference on Robotics and Automation (ICRA) (pp. 5958-5964). IEEE.
    [14] Li, Z., & Snavely, N. (2018). Megadepth: Learning single-view depth prediction from internet photos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2041-2050).
    [15] Godard, C., Mac Aodha, O., Firman, M., & Brostow, G. J. (2019). Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE international conference on computer vision (pp. 3828-3838).
    [16] Sakaridis, C., Dai, D., & Van Gool, L. (2018). Semantic foggy scene understanding with synthetic data.International Journal of Computer Vision, 126(9), 973-992.
    [17] Zhang, N., Zhang, L., & Cheng, Z. (2017, November). Towards simulating foggy and hazy images and evaluating their authenticity. InInternational Conference on Neural Information Processing (pp. 405-415). Springer, Cham.
    [18] W. Maddern, G. Pascoe, C. Linegar and P. Newman, "1 Year, 1000km: The Oxford RobotCar Dataset", The International Journal of Robotics Research (IJRR), 2016.
    [19] He, K., Sun, J., & Tang, X. (2010). Single image haze removal using dark channel prior. IEEE transactions on pattern analysis and machine intelligence, 33(12), 2341-2353.
    [20] Cai, B., Xu, X., Jia, K., Qing, C., & Tao, D. (2016). Dehazenet: An end-to- end system for single image haze removal. IEEE Transactions on Image Processing, 25(11), 5187-5198.
    [21] Li, Y., Tan, R. T., & Brown, M. S. (2015). Nighttime haze removal with glow and multiple light colors. In Proceedings of the IEEE international conference on computer vision (pp. 226-234).
    [22] Dong, H., Pan, J., Xiang, L., Hu, Z., Zhang, X., Wang, F., & Yang, M. H. (2020). Multi-Scale Boosted Dehazing Network with Dense Feature Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 2157-2167).
    [23] Qin, X., Wang, Z., Bai, Y., Xie, X., & Jia, H. (2020). FFA-Net: Feature Fusion Attention Network for Single Image Dehazing. In AAAI (pp. 11908-11915).
    [24] Wang, Z., Bovik, A. C., Sheikh, H. R., & Simoncelli, E. P. (2004). Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4), 600-612.
    [25] Zhang, R., Isola, P., Efros, A. A., Shechtman, E., & Wang, O. (2018). The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 586-595).
    [26] Mittal, A., Moorthy, A. K., & Bovik, A. C. (2012). No-reference image quality assessment in the spatial domain.IEEE Transactions on image processing, 21(12), 4695-4708.
    [27] Mittal, A., Soundararajan, R., & Bovik, A. C. (2012). Making a “completely blind” image quality analyzer. IEEE Signal processing letters, 20(3), 209-212.
    [28] Yu, F., Chen, H., Wang, X., Xian, W., Chen, Y., Liu, F., ... & Darrell, T. (2020). BDD100K: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 2636-2645).
    [29] Li, B., Ren, W., Fu, D., Tao, D., Feng, D., Zeng, W., & Wang, Z. (2017). Reside: A benchmark for single image dehazing. arXiv preprint arXiv:1712.04143, 1.
    [30] Jocher, G. (2020). Yolov5. Code repository https://github.com/ultralytics/yolov5.
    [31] Tzutalin. LabelImg. Git code (2015). https://github.com/tzutalin/labelImg
    [32] IEEE International Conference on Multimedia and Expo 2020 Grand Challenge – Embedded Deep Learning Object Detection Model Compression Competition for Traffic in Asian Countries. http://2020.ieeeicme.org/www.2020.ieeeicme.org/index.php/grand- challenges/index.html
    [33] Oxford Robotics Institute. Software Development Kit for the Oxford Robotcar Dataset. Git code repository. https://github.com/ori-mrg/robotcar-dataset- sdk
    [34] 交通部高速公路局「交通資料庫」CCTV 動態資訊(v1.1) https://tisvcloud.freeway.gov.tw/

    QR CODE
    :::