일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | ||||
4 | 5 | 6 | 7 | 8 | 9 | 10 |
11 | 12 | 13 | 14 | 15 | 16 | 17 |
18 | 19 | 20 | 21 | 22 | 23 | 24 |
25 | 26 | 27 | 28 | 29 | 30 |
- GAN
- Deep learning
- meta-learning
- cs231n
- Interpretability
- 코딩 테스트
- grad-cam
- 딥러닝
- keras
- 설명가능한 인공지능
- xai
- 기계학습
- 메타러닝
- Cam
- 시계열 분석
- AI
- Score-CAM
- 설명가능한
- coding test
- Class activation map
- Explainable AI
- 머신러닝
- 백준
- python
- Artificial Intelligence
- Unsupervised learning
- SmoothGrad
- 인공지능
- 코딩테스트
- Machine Learning
- Today
- Total
목록2021/04 (8)
iMTE
논문 제목 : Interpretable and fine-grained visual explanations for CNNs 논문 주소 : openaccess.thecvf.com/content_CVPR_2019/html/Wagner_Interpretable_and_Fine-Grained_Visual_Explanations_for_Convolutional_Neural_Networks_CVPR_2019_paper.html CVPR 2019 Open Access Repository Interpretable and Fine-Grained Visual Explanations for Convolutional Neural Networks Jorg Wagner, Jan Mathias Kohler, Tobias Gindel..
논문 제목 : Sanity checks for saliency maps 논문 주소 : arxiv.org/abs/1810.03292 Sanity Checks for Saliency Maps Saliency methods have emerged as a popular tool to highlight features in an input deemed relevant for the prediction of a learned model. Several saliency methods have been proposed, often guided by visual appeal on image data. In this work, we propose an a arxiv.org 주요 내용 : 1) Saliency map은 학..
논문 제목 : Sanity checks for saliency maps 논문 주소 : arxiv.org/abs/1810.03292 Sanity Checks for Saliency Maps Saliency methods have emerged as a popular tool to highlight features in an input deemed relevant for the prediction of a learned model. Several saliency methods have been proposed, often guided by visual appeal on image data. In this work, we propose an a arxiv.org 주요 수식 정리: 0) Definition in..
논문 제목 : SmoothGrad : removing noise by adding noise 논문 주소 : arxiv.org/abs/1706.03825 SmoothGrad: removing noise by adding noise Explaining the output of a deep network remains a challenge. In the case of an image classifier, one type of explanation is to identify pixels that strongly influence the final decision. A starting point for this strategy is the gradient of the class score arxiv.org 주요 ..
논문 제목 : Smooth Grad-CAM++: An Enhanced Inference Level Visualization Technique for Deep Convolutional Neural Network Models 논문 주소 : arxiv.org/abs/1908.01224 Smooth Grad-CAM++: An Enhanced Inference Level Visualization Technique for Deep Convolutional Neural Network Models Gaining insight into how deep convolutional neural network models perform image classification and how to explain their outpu..
논문 제목 : Grad-CAM++: Generalized Gradient-based Visual Explanations for Deep Convolutional Networks 논문 주소 : arxiv.org/pdf/1710.11063.pdf IEEE WACV (2018, ieeexplore.ieee.org/document/8354201)에 나온 논문을 바탕으로 이해하고 내용을 작성한다. arixv에서 나온 버전이 좀 더 extended version임으로 Grad-CAM++에 더 깊은 이해를 위해서는 extended version을 읽는 것을 추천한다. 주요 내용 : 1) Deep models은 "black box"로서 internal function을 이해하는데에는 어려움이 있다. 이를 해결하기 위해..
논문 제목 : Learning deep features for discriminative localization 논문 주소 : openaccess.thecvf.com/content_iccv_2017/html/Selvaraju_Grad-CAM_Visual_Explanations_ICCV_2017_paper.html ICCV 2017 Open Access Repository Grad-CAM: Visual Explanations From Deep Networks via Gradient-Based Localization Ramprasaath R. Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, Dhruv Batra; Pr..
논문 제목 : Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization 논문 주소 : openaccess.thecvf.com/content_iccv_2017/html/Selvaraju_Grad-CAM_Visual_Explanations_ICCV_2017_paper.html ICCV 2017 Open Access Repository Grad-CAM: Visual Explanations From Deep Networks via Gradient-Based Localization Ramprasaath R. Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, De..