Last active
January 26, 2023 05:12
-
-
Save ZHAOZHIHAO/ebdeb8f790e78ae8a85e241f16c2fde8 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Papers: | |
When Saliency Meets Sentiment: Understanding How Image Content Invokes Emotion and Sentiment | |
1. Very very good result, code available on github | |
Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields | |
2. Deep Feature Flow for Video Recognition | |
code: https://github.com/msracver/Deep-Feature-Flow | |
Fast object detection and segmentation in a video. | |
22.25fps reported on the Cityscapes andImageNet VID dataset with NVIDIA K40 GPU and Intel Core i7-4790 CPU. | |
The naive per-frame method runs 4.05 fps. | |
Comparsion between the naive method, which can be seen as any state of the art segmentation and detection algorithom perform on every frame, | |
and the proposed method, is as follwing: | |
per-frame proposed dataset | |
detection 73.9 73.1 ImageNet VID | |
segmetation 71.1 69.2 Cityscapes | |
I think the small loss of accuracy can be ingored. | |
When I test the code, the result is not good at all. | |
3. One-Shot Video Object Segmentation | |
code: http://www.vision.ee.ethz.ch/~cvlsegmentation/osvos/ | |
4. Zero-Shot Learning - The Good, the Bad and the Ugly | |
Many zero-shot got very very bad result in the wild, only good at the dataset the authors used. | |
The author tried almost all zero shot methods and released code at https://www.mpi-inf.mpg.de/departments/computer-vision-and-multimodal-computing/research/zero-shot-learning/zero-shot-learning-the-good-the-bad-and-the-ugly/ | |
5. Deep Image Matting | |
Code(implenmented a baidu engineer): https://github.com/Joker316701882/Deep-Image-Matting | |
The code is easy to run: having tensorflow prepared, download the pretrained model, then run it. | |
His comment here: http://blog.leanote.com/post/calebge/Deep-Image-Matting%E5%A4%8D%E7%8E%B0%E8%BF%87%E7%A8%8B%E6%80%BB%E7%BB%93 | |
5. Focal Loss for Dense Object Detection | |
Unoffical code: https://github.com/unsky/focal-loss | |
6. Understanding Black-box Predictions via Influence Functions | |
Someone's comment: https://github.com/eolecvk/InfluenceFunctions/blob/master/review.pdf | |
Another comment: http://nooverfit.com/wp/icml-2017%E8%AE%BA%E6%96%87%E7%B2%BE%E9%80%891-%E7%94%A8%E5%BD%B1%E5%93%8D%E5%87%BD%E6%95%B0%E7%90%86%E8%A7%A3%E6%9C%BA%E5%99%A8%E5%AD%A6%E4%B9%A0%E4%B8%AD%E7%9A%84%E9%BB%91%E7%9B%92%E9%A2%84/ | |
7. Convexified Convolutional Neural Networks | |
official code: https://github.com/zhangyuc/CCNN | |
8. On Calibration of Modern Neural Networks | |
[来自知乎] 也是在deep learning session, 也是百分之百的实验。不过它报告了一个俺从来没有注意过的有趣现象: | |
基于深度网络的分类器输出的class label probability 都over-confident;而这种现象在浅层网络中并不存在。这后面没准有东西可以挖一挖。 | |
9. Learned Invariant Feature Transform | |
Use CNN to generate key point descriptors. | |
Code: https://github.com/cvlab-epfl/LIFT | |
10. Person Re-identification in the Wild | |
11. NetVLAD: CNN architecture for weakly supervised place recognition | |
12. These are learning with nosiy labels: | |
Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach CVPR 2017 | |
See my comment. | |
Toward robustness against label noise in training deep discriminative neural networks ICLR 2017 | |
In this paper, their model are not suitable for unknown T in "Making Deep Neural Networks Robust to Label Noise: a Loss | |
Correction Approach". Known T is used. | |
Learning from noisy large-scale datasets with minimal supervision CVPR 2017 | |
i. Their algorithm can be explained with one picture(fig 3). But you probably need to read section 3 to understand figure 3. | |
ii. From table 2(their result), I think their result is not good, since they got only 2(on AP), and 1(on MAP) point precision | |
increasement, compared to "Fine-Tuning with clean labels", which is a very simple and naive method. They have a very very | |
complex network compared to simple "Fine-Tuning with clean labels". | |
I think, compared to "Fine-Tuning with clean labels", of course, the model can get better performance, since it utilizes | |
prior knowledge. But the result is not much better than simple, naive method. | |
iii. From ii, I think they didn't find the crucial factor, or didn't find a good way, to solve this problem. | |
iv. On the dataset used in paper of Multi-label fashion image classification with minimal human supervision, | |
their method is even worser than "Fine-Tuning with clean labels", lmao. | |
Learning visual n-grams from web data ICCV 2017 | |
Multi-label fashion image classification with minimal human supervision ICCV 2017 | |
i. Almost the same models as Learning from noisy large-scale datasets with minimal supervision, | |
the difference can be seen from section 4.2 or Figure 5. | |
ii. Got less than 1 point improvement on metric AP & MAP, compared to "Fine-Tuning with clean labels", see Table 2. | |
Iterative Learning with Open-set Noisy Labels CVPR 2018 spotlight | |
The key point in this paper is that they uses a algorithm, based on feature distance, to detect outlier/wrong labeled image. | |
Their whole algorithm can be seen in figure 3. | |
AMBIENTGAN: GENERATIVE MODELS FROM LOSSY MEASUREMENTS ICLR 2018 ORAL | |
i. This paper is the greatest one in writing I ever read till now. | |
ii. The idea is quite simple that can be said in one sentence: | |
The generator learns the mapping f: image -> nosiy image, by applying its conjectural f' to an image and | |
sending this image to discriminator. | |
Longer explanation | |
"rather than distinguish a real image from a generated image as in a traditional GAN, our discriminator must | |
distinguish a real measurement from a simulated measurement of a generated image;" | |
The word "measurement" in the above sentence means a distorted image. The distorted image may lacks some patchs, or some | |
pixels are set to zero, etc... The discriminator needs to classify if an image is real measurement. Real measurement is | |
groundtruth that artifically made(so these are used for training the discriminator), simulated measurement is result of | |
performing a distortion(f(theta)) on lossless images(so these images want to fool the discriminator). | |
learning from noisy single-labeled data, ICLR 2018 | |
Unsupervised Learning by Predicting Noise, ICML 2017 | |
Others' notes on zhihu & google. | |
https://zhuanlan.zhihu.com/p/27614220 | |
unsupervised representation learning by predicting image rotations | |
Learn feature representation to predict image rotation. | |
From Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels | |
i. Other approaches(correction) include treating the true labels as a latent variable and the noisy labels as an observed variable so that | |
EM-like algorithms can be used to learn true label distribution of the dataset. | |
ii. Techniques to re-weight confident samples have also been proposed | |
iii. Another popular approach attempts at cleaning up noisy labels. | |
a. label cleaning network with some clean labels | |
b. prune the correct samples based on softmax outputs | |
iv. loss | |
The traditional CNN architecture is designed for clean dataset. For nosiy dataset, I think, we can use the CNN extract features, | |
but the other parts should be changed. | |
13. Global Optimality in Neural Network Training | |
14. Learning from Simulated and Unsupervised Images through Adversarial Training | |
15. One shot | |
Matching Networks for One Shot Learning | |
Comment by Andrej Karpathy(OpenAI, Li FeiFei's student) | |
https://github.com/karpathy/paper-notes/blob/master/matching_networks.md | |
16. deep learning & unsupervised | |
Clustering with Deep Learning: Taxonomy and New Methods, rejected by ICLR 2018, but a nice survey | |
comments by ICLR reviewers: https://openreview.net/forum?id=B1CEaMbR- | |
Learning Discrete Representations via Information Maximizing Self-Augmented Training, ICML 2017 | |
CNN-Based Joint Clustering and Representation Learning with Feature Drift Compensation for Large-Scale Image Data | |
Published in: IEEE Transactions on Multimedia ( Volume: 20, Issue: 2, Feb. 2018) | |
17. Problems in ML/DL/CV area | |
Troubling Trends in Machine Learning Scholarship | |
blog: https://news.cnblogs.com/n/601590/ | |
Winner's Curse? On Pace, Progress, and Empirical Rigor | |
blog: http://www.sohu.com/a/224432485_114877 | |
18. People Count | |
CVPR 2016, Single-image crowd counting via multi-column convolutional neural network | |
unofficial code: https://github.com/svishwa/crowdcount-mcnn | |
Nice summary till April 2018: http://gjy3035.github.io/reading/Awesome-Crowd-Counting.html | |
19. VQA | |
CVPR 2018, Tips and tricks for visual question answering: Learnings from the 2017 challenge | |
ECCV 2018, Learning Visual Question Answering by Bootstrapping Hard Attention | |
hard attention, where some information is selectively ignored is used instead of | |
soft attention, where information is re-weighted and aggregated, but never filtered out. | |
CVPR 2018, Learning Answer Embeddings for Visual Question Answering | |
The proposed approach takes the semantic relationships (as characterized by the embeddings) among answers into | |
consideration, instead of viewing them as independent ordinal numbers. The learned embedded function can be used | |
to embed unseen answers (in the training dataset). | |
Key: These properties make the approach particularly appealing for transfer learning for open-ended Visual QA, where | |
the source dataset on which the model is learned has limited overlapping with the target dataset in the space of | |
answers. | |
CVPR 2018, Cross-Dataset Adaptation for Visual Question Answering | |
cross-dataset adaptation for visual question answering, our goal is to train a Visual QA model on a source dataset | |
but apply it to another target one. | |
The key challenge is that the two datasets are constructed differently, resulting in the cross-dataset mismatch | |
on images, questions, or answers. | |
We overcome this difficulty by proposing a novel domain adaptation algorithm. | |
CVPR 2018, Embodied Question Answering | |
We present a new AI task – Embodied Question Answering (EmbodiedQA) – where an agent is spawned at a random | |
location in a 3D environment and asked a question (‘What color is the car?’). In order to answer, the agent must | |
first intelligently navigate to explore the environment, gather necessary visual information through | |
first-person(egocentric) vision, and then answer the question (‘orange’). | |
In this work, we develop a dataset of questions and answers in House3D environments [1], evaluation metrics, and | |
a hierarchical model trained with imitation and reinforcement learning. | |
CVPR 2018, Differential Attention for Visual Question Answering | |
In a word, they got better attention via complex network | |
See figure 1 & figure 5 to briefly see what they do. | |
However, the regions that the previous systems focus on are not correlated with the regions that humans focus on. | |
(The previous methods' attention is not like human.) The accuracy is limited due to this drawback. | |
In this paper, we propose to solve this problem by using an exemplar based method. We obtain one or more supporting | |
and opposing exemplars to obtain a differential attention region. | |
This differential attention is closer to human attention than other image based attention methods. It also helps in | |
obtaining improved accuracy when answering questions. | |
CVPR 2018, Learning by Asking Questions | |
Question in VQA is also learned(there is a question generator). See Figure 3. | |
blog, VQA 之 Multimodal Compact Bilinear Pooling | |
https://blog.csdn.net/bea_tree/article/details/72903566 | |
20. Shadow Detection | |
ECCV 2018, Bidirectional Feature Pyramid Network with Recurrent Attention Residual Modules for Shadow Detection | |
Code on github, just modify the network. | |
21. | |
CVPR 2018, PoseTrack: A Benchmark for Human Pose Estimation and Tracking | |
CVPR 2018, Detect-and-Track: Efficient Pose Estimation in Videos | |
CVPR 2017, ArtTrack: Articulated Multi-person Tracking in the Wild | |
pose flow: Efficient online pose tracking | |
Simple Baselines for Human Pose Estimation and Tracking | |
blog on multi methods: https://blog.csdn.net/white_xiaohao/article/details/79087565 | |
Comparing MASK RCNN with Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields, the architecture may doesn't | |
matter as long as you get the correct core module/element/way. | |
22. Learning Dual Convolutional Neural Networks for Low-Level Vision | |
23. Interaction reasoning | |
NIPS 2016, Interaction Networks for Learning about Objects, Relations and Physics | |
Reasoning the 4th frame of toy data. | |
reviews: https://media.nips.cc/nipsbooks/nipspapers/paper_files/nips29/reviews/2244.html | |
NIPS 2017, Visual Interaction Networks | |
Reasoning the 4th, 5th, 6th ... frame of toy data. | |
reviews: https://media.nips.cc/nipsbooks/nipspapers/paper_files/nips30/reviews/2374.html | |
NIPS 2017, A simple neural network module for relational reasoning | |
The paper proposes a plug and play module (called Relation Networks (RNs)) specialized for relational reasoning. | |
The module is composed of Multi Layer Perceptrons and considers relations between all pairs of objects. | |
reviews: https://media.nips.cc/nipsbooks/nipspapers/paper_files/nips30/reviews/2565.html | |
ICLR 2018, Graph Networks as Learnable Physics Engines for Inference and Control | |
Using graph networks to learn physics. | |
arxiv June 2018, Relational Deep Reinforcement Learning | |
arxiv June 2018, Flexible Neural Representation for Physics Prediction | |
arxiv Sep 2018, Neural Allocentric Intuitive Physics Prediction from Real Videos | |
ICML 2017, Neural Message Passing for Quantum Chemistry | |
Using NN to predict chemistry properties. | |
Talk on ICML 2017, https://vimeo.com/238221016 | |
digression: | |
Learning to Decompose and Disentangle Representations for Video Prediction | |
24. ECCV 2018 to read | |
The Devil of Face Recognition is in the Noise | |
GANimation: Anatomically-aware Facial Animation from a Single Image | |
Learning to Blend Photos | |
Imagine This! Scripts to Compositions to Videos | |
CVPR 2018 to read | |
Learning to Segment Every Thing | |
LayoutNet: Reconstructing the 3D Room Layout from a Single RGB Image | |
Taskonomy: Disentangling Task Transfer Learning | |
https://zhuanlan.zhihu.com/p/38425434 | |
Cross-Domain Weakly-Supervised Object Detection through Progressive Domain Adaptation | |
Unsupervised Discovery of Object Landmarks as Structural Representations | |
What have we learned from deep representations for action recognition? | |
Semi-parametric Image Synthesis | |
VITON: An Image-based Virtual Try-on Network | |
others: | |
LINE ARTIST: A Multi-style Sketch to Painting Synthesis Scheme | |
25. Convolutional Neural Networks on Graphs | |
PPT: http://helper.ipam.ucla.edu/publications/dlt2018/dlt2018_14506.pdf | |
video on this PPT: https://www.youtube.com/watch?v=v3jZRkvIOIM&t=0s&list=WL&index=5 | |
a work by this guy on ICLR 2018 workshop, An Experimental Study of Neural Networks for Variable Graphs, and the review | |
https://openreview.net/forum?id=SJexcZc8G | |
26. Placehoder title:) | |
ICLR 2018, i-RevNet: Deep Invertible Networks | |
reviews: https://openreview.net/forum?id=HJsjkMb0Z | |
ICLR 2019 under review, The Forward-Backward Embedding of Directed Graphs | |
Code for this paper is available on github :) | |
ICLR 2019 under review, Invariance and Inverse Stability under ReLU | |
Githubs: | |
Openai's platform that you can write your code to play game, then test with this platform. I think it can reduce a lot of your time. | |
https://github.com/openai/universe | |
Neural Complete | |
Use a network to generate another network, the generated network is in Python keras code. | |
https://github.com/kootenpv/neural_complete | |
Websites: | |
1. 如何学习统计学,或我的学习之路——初学者写给初学者 | |
https://cosx.org/2008/11/how-to-learn-statistics-by-jthu/ | |
讲的挺对我的胃口的,感觉这才是真的学习科学的方法。 | |
2. Funny: "Tips on publishing in NIPS, ICML or any top tier conferences for ML" | |
https://www.reddit.com/r/MachineLearning/comments/3x3urc/tips_on_publishing_in_nips_icml_or_any_top_tier/ | |
3. Changing gazes of people in photos | |
You can test it online. I tested with my own photo, amazing result | |
http://163.172.78.19/ | |
4. deep-learning-most-amazing-applications | |
http://www.yaronhadad.com/deep-learning-most-amazing-applications/ | |
5. Comments on many CVPR2017 papers | |
http://blog.csdn.net/zhangjunhit/article/category/6801399 | |
6. a blog on MXNET: simple introduction | |
http://blog.csdn.net/u013713010/article/details/71635814http://blog.csdn.net/u013713010/article/details/71635814 | |
7. comments on CVPR 2017 | |
https://medium.com/@Synced/cvpr-2017-the-fusion-of-deep-learning-and-computer-vision-whats-next-f4d8e9efe2c | |
8. 图像拼接近年发展,介绍了各个主要方法的核心内容,写的很好 | |
https://www.zhihu.com/question/34535199/answer/135169187 | |
9. Machine Learning Top 10 Open Source Projects (v.Feb 2018) | |
https://medium.mybridge.co/machine-learning-top-10-open-source-projects-v-feb-2018-d1d39062bd20 | |
10. CS231n Lecture 10 - Recurrent Neural Networks, Image Captioning, LSTM | |
Very clear, not hard to understand for new beginners, nice. | |
https://www.youtube.com/watch?v=iX5V1WpxxkY | |
11. Understanding LSTM Networks | |
http://colah.github.io/posts/2015-08-Understanding-LSTMs/ | |
There is no detail equations, for the equations in the article, see | |
https://www.cnblogs.com/zhangchaoyang/articles/6684906.html | |
12. CVPR2017 Image Caption有关论文总结 | |
https://zhuanlan.zhihu.com/p/30893160 | |
My: 第三篇Skeleton Key: Image Captioning by Skeleton-Attribute Decomposition里面的黑体"ours"不是最好的"ours",而是“with a removed”的一个测试, | |
证明他们的trick有效的。第三篇效果比第一篇效果要好的。可以对比第一篇的table 4和第三篇的table 2 | |
My: 除了这四篇,CVPR2017还有如下image cpation文章: | |
Deep Reinforcement Learning-Based Image Captioning With Embedding Reward | |
Captioning Images With Diverse Objects | |
Self-Critical Sequence Training for Image Captioning | |
13. The Unreasonable Effectiveness of Recurrent Neural Networks | |
http://karpathy.github.io/2015/05/21/rnn-effectiveness/ | |
14. A significant drawback of image caption | |
https://zhuanlan.zhihu.com/p/34188400 | |
15. Very detailed explanation on Hough transformation | |
https://blog.csdn.net/sunshine_in_moon/article/details/45273909 | |
clear, short, with python code | |
https://alyssaq.github.io/2014/understanding-hough-transform/ | |
16. 2018-2019 International Conferences in Artificial Intelligence, Computer Vision and Image Processing | |
https://jackietseng.github.io/conference_call_for_paper/2018-2019-conferences.html | |
17. 为什么要学数学?——《魔鬼数学》[美]乔丹•艾伦伯格 | |
it's funny | |
https://www.youtube.com/watch?v=9uqq2ottHnw | |
18. CVPR2018-Segmentation相关论文整理 | |
very nice | |
https://blog.csdn.net/qq_16761599/article/details/80727385 | |
19. CV 届的金鸡百花奖:盘点我心中的 CVPR 2018 创意 TOP10 | |
nice | |
https://www.leiphone.com/news/201807/jWMEVsMly1UigfHk.html | |
20. ICLR2018录用论文,七大趋势全解读 | |
nice | |
http://baijiahao.baidu.com/s?id=1599776977066292834&wfr=spider&for=pc | |
21. Asking questions to images with deep learning: a visual-question-answering tutorial | |
a nice introduction tutorial though it doesn't contain latest contents. | |
https://blog.floydhub.com/asking-questions-to-images-with-deep-learning/ | |
22. visual question answering papers & projects, it's up to latest | |
https://github.com/handong1587/handong1587.github.io/blob/d83f99c0dbb313cdc5f3870d94e52d7bf69a95ed/_posts/deep_learning/2015-10-09-vqa.md | |
23. Introduction to Visual Question Answering: Datasets, Approaches and Evaluation | |
https://tryolabs.com/blog/2018/03/01/introduction-to-visual-question-answering/ | |
24. Research Topics of Computer Vision & Graphics Group | |
https://www.hhi.fraunhofer.de/en/departments/vit/research-groups/computer-vision-graphics/research-topics.html | |
25. Hinton 新作「在线蒸馏」,提升深度学习分布式训练表现的利器 | |
well written, motivation and how it works are well introduced | |
http://baijiahao.baidu.com/s?id=1598625949059208402&wfr=spider&for=pc | |
26. 3D segmentation | |
a nice introduction | |
https://thegradient.pub/beyond-the-pixel-plane-sensing-and-learning-in-3d/ | |
27. Online segmentation demo, SegNet | |
http://mi.eng.cam.ac.uk/projects/segnet/demo.php#demo | |
28. very short summary on many many 2017 ICLR papers | |
quality 奈斯! | |
https://zhuanlan.zhihu.com/p/26641697 | |
29. 关系推理水平超越人类:DeepMind展示全新神经网络推理预测技术 | |
https://baijiahao.baidu.com/s?id=1569520267563057&wfr=spider&for=pc | |
30. 图深度学习(GraphDL),下一个人工智能算法热点?一文了解最新GDL相关文章 | |
https://mp.weixin.qq.com/s?__biz=MzU2OTA0NzE2NA%3D%3D&mid=2247490830&idx=1&sn=6356b25bbf7e0983caad352b513c2994&scene=45#wechat_redirect | |
One Oxford student's note on graphDL paper | |
https://yobibyte.github.io/pages/paper-notes.html | |
31. 神经网络也可以有逻辑——解析视觉推理(Visual Reasoning) | |
http://www.sohu.com/a/167168537_633698 | |
32. 深度学习结合SLAM 语义slam 文章和代码 | |
https://blog.csdn.net/qq_40213457/article/details/81407786 | |
33. ECCV 2018论文解读及资源集锦(9月21日更新,含全部论文下载链接) | |
http://www.cvmart.net/community/article/detail/298 | |
CVPR 2018 论文解读集锦(8月21日更新) | |
http://www.cvmart.net/community/article/detail/214 | |
https://zhuanlan.zhihu.com/cvpr2018 | |
34. 费曼谈巴西的物理教育 | |
http://blog.sciencenet.cn/blog-1319915-1012178.html | |
35. SLAM的前世今生 终于有人说清楚了 | |
全面,理论和实际(市场)都有 | |
https://www.cnblogs.com/zengcv/p/5994587.html | |
36. Applications of deep learning on image processing | |
https://blog.csdn.net/c2a2o2/article/details/77701181 | |
37. MIT在读博士心得:做好AI科研,你需要注意什么? | |
https://www.sohu.com/a/230352973_236505 | |
38. Artificial Intelligence — The Revolution Hasn’t Happened Yet | |
https://medium.com/@mijordan3/artificial-intelligence-the-revolution-hasnt-happened-yet-5e1d5812e1e7 | |
39. Capsule network | |
Hinton's talk at MIT. He also has this talk at other places. | |
https://www.youtube.com/watch?v=rTawFwUvnLE | |
Hinton said, "This is an amazingly good video. I wish I could explain capsules that well." | |
https://www.reddit.com/r/MachineLearning/comments/7ew7ba/d_capsule_networks_capsnets_tutorial/ | |
Chinese translation of this video, in text format. | |
https://mp.weixin.qq.com/s/9BIbthQvePqeVdLWil7vgQ | |
Capsule related papers: | |
CVPR 2018, FACSCaps: Pose-Independent Facial Action Coding with Capsule | |
ECCV 2018, Neural Network Encapsulation | |
ICLR 2018 workshop, Spectral Capsule Networks | |
40. Interpreting and Explaining Deep Models in Computer Vision, CVPR 2018 Tutorial | |
http://interpretable-ml.org/cvpr2018tutorial/readinglist.html | |
41. Deep Reinforcement Learning Doesn't Work Yet | |
https://www.alexirpan.com/2018/02/14/rl-hard.html | |
Chinese translation | |
https://blog.csdn.net/r1unw1w/article/details/79385925 | |
42. Two nice answers on understanding disparity map | |
https://stackoverflow.com/questions/17607312/difference-between-disparity-map-and-disparity-image-in-stereo-matching | |
43. A short (3:13 mins) video on introducing epipolar plane images | |
https://www.youtube.com/watch?v=1F_5c_escis | |
44. Understanding the Basis of the Kalman Filter Via a Simple and Intuitive Derivation | |
This paper can be found by goole it. | |
45. How to name mathematical variables in a paper? | |
https://math.stackexchange.com/questions/85963/variable-naming-convention-in-mathematical-modeling | |
https://math.stackexchange.com/questions/866885/is-there-a-list-of-typical-variable-letters-to-use-in-a-given-context | |
46. A minimal solution to hand motion capture from a single color camera at over 100fps. Easy to use, plug to run. | |
https://github.com/CalciferZh/minimal-hand | |
https://www.youtube.com/watch?v=OIRulRoBdL4 | |
47. Time Derivative of Rotation Matrices: A Tutorial | |
Abstract — The time derivative of a rotation matrix equals the product of a skew-symmetric matrix and the rotation matrix itself. | |
This article gives a brief tutorial on the well-known result. | |
https://arxiv.org/pdf/1609.06088.pdf | |
48. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead (HIGH QUALITY) | |
https://www.nature.com/articles/s42256-019-0048-x | |
49. Explainable Artificial Intelligence/ interpretable ai | |
Explainable Artificial Intelligence (XAI): Concepts, Taxonomies,Opportunities and Challenges toward Responsible AI | |
Explainable Artificial Intelligence: a Systematic Review | |
Opportunities and Challenges in ExplainableArtificial Intelligence (XAI): A Survey | |
https://github.com/jphall663/awesome-machine-learning-interpretability#free-books | |
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead | |
Molnar, Christoph. Interpretable Machine Learning. Lulu. com, 2020 | |
50. learning from delayed rewards | |
A PhD thsis. This paper creats Q-learning. The concepts are super-clearly explained. | |
This paper would teach people like me on how to write good papers but not hype papers like most CVPR papers do. | |
These old paper would also inspire people, because every detail is clearly explained. More knowledge would give us more imagination. | |
http://www.cs.rhul.ac.uk/~chrisw/new_thesis.pdf | |
51. TEN SIMPLE RULES FOR MATHEMATICAL WRITING | |
actually not only for MATHEMATICAL writting, but also natural language writing. | |
http://web.mit.edu/dimitrib/www/Ten_Rules.pdf | |
51. 图解机器学习的数学基础专辑(完结)总结:who, why, what | |
https://www.bilibili.com/video/BV1iW411T781?p=23 | |
52. neural networks flops computation | |
https://zhuanlan.zhihu.com/p/456046786 | |
53. Deconstructing the Homography Matrix | |
Euclidean, Similarity, Affine, Projective | |
https://medium.com/@insight-in-plain-sight/deconstructing-the-homography-matrix-35989ecc0b2 | |
54. Build opencv python gpu | |
https://forum.opencv.org/t/can-i-use-opencv-python-with-gpu/8947 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment