- 1x1 convolutions have been extensively used to reduce the number of parameters without affecting the results much
- Deep Mutual Learning: Unlike bagging/ boosting, models learn jointly, and help each other to fit well
- Skip connections: Help solving
degradation problem
without adding parameters. Hard Sample Mining
Last active
August 13, 2021 11:16
-
-
Save harshraj22/cc64a15c658f964f588ff5c13bacbcde to your computer and use it in GitHub Desktop.
New ideas for BTP: Visual Question Answering
- There exist better evaluation metrics than rule based ones (BLEU, CIDER) as @akhileshkb was mentioning yesterday. reference
- Loss functions similar to classification problem might not be the best idea. Rather exploration of loss function which takes into account the meaning of words might be helpful.
Would create a repo soon, so that all the discussion goes into discussion section of the repo
Knowledge distillation
for building small models (small num of parameters) with high performance.
Discussion moved to vqa/wiki
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
List of some fusion techniques [This could help as we will have to fuse two branches i.e image and text.]
Space Efficient tensor operations: