Skip to content

Instantly share code, notes, and snippets.

@harshraj22
Last active August 13, 2021 11:16
Show Gist options
  • Save harshraj22/cc64a15c658f964f588ff5c13bacbcde to your computer and use it in GitHub Desktop.
Save harshraj22/cc64a15c658f964f588ff5c13bacbcde to your computer and use it in GitHub Desktop.
New ideas for BTP: Visual Question Answering
  • 1x1 convolutions have been extensively used to reduce the number of parameters without affecting the results much
  • Deep Mutual Learning: Unlike bagging/ boosting, models learn jointly, and help each other to fit well
  • Skip connections: Help solving degradation problem without adding parameters.
  • Hard Sample Mining
@JanhaviDadhania
Copy link

List of some fusion techniques [This could help as we will have to fuse two branches i.e image and text.]

  1. https://arxiv.org/pdf/2003.10758.pdf
  2. https://arxiv.org/pdf/2103.09422v1.pdf
  3. http://www.zhengzhu.net/upload/P6938bc861e8d4583bf47d47d64ed9598.pdf
  4. concatenation
  5. convolution

Space Efficient tensor operations:

  1. https://arxiv.org/pdf/1901.11261.pdf

@harshraj22
Copy link
Author

harshraj22 commented Aug 6, 2021

  1. There exist better evaluation metrics than rule based ones (BLEU, CIDER) as @akhileshkb was mentioning yesterday. reference
  2. Loss functions similar to classification problem might not be the best idea. Rather exploration of loss function which takes into account the meaning of words might be helpful.

Would create a repo soon, so that all the discussion goes into discussion section of the repo

@harshraj22
Copy link
Author

Knowledge distillation for building small models (small num of parameters) with high performance.

@harshraj22
Copy link
Author

Discussion moved to vqa/wiki

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment