Skip to content

Instantly share code, notes, and snippets.

@shagunsodhani
shagunsodhani / bAbI.md
Created May 25, 2016 17:18
Notes for "Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks" Paper

Introduction

The paper presents a framework and a set of synthetic toy tasks (classified into skill sets) for analyzing the performance of different machine learning algorithms.

Tasks

  • Single/Two/Three Supporting Facts: Questions where a single(or multiple) supporting facts provide the answer. More is the number of supporting facts, tougher is the task.
  • Two/Three Supporting Facts: Requires differentiation between objects and subjects.
  • Yes/No Questions: True/False questions.
  • Counting/List/Set Questions: Requires ability to count or list objects having a certain property.
@shagunsodhani
shagunsodhani / Question Answering with Subgraph Embeddings.md
Created June 1, 2016 16:33
Notes for "Question Answering with Subgraph Embeddings" paper

Question Answering with Subgraph Embeddings

Introduction

  • Open-domain Question Answering (Open QA) - efficiently querying large-scale knowledge base(KB) using natural language.
  • Two main approaches:
    • Information Retrieval
      • Transform question (in natural language) into a valid query(in terms of KB) to get a broad set of candidate answers.
      • Perform fine-grained detection on candidate answers.
  • Semantic Parsing
@shagunsodhani
shagunsodhani / The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training.md
Created June 8, 2016 17:10
Summary of "The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training" paper

The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training

Introduction

  • The paper explores the challenges involved in training deep networks, the effect of unsupervised pre-training on training process and visualizes the error function landscape for deep architectures.
  • Link to the paper

Experiments

  • Datasets used - Shapeset and MNIST.
@shagunsodhani
shagunsodhani / SeqToSeq.md
Created June 15, 2016 16:28
Summary of "Sequence to Sequence Learning with Neural Networks" paper

Sequence to Sequence Learning with Neural Networks

Introduction

  • The paper proposes a general and end-to-end approach for sequence learning that uses two deep LSTMs, one to map input sequence to vector space and another to map vector to the output sequence.
  • For sequence learning, Deep Neural Networks (DNNs) requires the dimensionality of input and output sequences be known and fixed. This limitation is overcome by using the two LSTMs.
  • Link to the paper

Model

@shagunsodhani
shagunsodhani / QRN.md
Last active September 24, 2016 19:01
Notes for "Query Regression Networks for Machine Comprehension" Paper

Query Regression Networks for Machine Comprehension

Introduction

  • Machine Comprehension (MC) - given a natural language sentence, answer a natural language question.
  • End-To-End MC - can not use language resources like dependency parsers. The only supervision during training is the correct answer.
  • Query Regression Network (QRN) - Variant of Recurrent Neural Network (RNN).
  • Link to the paper

Related Work

@shagunsodhani
shagunsodhani / Advances In Optimizing Recurrent Networks.md
Created June 26, 2016 16:44
Notes for "Advances In Optimizing Recurrent Networks" paper

Advances In Optimizing Recurrent Networks

Introduction

  • Recurrent Neural Networks (RNNs) are very powerful at modelling sequences but they are not good at learning long-term dependencies.
  • The paper discusses the reasons behind this difficulty and some suggestions to mitigate it.
  • Link to the paper.

Optimization Difficulty

@shagunsodhani
shagunsodhani / KeyValueMemNN.md
Last active April 30, 2023 04:13
Summary of paper "Key-Value Memory Networks for Directly Reading Documents"

Key-Value Memory Networks for Directly Reading Documents

Introduction

  • Knowledge Bases (KBs) are effective tools for Question Answering (QA) but are often too restrictive (due to fixed schema) and too sparse (due to limitations of Information Extraction (IE) systems).
  • The paper proposes Key-Value Memory Networks, a neural network architecture based on Memory Networks that can leverage both KBs and raw data for QA.
  • The paper also introduces MOVIEQA, a new QA dataset that can be answered by a perfect KB, by Wikipedia pages and by an imperfect KB obtained using IE techniques thereby allowing a comparison between systems using any of the three sources.
  • Link to the paper.

Related Work

@shagunsodhani
shagunsodhani / A Neural Conversational Model.md
Created July 9, 2016 12:53
Notes for paper titled A Neural Conversational Model
@shagunsodhani
shagunsodhani / Deep Math.md
Created July 16, 2016 10:09
Notes for Deep Math paper

Deep Math: Deep Sequence Models for Premise Selection

Introduction

  • Automated Theorem Proving (ATP) - Attempting to prove mathematical theorems automatically.
  • Bottlenecks in ATP:
    • Autoformalization - Semantic or formal parsing of informal proofs.
    • Automated Reasoning - Reasoning about already formalised proofs.
  • Paper evaluates the effectiveness of neural sequence models for premise selection (related to automated reasoning) without using hand engineered features.
  • Link to the paper
@shagunsodhani
shagunsodhani / RecurrentNeuralNetworkRegularization.md
Created July 24, 2016 15:01
Notes for 'Recurrent Neural Network Regularization' paper

Recurrent Neural Network Regularization

Introduction

  • The paper explains how to apply dropout to LSTMs and how it could reduce overfitting in tasks like language modelling, speech recognition, image caption generation and machine translation.
  • Link to the paper
  • Regularisation method that drops out (or temporarily removes) units in a neural network.