Skip to content

Instantly share code, notes, and snippets.

@shagunsodhani
shagunsodhani / ElasticNet.md
Created March 13, 2016 17:51
Notes for "Regularization and variable selection via the elastic net" paper.

Regularization and variable selection via the elastic net

Introduction to elastic net

  • Regularization and variable selection method.
  • Sparse Representation
  • Exihibits grouping effect.
  • Prticulary useful when number of predictors (p) >> number of observations (n).
  • LARS-EN algorithm to compute elastic net regularization path.
  • Link to paper.
@shagunsodhani
shagunsodhani / Word2Vec.md
Created March 20, 2016 15:04
Summary of paper titled "Efficient Estimation of Word Representations in Vector Space"

Efficient Estimation of Word Representations in Vector Space

Introduction

Model Architecture

@shagunsodhani
shagunsodhani / DistBelief.md
Created April 2, 2016 07:10
Notes for "Large Scale Distributed Deep Networks" paper

Large Scale Distributed Deep Networks

Introduction

  • In machine learning, accuracy tends to increase with an increase in the number of training examples and number of model parameters.
  • For large data, training becomes slow on even GPU (due to increase CPU-GPU data transfer).
  • Solution: Distributed training and inference - DistBelief
  • Link to paper

DistBelief

@shagunsodhani
shagunsodhani / DistributedGraphLab.md
Created April 11, 2016 02:01
Notes for "Distributed GraphLab: A Framework for Machine Learning and Data Mining in the Cloud" paper.

Introduction

  • GraphLab abstraction exposes asynchronous, dynamic, graph-parallel computation model in the shared-memory setting.
  • This paper extends the abstraction to the distributed setting.
  • Link to the paper.

Characteristics of MLDM (Machine Learning and Data Mining)

  • Graph Structured Computation
  • Sometimes computation requires modeling dependencies between data.
@shagunsodhani
shagunsodhani / Learning to execute.md
Created April 18, 2016 17:41
Notes for "Learning To Execute" paper

Learning To Execute

Problem Statement

  • Evaluating if LSTMs can express and learn short, simple programs (linear time, constant memory) in the sequence-to-sequence framework.
  • Link to paper

Approach

@shagunsodhani
shagunsodhani / Memory Networks.md
Last active March 28, 2024 11:17
Notes for "Memory Networks" paper

Memory Networks

Introduction

  • Memory Networks combine inference components with a long-term memory component.
  • Used in the context of Question Answering (QA) with memory component acting as a (dynamic) knowledge base.
  • Link to the paper.

Related Work

@shagunsodhani
shagunsodhani / End-To-End Memory Networks.md
Created April 24, 2016 14:19
Notes for End-To-End Memory Networks paper

End-To-End Memory Networks

Introduction

  • Neural Network with a recurrent attention model over a large external memory.
  • Continous form of Memory-Network but with end-to-end training so can be applied to more domains.
  • Extension of RNNSearch and can perform multiple hops (computational steps) over the memory per symbol.
  • Link to the paper.
  • Link to the implementation.
@shagunsodhani
shagunsodhani / CurriculumLearning.md
Created May 8, 2016 17:14
Notes for Curriculum Learning paper

Curriculum Learning

Introduction

  • Curriculum Learning - When training machine learning models, start with easier subtasks and gradually increase the difficulty level of the tasks.
  • Motivation comes from the observation that humans and animals seem to learn better when trained with a curriculum like a strategy.
  • Link to the paper.

Contributions of the paper

@shagunsodhani
shagunsodhani / t-SNE.md
Created May 15, 2016 09:30
Notes for t-SNE paper

Visualizing Data using t-SNE

Introduction

  • Method to visualize high-dimensional data points in 2/3 dimensional space.
  • Data visualization techniques like Chernoff faces and graph approaches just provide a representation and not an interpretation.
  • Dimensionality reduction techniques fail to retain both local and global structure of the data simultaneously. For example, PCA and MDS are linear techniques and fail on data lying on a non-linear manifold.
  • t-SNE approach converts data into a matrix of pairwise similarities and visualizes this matrix.
  • Based on SNE (Stochastic Neighbor Embedding)
  • Link to paper
@shagunsodhani
shagunsodhani / LargeVis.md
Last active November 10, 2023 02:05
Notes for LargeVis paper

#Visualizing Large-scale and High-dimensional Data

Introduction

  • LargeVis - a technique to visualize large-scale and high-dimensional data in low-dimensional space.
  • Problem relates to both information visualization and machine learning (and data mining) domain.
  • Link to the paper

t-SNE