Skip to content

Instantly share code, notes, and snippets.

What

Roll your own iPython Notebook server with Amazon Web Services (EC2) using their Free Tier.

What are we using? What do you need?

  • An active AWS account. First time sign-ups are eligible for the free tier for a year
  • One Micro Tier EC2 Instance
  • With AWS we will use the stock Ubuntu Server AMI and customize it.
  • Anaconda for Python.
  • Coffee/Beer/Time
@mrgloom
mrgloom / dnn_compare_optims.py
Last active February 9, 2018 18:04 — forked from syhw/dnn_compare_optims.py
comparing SGD vs SAG vs Adadelta vs Adagrad
"""
A deep neural network with or w/o dropout in one file.
"""
import numpy
import theano
import sys
import math
from theano import tensor as T
from theano import shared
@mrgloom
mrgloom / kmtransformer.py
Last active August 31, 2015 13:46 — forked from larsmans/kmtransformer.py
k-means feature mapper for scikit-learn
from sklearn.base import BaseEstimator, TransformerMixin
from sklearn.metrics.pairwise import rbf_kernel
class KMeansTransformer(BaseEstimator, TransformerMixin):
def __init__(self, centroids):
self.centroids = centroids
def fit(self, X, y=None):
return self
@mrgloom
mrgloom / rbm.py
Last active September 11, 2015 09:59
Some fairly clean (and fast) code for Restricted Boltzmann machines.
"""
Code for training RBMs with contrastive divergence. Tries to be as
quick and memory-efficient as possible while utilizing only pure Python
and NumPy.
"""
# Copyright (c) 2009, David Warde-Farley
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
@mrgloom
mrgloom / pegasos.py
Created October 13, 2015 16:58 — forked from alextp/pegasos.py
class OnlineLearner(object):
def __init__(self, **kwargs):
self.last_misses = 0.
self.iratio = 0.
self.it = 1.
self.l = kwargs["l"]
self.max_ratio = -np.inf
self.threshold = 500.
def hinge_loss(self, vector, cls, weight):
name: "VGG_ILSVRC_16_layer"
layer {
name: "data"
type: "DenseImageData"
top: "data"
top: "label"
dense_image_data_param {
source: "/home/myuser/Downloads/SegNet/SegNet-Tutorial/CamVid/train.txt" # Change this to the absolute path to your data file
batch_size: 1 # Change this number to a batch size that will fit on your GPU
shuffle: true
@mrgloom
mrgloom / segnet_simple_train.prototxt
Created March 31, 2016 14:18
segnet_simple_train.prototxt
name: "segnet"
layer {
name: "data"
type: "DenseImageData"
top: "data"
top: "label"
dense_image_data_param {
source: "/SegNet/CamVid/train.txt" # Change this to the absolute path to your data file
batch_size: 4 # Change this number to a batch size that will fit on your GPU
@mrgloom
mrgloom / detect_multiscale.cpp
Created May 23, 2016 13:20 — forked from thorikawa/detect_multiscale.cpp
Simple example for CascadeClassifier.detectMultiScale
#include <opencv2/opencv.hpp>
#include <vector>
using namespace cv;
using namespace std;
int main () {
Mat img = imread("lena.jpg");
CascadeClassifier cascade;
if (cascade.load("haarcascade_frontalface_alt.xml")) {
@mrgloom
mrgloom / Find header and libs cheatsheet
Last active May 2, 2017 15:41
Find OpenCV function in headers and libs
#To find function in headers
grep -n -r <function_name> <path_to_opencv_include_folder>
#To find function in libs
nm -C -A <path_to_opencv_lib_folder>/*.so | grep <function_name> | grep -v U
#Find package installed via apt-get
apt list --installed | grep <package_name>
#Find files related to package
dpkg -L <package_name>
name: "VGG_ILSVRC_16_layers"
layer {
name: "train-data"
type: "Data"
top: "data"
top: "label"
include {
stage: "train"
}