http://scikit-learn.org/stable/related_projects.html
[operating-hadoop]
HBase is used widely at Facebook and one of the biggest usecase is Facebook Messages. With a billion users there are a lot of reliability and performance challenges on both HBase and HDFS. HDFS was originally designed for a batch processing system like MapReduce/Hive. A realtime usecase like Facebook Messages where the p99 latency can`t be more than a couple hundreds of milliseconds poses a lot of challenges for HDFS. In this talk we will share the work the HDFS team at Facebook has done to support a realtime usecase like Facebook Messages : (1) Using system calls to tune performance; (2) Inline checksums to reduce iops by 40%; (3) Reducing the p99 for read and write latencies by about 10x; (4) Tools used to determine root cause of outliers. We will discuss the details of each technique, the challenges we faced, lessons learned and results showing the impact of each improvement.
speaker: Pritam Damania
| import logging | |
| import matplotlib.pyplot as plt | |
| import numpy as np | |
| import os | |
| import scipy.stats as stats | |
| import sys | |
| def read_data(filename): | |
| """Reads a data file assumed to have at least 2 columns: 1) lat, 2) lng.""" |
TL;DR
Install Postgres 9.5, and then:
sudo pg_dropcluster 9.5 main --stop
sudo pg_upgradecluster 9.3 main
sudo pg_dropcluster 9.3 main| # original example from Digg Data website (Takashi J. OZAKI, Ph. D.) | |
| # http://diggdata.in/post/58333540883/k-fold-cross-validation-in-r | |
| library(plyr) | |
| library(randomForest) | |
| data <- iris | |
| # in this cross validation example, we use the iris data set to |
| import numpy | |
| from scipy.ndimage.interpolation import map_coordinates | |
| from scipy.ndimage.filters import gaussian_filter | |
| def elastic_transform(image, alpha, sigma, random_state=None): | |
| """Elastic deformation of images as described in [Simard2003]_. | |
| .. [Simard2003] Simard, Steinkraus and Platt, "Best Practices for | |
| Convolutional Neural Networks applied to Visual Document Analysis", in |
| import numpy as np | |
| from keras.datasets import imdb | |
| from keras.preprocessing.sequence import pad_sequences | |
| from keras.models import Sequential | |
| from keras.layers import containers | |
| from keras.layers.noise import GaussianNoise | |
| from keras.layers.core import Dense, AutoEncoder | |
| from keras.utils import np_utils | |
| from sklearn.metrics import (precision_score, recall_score, |
| #How to download streaming video | |
| Streaming just means a download that they don't want you to keep. But Chrome's developer tools make it easy to access what's really going on under the hood. | |
| ##Open Developer Tools | |
| From the page where you want to download some things, go into your chrome menu to open the developer tools. You can either: | |
| 1. (On a mac): Command-option-J | |
| 2. (On a PC): Control-alt-J |