brew install redis
Set up launchctl to auto start redis
$ ln -sfv /usr/local/opt/redis/*.plist ~/Library/LaunchAgents
/usr/local/opt/redis/
is a symlink to /usr/local/Cellar/redis/x.y.z
(e.g., 2.8.7
)
from sqlalchemy import create_engine, Table, Column, Integer, Unicode, MetaData, String, Text, update, and_, select, func, types | |
# create engine, reflect existing columns, and create table object for oldTable | |
srcEngine = create_engine('mysql+mysqldb://username:[email protected]/database') # change this for your source database | |
srcEngine._metadata = MetaData(bind=srcEngine) | |
srcEngine._metadata.reflect(srcEngine) # get columns from existing table | |
srcTable = Table('oldTable', srcEngine._metadata) | |
# create engine and table object for newTable | |
destEngine = create_engine('mysql+mysqldb://username:password@localhost/database') # change this for your destination database |
First, install the following libraries:
$ brew install unixodbc
$ brew install freetds --with-unixodbc
FreeTDS should already work now, without configuration:
$ tsql -S [IP or hostname] -U [username] -P [password]
locale is "en_US.UTF-8"
locale charset is "UTF-8"
cat data.json | jq -c -M '.data[]' | \ | |
while read line; do echo $line > parts/$(date +%s%N).json; done |
var fs = require('fs'); | |
var gulp = require('gulp'); | |
var runWintersmith = require('run-wintersmith'); | |
var s3 = require("gulp-s3"); | |
var options = { headers: {'Cache-Control': 'public'} } | |
gulp.task('default', function() { | |
console.log('`gulp deploy` to deploy') | |
}); |
I ran into an issue today where I had to perform a bulk insert into a postgres DB. I was already using SQLAlchemy and Flask-SQLAlchemy to manage the connections to the db and I didn't want to have to use things like psycopg2
directly.
Note: SQLAlchemy provides an ORM. It isn't just an ORM. That is an important thing to be kept in mind. This means that you can bypass choose to not use the ORM layer when you don't want it. The idea with an ORM is to track changes to objects and when you have a case like that is when you'd use the ORM. In a bulk upload scenario, you don't need to track changes to objects. All you care is that everything be pushed into the DB.
var accounts = [ | |
{ name: 'James Brown', msgCount: 123 }, | |
{ name: 'Stevie Wonder', msgCount: 22 }, | |
{ name: 'Sly Stone', msgCount: 16 }, | |
{ name: 'Otis Redding', msgCount: 300 } // Otis has the most messages | |
]; | |
// get sum of msgCount prop across all objects in array | |
var msgTotal = accounts.reduce(function(prev, cur) { | |
return prev + cur.msgCount; |
More details - http://blog.gbaman.info/?p=791
For this method, alongside your Pi Zero, MicroUSB cable and MicroSD card, only an additional computer is required, which can be running Windows (with Bonjour, iTunes or Quicktime installed), Mac OS or Linux (with Avahi Daemon installed, for example Ubuntu has it built in).
1. Flash Raspbian Jessie full or Raspbian Jessie Lite onto the SD card.
2. Once Raspbian is flashed, open up the boot partition (in Windows Explorer, Finder etc) and add to the bottom of the config.txt
file dtoverlay=dwc2
on a new line, then save the file.
3. If using a recent release of Jessie (Dec 2016 onwards), then create a new file simply called ssh
in the SD card as well. By default SSH i
# make sure to replace `<hash>` with your gist's hash
git clone https://gist.github.com/<hash>.git # with https
git clone [email protected]:<hash>.git # or with ssh
# to run: docker-compose run | |
# | |
# Create a .evn file in the same folder as this file and change the variables. | |
# MOUNT_POINT=/tmp/ | |
# VPN_PROVIDER=changeme | |
# VPN_CONFIG=changeme | |
# VPN_USERNAME=changeme | |
# VPN_PASSWORD=changeme | |
# | |
# |