Commonly used scientific symbols in pandoc markdown
encoding is UTF-8, needs xelatex, like this:
---
output:
pdf_document:
latex_engine: xelatex
---
#!/usr/bin/php | |
<?php | |
/** | |
* Checks a list of domains and outputs any that use Google. | |
* | |
* usage: php ./googlecheck.php filewithdomains | |
* | |
*/ | |
stream_set_blocking(STDIN, 0); |
cd /Library/Frameworks/R.framework/Resources/lib | |
# for vecLib use | |
ln -sf libRblas.vecLib.dylib libRblas.dylib | |
# for R reference BLAS use | |
ln -sf libRblas.0.dylib libRblas.dylib |
##################################################### | |
### How to compile R and link it against OpenBLAS ### | |
##################################################### | |
# Assuming no root priviledges on the destination machine, | |
# so we have to compile our software locally | |
# set default installation directory: | |
INSTDIR=$HOME/opt |
Commonly used scientific symbols in pandoc markdown
encoding is UTF-8, needs xelatex, like this:
---
output:
pdf_document:
latex_engine: xelatex
---
This is a set up for projects which want to check in only their source files, but have their gh-pages branch automatically updated with some compiled output every time they push.
A file below this one contains the steps for doing this with Travis CI. However, these days I recommend GitHub Actions, for the following reasons:
library(foreach) | |
library(iterators) | |
library(doParallel) | |
library(tcltk) | |
# Choose number of iterations | |
n <- 1000 | |
cl <- makeCluster(8) |
#plotly box or lasso select linked to | |
# DT data table | |
# using Wage data | |
# the out group: is sex:Male, region:Middle Atlantic + | |
library(ggplot2) | |
library(plotly) | |
library(dplyr) | |
library(ISLR) |
This quick guide for getting a Jupyter Notebook up and running on Bridges, a supercomputer managed by the Pittsburgh Supercomputing Center. Bridges is a new machine designed to accommodate non-traditional uses of High Performance Computing (HPC) resources like data science and digital humanities. Bridges is available through XSEDE, which is the system that manages access to multiple supercomputing resources. Through XSEDE, Bridges is available researchers or educators at US academic or non-profit research institutions (see the XSEDE eligibility policies) Allocations are free, but there is a somewhat difficult to understand application process filled with jargon and acronyms that take time to understand. See the XSEDE getting started guide for more information about getting acc
I followed these two blogs to install server | |
1. http://koo.fi/blog/2015/03/19/openstreetmap-nominatim-server-for-geocoding/#Compile_Nominatim | |
This explains ( and is the main blog which I followed) various steps | |
2. https://www.linuxbabe.com/linux-server/openstreetmap-tile-server-ubuntu-16-04 | |
This explains how to setup swap files and install tiles if needed. | |
I will use Ubuntu 14.04 LTS as the platform. Just a basic install with ssh server. We will install Apache to serve http requests. Make sure you have enough disk space and RAM to hold the data and serve it efficiently. I used the Finland extract, which was about a 200 MB download. The resulting database was 26 GB after importing, indexing and adding Wikipedia data. The Wikipedia data probably actually took more disk space than the OSM data. My server has 4 GB RAM, which seems to be enough for this small data set. | |
1. Sofware requirements |