Add --allow-privileged=true
to:
# kubelet config
sudo vim /var/snap/microk8s/current/args/kubelet
#kube-apiserver config
sudo vim /var/snap/microk8s/current/args/kube-apiserver
Restart services:
Add --allow-privileged=true
to:
# kubelet config
sudo vim /var/snap/microk8s/current/args/kubelet
#kube-apiserver config
sudo vim /var/snap/microk8s/current/args/kube-apiserver
Restart services:
Add --allow-privileged=true
to:
# kubelet config
sudo vim /var/snap/microk8s/current/args/kubelet
#kube-apiserver config
sudo vim /var/snap/microk8s/current/args/kube-apiserver
Restart services:
sysctl -w fs.file-max=12000500
sysctl -w fs.nr_open=20000500
# Set the maximum number of open file descriptors
ulimit -n 20000000
# Set the memory size for TCP with minimum, default and maximum thresholds
sysctl -w net.ipv4.tcp_mem='10000000 10000000 10000000'
/* | |
Small demo to show chaining of API calls. We're using these APIs: | |
https://weathers.co for current weather in 3 cities, and | |
https://api.github.com/search/repositories?q=<term> to search GitHub repos for | |
the 'hottest' city. For example, we'll grab weather for 3 cities, pick the hottest one, | |
then hit GitHub with a search for repos that have that city's name. A bit nonsensical | |
but I just wanted to demonstrate using the results of one API call to populate a second. | |
The problem we're trying to solve is that the API calls are asynch, and one of them | |
(getCityTemperatures) loops through several asynch calls (one for each city), and they |
To remove a submodule you need to:
Originall From: Posted 2015-05-29 http://ubwg.net/b/full-list-of-ffmpeg-flags-and-options | |
This is the complete list that’s outputted by ffmpeg when running ffmpeg -h full. | |
usage: ffmpeg [options] [[infile options] -i infile]… {[outfile options] outfile}… | |
Getting help: | |
-h — print basic options | |
-h long — print more options | |
-h full — print all options (including all format and codec specific options, very long) |
# uname -a | |
# Linux ip-172-31-1-128 5.4.0-1038-aws #40-Ubuntu SMP Fri Feb 5 23:53:34 UTC 2021 aarch64 aarch64 aarch64 GNU/Linux | |
git clone --depth=1 --branch=docker_server_from_ci_builds https://github.com/filimonov/ClickHouse.git | |
cd ClickHouse/docker/server/ | |
docker build . --network host --build-arg single_binary_location="https://builds.clickhouse.tech/master/aarch64/clickhouse" | |
docker image ls | |
docker tag 98f169cda25a altinity/clickhouse-server:21.4.1.6307-testing-arm | |
docker login |
SELECT table, | |
formatReadableSize(sum(bytes)) as size, | |
min(min_date) as min_date, | |
max(max_date) as max_date | |
FROM system.parts | |
WHERE active | |
GROUP BY table |
# You don't need Fog in Ruby or some other library to upload to S3 -- shell works perfectly fine | |
# This is how I upload my new Sol Trader builds (http://soltrader.net) | |
# Based on a modified script from here: http://tmont.com/blargh/2014/1/uploading-to-s3-in-bash | |
S3KEY="my aws key" | |
S3SECRET="my aws secret" # pass these in | |
function putS3 | |
{ | |
path=$1 |