Skip to content

Instantly share code, notes, and snippets.

@alexclear
alexclear / gist:b291787701026e28910f
Last active August 29, 2015 14:16
LeoFS node recovery issue(?)
When I start "recover node <storage_node_name>" message of type QUEUE_ID_RECOVERY_NODE seems to be
blocking processing of other messages (mainly, QUEUE_ID_PER_OBJECT ones).
I tried to visualize message processing using StatsD:
http://ns2.1888.spb.ru/2015-02-24-024731_1366x768_scrot.png
A graph at the lower right corner depicts calls to
handle_call({consume, ?QUEUE_ID_RECOVERY_NODE, MessageBin}).
It is my understanding that recover_node_callback(Node) should generate a lot
of QUEUE_TYPE_PER_OBJECT messages synchronously for a single QUEUE_ID_RECOVERY_NODE
message but the graph shows that QUEUE_ID_RECOVERY_NODE messages are constantly processed.
@psayre23
psayre23 / gist:c30a821239f4818b0709
Last active March 31, 2025 12:58
Runtime Complexity of Java Collections
Below are the Big O performance of common functions of different Java Collections.
List | Add | Remove | Get | Contains | Next | Data Structure
---------------------|------|--------|------|----------|------|---------------
ArrayList | O(1) | O(n) | O(1) | O(n) | O(1) | Array
LinkedList | O(1) | O(1) | O(n) | O(n) | O(1) | Linked List
CopyOnWriteArrayList | O(n) | O(n) | O(1) | O(n) | O(1) | Array
/**
* Written by Gil Tene of Azul Systems, and released to the public domain,
* as explained at http://creativecommons.org/publicdomain/zero/1.0/
*/
package org.HdrHistogram;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicLongFieldUpdater;
import java.util.concurrent.locks.ReentrantLock;
@joyrexus
joyrexus / README.md
Created March 27, 2014 18:17
groupby and countby for python

groupby and countby for python

Python has the standard methods for applying functions over iterables, viz. map, filter, and reduce.

For example, we can use filter to filter some numbers by some criterion:

even = lambda x: x % 2 is 0
odd  = lambda x: not even(x)
data = [1, 2, 3, 4]
@kevin-smets
kevin-smets / iterm2-solarized.md
Last active May 17, 2025 20:49
iTerm2 + Oh My Zsh + Solarized color scheme + Source Code Pro Powerline + Font Awesome + [Powerlevel10k] - (macOS)

Default

Default

Powerlevel10k

Powerlevel10k

@grantcarthew
grantcarthew / Connect-Telnet.ps1
Last active June 1, 2022 13:48
A full graceful telnet client using PowerShell and the .NET Framework. http://uglygizmo.blogspot.com.au/
<#
.SYNOPSIS
A full graceful telnet client using PowerShell and the .NET Framework.
.DESCRIPTION
This script was made with a view of using it to have full control over the text
stream for automating Cisco router and switch configurations.
.PARAMETER TelnetHost
The address of the server or router hosting the telnet service.

Moved

Now located at https://github.com/JeffPaine/beautiful_idiomatic_python.

Why it was moved

Github gists don't support Pull Requests or any notifications, which made it impossible for me to maintain this (surprisingly popular) gist with fixes, respond to comments and so on. In the interest of maintaining the quality of this resource for others, I've moved it to a proper repo. Cheers!

@wacko
wacko / gist:5577187
Last active July 13, 2024 00:48
SSH between Mac OS X host and Virtual Box guest

On Mac OS (host):

Shutdown your VM and do:

VirtualBox > Settings > Network > Add (you will get vboxnet0)

On a terminal ifconfig will show you new interface vboxnet0

VM's Settings > System > check "Enable I/O APIC." VM's Settings > Network > Adapter 2 > host-only vboxnet0

@nimbus154
nimbus154 / dictionary.py
Created April 21, 2013 23:54
An example of how to write functional tests for a RESTful API using the Bottle microframework.
from bottle import get, run, request, post, Bottle, abort, error, response, debug, redirect
# This is a dictionary endpoint. It retrieves definitions for words.
# You can also add words to the dictionary.
# this allows our bottle application to be accessible outside this file
app = Bottle()
dictionary = {
"lugubrious": "extremely sad",
# you can make a text file of request times (in ms, one number per line) and import it here, or you can use a probability distribution to simulate request times (see below where setting req_durations_in_ms)
# rq = read.table("~/Downloads/request_times.txt", header=FALSE)$V1
# argument notes:
# parallel_router_count is only relevant if router_mode is set to "intelligent"
# choice_of_two, power_of_two, and unicorn_workers_per_dyno are only relevant if router_mode is set to "naive"
# you can only select one of choice_of_two, power_of_two, and unicorn_workers_per_dyno
run_simulation = function(router_mode = "naive",
reqs_per_minute = 9000,