Skip to content

Instantly share code, notes, and snippets.

View gadenbuie's full-sized avatar

Garrick Aden-Buie gadenbuie

View GitHub Profile
@geoffjay
geoffjay / pr-stats.sh
Created November 3, 2022 21:07
GitHub GraphQL API query to get some simple PR metrics
#!/bin/bash -e
QUERY='{
search(type: ISSUE, query: "is:pr state:closed repo:clio/grow", first: 100) {
nodes {
... on PullRequest {
title
number
additions
deletions
@moodymudskipper
moodymudskipper / gist:5f227be37c5d9a6b66fe8fa6a2608d32
Last active April 2, 2023 14:25
Custom roxygen2 tags : "tip" example
# You'll also need to add roxygen2 to Suggests in the DESCRIPTION file to satisfy checks
# or just call `usethis::use_package("roxygen2", "Suggests")`
# first we define the tag used in examples in the doc ----------------------------------------------
#' @export
#' @importFrom roxygen2 roxy_tag_parse
roxy_tag_parse.roxy_tag_tip <- function(x) {
roxygen2::tag_markdown(x)
}
@hyperupcall
hyperupcall / settings.jsonc
Last active June 26, 2025 09:47
VSCode config to disable popular extensions' annoyances (telemetry, notifications, welcome pages, etc.)
// I'm tired of extensions that automatically:
// - show welcome pages / walkthroughs
// - show release notes
// - send telemetry
// - recommend things
//
// This disables all of that stuff.
// If you have more config, leave a comment so I can add it!!
{
@adrienbrault
adrienbrault / llama2-mac-gpu.sh
Last active April 8, 2025 13:49
Run Llama-2-13B-chat locally on your M1/M2 Mac with GPU inference. Uses 10GB RAM. UPDATE: see https://twitter.com/simonw/status/1691495807319674880?s=20
# Clone llama.cpp
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
# Build it
make clean
LLAMA_METAL=1 make
# Download model
export MODEL=llama-2-13b-chat.ggmlv3.q4_0.bin

Experimental function to concatenate lots of HTML dependencies' JavaScript files into one single one.

A number of caveats apply:

  1. Strict mode: Many JavaScript files contain "use strict"; at the top of the file. This causes slightly different behavior in the JS engine. If the concatenated JS files have different expectations regarding strict mode, there could be problems.
  2. Hosted scripts: This script does not attempt to concatenate JS files that are hosted on CDNs. Instead, they're left in their original form. It's possible that files we do concatenate, either assume that a CDN-hosted script is loaded before it itself is, or, the opposite. combine_js has a parameter that lets you decide whether to put the combined script first or last, there's no way for us to decide automatically (assuming there's even a right answer).
  3. ES6 modules: We leave type="module" alone, because for ES6 modules the file boundarie
# Instructions:
# - Download LM Studio
# - Download Phi-3 Model (within LM Studio)
# - Load the model into LM Studio
# - Start the Local Server (instructions here: https://lmstudio.ai/docs/local-server)
# Load necessary libraries
library(httr)
library(jsonlite)
library(tidyverse)
@LoueeD
LoueeD / bsky-comments.js
Last active January 24, 2025 04:10
bluesky comments web component - inspired by emilyliu and coryzue
// Moved to a github repo for versioning, the file can now be loaded from a CDN
// <script type="module" src="https://esm.sh/gh/loueed/[email protected]/comments"></script>
class BskyComments extends HTMLElement {
constructor() {
super();
this.attachShadow({ mode: "open" });
this.visibleCount = 3;
this.thread = null;
this.error = null;