Skip to content

Instantly share code, notes, and snippets.

View zsarge's full-sized avatar

Zack Sargent zsarge

View GitHub Profile
@adtac
adtac / README.md
Last active June 23, 2025 13:37
Using your Kindle as an e-ink monitor

3.5 fps, Paperwhite 3
@adtac_

step 1: jailbreak your Kindle

mobileread.com is your best resource here, follow the instructions from the LanguageBreak thread

I didn't really follow the LanguageBreak instructions because I didn't care about most of the features + I was curious to do it myself, but the LanguageBreak github repo was invaluable for debugging

@edrisranjbar
edrisranjbar / natas.txt
Last active January 17, 2025 21:39
Natas Hacking Challenge Passwords
natas0:
natas1: 0nzCigAq7t2iALyvU9xcHlYN4MlkIwlq
natas2: TguMNxKo1DSa1tujBLuZJnDUlCcUAPlI
natas3: 3gqisGdR0pjm6tpkDKdIWO2hSvchLeYH
natas4: QryZXc2e0zahULdHrtHxzyYkj59kUxLQ
natas5: 0n35PkggAPm2zbEpOU802c0x0Msn1ToK
natas6: 0RoJwHdSKWFTYR5WuiAewauSuNaBXned
natas7: bmg8SvU1LizuWjx3y7xkNERkHxGre0GS
natas8: xcoXLmzMkoIP9D7hlgPlh9XD7OgLAe5Q
natas9: ZE1ck82lmdGIoErlhQgWND6j2Wzz6b6t
@quad
quad / 0-unnamed-architecture.md
Last active April 14, 2024 05:45
What is this architecture called?

What is this architecture called?

I rarely see the classical three-tier architecture in the wild; I frequently see a different architecture.

I don't know this architecture's name. Do you?

The Three-Tier Architecture

The "three-tier architecture" has been the reference pattern for Internet services:

@LiquidityC
LiquidityC / Makefile
Last active March 21, 2025 07:48
Generic drop in Makefile
VERSION = \"1.0.0\"
PREFIX ?= out
INCDIR = include
SRCDIR = src
LANG = c
OBJDIR = .obj
MODULE ?= binary_name
CC ?= gcc
@kbdharun
kbdharun / downgrade-fedora-kernel.md
Last active September 1, 2024 20:51
Downgrading Kernel in Fedora

Steps

  • List kernel versions:
dnf list kernel --showduplicates
  • Now, install/modify/reinstall the latest kernel:

Mara's statement on the retraction of ThePhd's RustConf keynote

I was one of the people who didn't vote for ThePhd's keynote; originally because I simply preferred another promising candidate. Later, after hearing technical concerns from an expert that I mostly agreed with, also because of the topic, although I must admit I am no expert on the topic.

The candidate I voted for originally got a few approving comments at first, but ended up being mostly ignored later, mostly because of our lack of process and proper voting.

When it was brought up in the leadership chat that 'there are concerns', I focused on the talk itself rather than focusing on the process failure. That was a mistake, for which I apologize. I am not an expert on this topic, and should not have rushed myself into talking about things outside my expertise, even under pressure.

In another situation this could have been a minor mistake with no consequences, just one of several opinions in a discussion, but in this situation it became part

@rain-1
rain-1 / llama-home.md
Last active June 24, 2025 11:12
How to run Llama 13B with a 6GB graphics card

This worked on 14/May/23. The instructions will probably require updating in the future.

llama is a text prediction model similar to GPT-2, and the version of GPT-3 that has not been fine tuned yet. It is also possible to run fine tuned versions (like alpaca or vicuna with this. I think. Those versions are more focused on answering questions)

Note: I have been told that this does not support multiple GPUs. It can only use a single GPU.

It is possible to run LLama 13B with a 6GB graphics card now! (e.g. a RTX 2060). Thanks to the amazing work involved in llama.cpp. The latest change is CUDA/cuBLAS which allows you pick an arbitrary number of the transformer layers to be run on the GPU. This is perfect for low VRAM.

  • Clone llama.cpp from git, I am on commit 08737ef720f0510c7ec2aa84d7f70c691073c35d.
@rain-1
rain-1 / LLM.md
Last active June 28, 2025 14:59
LLM Introduction: Learn Language Models

Purpose

Bootstrap knowledge of LLMs ASAP. With a bias/focus to GPT.

Avoid being a link dump. Try to provide only valuable well tuned information.

Prelude

Neural network links before starting with transformers.

@NathanBaulch
NathanBaulch / main.go
Last active July 11, 2024 20:27
Optimized solution to APS #038 unique words with unique letters
package main
import (
"bufio"
"fmt"
"io"
"math"
"os"
"runtime"
"sort"