Skip to content

Instantly share code, notes, and snippets.

View paramsiddharth's full-sized avatar
❤️
For the mother's pride, for the times I cried, got to stay alive. I'll survive.

Param Siddharth paramsiddharth

❤️
For the mother's pride, for the times I cried, got to stay alive. I'll survive.
View GitHub Profile
@rohitg00
rohitg00 / llm-wiki.md
Last active April 22, 2026 10:15 — forked from karpathy/llm-wiki.md
LLM Wiki v2 — extending Karpathy's LLM Wiki pattern with lessons from building agentmemory

LLM Wiki v2

A pattern for building personal knowledge bases using LLMs. Extended with lessons from building agentmemory, a persistent memory engine for AI coding agents.

This builds on Andrej Karpathy's original LLM Wiki idea file. Everything in the original still applies. This document adds what we learned running the pattern in production: what breaks at scale, what's missing, and what separates a wiki that stays useful from one that rots.

What the original gets right

The core insight is correct: stop re-deriving, start compiling. RAG retrieves and forgets. A wiki accumulates and compounds. The three-layer architecture (raw sources, wiki, schema) works. The operations (ingest, query, lint) cover the basics. If you haven't read the original, start there.

LLM Wiki

A pattern for building personal knowledge bases using LLMs.

This is an idea file, it is designed to be copy pasted to your own LLM Agent (e.g. OpenAI Codex, Claude Code, OpenCode / Pi, or etc.). Its goal is to communicate the high level idea, but your agent will build out the specifics in collaboration with you.

The core idea

Most people's experience with LLMs and documents looks like RAG: you upload a collection of files, the LLM retrieves relevant chunks at query time, and generates an answer. This works, but the LLM is rediscovering knowledge from scratch on every question. There's no accumulation. Ask a subtle question that requires synthesizing five documents, and the LLM has to find and piece together the relevant fragments every time. Nothing is built up. NotebookLM, ChatGPT file uploads, and most RAG systems work this way.

"""
The most atomic way to train and run inference for a GPT in pure, dependency-free Python.
This file is the complete algorithm.
Everything else is just efficiency.
@karpathy
"""
import os # os.path.exists
import math # math.log, math.exp
@Hegghammer
Hegghammer / working_moltbot_ollama_config.md
Last active April 8, 2026 12:16
Working Clawdbot/Moltbot setup with local Ollama model

Working Clawdbot/Moltbot setup with local Ollama model

[Update 2026-02-02: nemotron-3-nano also performs well on same setup; see comment below]

This is a guide to setting up Clawdbot/Moltbot with a local Ollama model that actually works -- meaning it has good tool use and decent speed. The main requirement is 48GB of VRAM. I have yet to find a model that fits on less than this and still works on Moltbot.

The setup involves creating a tool-tuned variant of qwen2.5:72b and modifying a range of configs in Moltbot. At the end you'll get a local Moltbot instance that can use tools (exec, read, write, web search), read skills, and perform agentic tasks without any cloud API dependencies. On my system I get ~16 t/s and have yet to come across a tool/skill that my bot can't use.

Claude Opus wrote the first draft of this Gist, then I (a human) checked and edited it.

Why I Left AT Protocol

I'm Tree. I think and write about decentralization - analyzing power dynamics, architectural principles, and how systems can preserve or undermine individual autonomy. I've been involved with AT Protocol development and community discussions for some time.

TL;DR: I deleted my atproto account not because of bullying, though that happened. I left because the community's response to my architectural critique demonstrated exactly the power dynamic I was warning about, while simultaneously making it impossible to discuss.

The Technical Argument

Infrastructure should be separate from ideology. This is a foundational principle for resilient decentralized systems.

@barbietunnie
barbietunnie / DecryptDBeaver.java
Last active March 12, 2026 07:03
How to retrieve/decrypt password stored in DBeaver connection
import javax.crypto.Cipher;
import javax.crypto.CipherInputStream;
import javax.crypto.NoSuchPaddingException;
import javax.crypto.SecretKey;
import javax.crypto.spec.IvParameterSpec;
import javax.crypto.spec.SecretKeySpec;
import java.io.ByteArrayInputStream;
import java.io.IOException;
import java.io.InputStream;
import java.nio.file.Files;
@ramonsmits
ramonsmits / HOWTO.md
Last active February 3, 2026 16:01
Install .NET 8 on Raspberry pi
@woss
woss / .env
Last active March 1, 2023 19:15
Remote signer setup
COMPOSE_PROJECT_NAME=remote-signer
ENABLE_API_KEY_SUPPORT=true
APPROVED_API_KEY=( UUID4 or anything really as long as this is exported on the sender side )
@abir-taheer
abir-taheer / instagram-follower-following.js
Last active April 20, 2026 20:47
"This is our community, this is our family, these are our friends." https://www.youtube.com/watch?v=gk7iWgCk14U&t=425s
if (window.location.origin !== "https://www.instagram.com") {
window.alert(
"Hey! You need to be on the instagram site before you run the code. I'm taking you there now but you're going to have to run the code into the console again.",
);
window.location.href = "https://www.instagram.com";
console.clear();
}
const fetchOptions = {
credentials: "include",
@sago35
sago35 / errors.go
Created May 17, 2022 11:52
TinyGo : fmt package for debugging the machine package
// Copyright 2018 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package fmt
import "errors"
// Errorf formats according to a format specifier and returns the string as a
// value that satisfies error.