Skip to content

Instantly share code, notes, and snippets.

View andrewssobral's full-sized avatar
🔴
I may be very slow to respond.

Andrews Cordolino Sobral andrewssobral

🔴
I may be very slow to respond.
View GitHub Profile

LLM Wiki

A pattern for building personal knowledge bases using LLMs.

This is an idea file, it is designed to be copy pasted to your own LLM Agent (e.g. OpenAI Codex, Claude Code, OpenCode / Pi, or etc.). Its goal is to communicate the high level idea, but your agent will build out the specifics in collaboration with you.

The core idea

Most people's experience with LLMs and documents looks like RAG: you upload a collection of files, the LLM retrieves relevant chunks at query time, and generates an answer. This works, but the LLM is rediscovering knowledge from scratch on every question. There's no accumulation. Ask a subtle question that requires synthesizing five documents, and the LLM has to find and piece together the relevant fragments every time. Nothing is built up. NotebookLM, ChatGPT file uploads, and most RAG systems work this way.

@alexlitz
alexlitz / tiny_adder_submission_autoregressive_gen.py
Last active March 12, 2026 22:30
Tiny Adder Autoregressive
#!/usr/bin/env python3
"""
TinyAdder: 36-parameter hand-crafted transformer for 10-digit addition.
Parameter counting:
- Identity mappings (direct copy): 0 params
- Broadcast (1 value to N outputs): 1 param
- Distinct values: count each
"""
import torch
"""
Dynamic NanoGPT Adder: 130 params + 0 buffers
transformer.wte.A [10, 1] = 10
transformer.wte.B [1, 4] = 4
transformer.h.0.attn.c_attn.weight [12, 4] = 48
transformer.h.0.attn.c_proj.weight [4, 4] = 16
transformer.h.0.mlp.c_fc.weight [4, 4] = 16
transformer.h.0.mlp.c_fc.bias [4] = 4
transformer.h.0.mlp.c_proj.u [4, 1] = 4
"""
The most atomic way to train and run inference for a GPT in pure, dependency-free Python.
This file is the complete algorithm.
Everything else is just efficiency.
@karpathy
"""
import os # os.path.exists
import math # math.log, math.exp
@neudinger
neudinger / trim.cpp
Last active November 24, 2025 18:20
#include <ranges>
#include <print>
#include <cctype>
#include <string>
#include <string_view>
using namespace std::literals;
using namespace std::string_view_literals;
@lodestone-rock
lodestone-rock / ramtorch_example.py
Created November 5, 2025 02:32
ramtorch single node multi gpu example
import os
import wandb
from tqdm import tqdm
import argparse
import torch
import torchvision
import torch.nn as nn
import torch.nn.functional as F
from torch.utils.data import DataLoader
@rekram1-node
rekram1-node / settings.json
Created October 20, 2025 22:30
OpenCode agent server (zed)
{
"agent_servers": {
"OpenCode": {
"command": "opencode",
"args": ["acp"]
}
}
}
@glennmatlin
glennmatlin / claude_code_hooks_for_uv.md
Last active April 24, 2026 11:13
Claude Code hooks for working with `uv`

Claude Code Hooks for working with uv

by Glenn Matlin / glennmatlin on all socials

What This Does

Prevents Claude Code from using pip, python, pytest, etc. directly in projects that use uv for Python environment management. Instead, Claude is guided to use uv run, uv add, etc.

Smart detection:

  • Only activates in directories with pyproject.toml
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
import os
from google import genai
from google.genai import types
client = genai.Client(api_key=os.getenv("GEMINI_API_KEY","xxx"))
# Repalce with the youtube url you want to analyze
youtube_url = "https://www.youtube.com/watch?v=RDOMKIw1aF4"
# Prompt to analyze and summarize the Youtube Video