Skip to content

Instantly share code, notes, and snippets.

View lvnilesh's full-sized avatar
💭
🏆 Vibranium Status Level

LV Nilesh lvnilesh

💭
🏆 Vibranium Status Level
View GitHub Profile
@ayushthakur
ayushthakur / understanding-roll-py.ipynb
Last active February 23, 2025 23:10
Understanding roll py.ipynb
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@usrbinkat
usrbinkat / README.md
Last active April 8, 2025 09:42
Ollama + Open-Webui + Nvidia/CUDA + Docker + docker-compose

image

UPDATE: This is tested and working on both Linux and Windows 11 used for LlaMa & DeepSeek

Here's a sample README.md file written by Llama3.2 using this docker-compose.yaml file that explains the purpose and usage of the Docker Compose configuration:

ollama-portal

A multi-container Docker application for serving OLLAMA API.

@btbytes
btbytes / README.md
Last active September 19, 2024 01:08
moby dick in postscript using claude

Asked Claude to:

write a one page pamphlet about moby dick using postscript.

Convert the PS to PDF using:

ps2pdf mobydick.ps
@btbytes
btbytes / README.md
Last active March 21, 2024 16:33
Gist created by gistash

gistash

Stash files on gist.github.com

Installation

$ wget https://gist.githubusercontent.com/btbytes/73d6ddb0cf940e730edf22a60208691b/raw/fbf9efb44ee1c88eaaa3f62d6437737402ea74ce/gistash 
$ chmod +x gistash
@carlosflorencio
carlosflorencio / .aerospace.toml
Created January 29, 2024 21:42
.aerospace.toml
start-at-login = true
accordion-padding = 0
gaps.inner.horizontal = 10
gaps.inner.vertical = 10
after-startup-command = [
'workspace 1', 'layout h_accordion horizontal',
'workspace 2', 'layout h_accordion horizontal',
]
@lvnilesh
lvnilesh / install_gpu_ubuntu_2204.sh
Created January 29, 2024 05:55 — forked from ovrnt/install_gpu_ubuntu_2204.sh
Yam Peleg's Ubuntu 22.04 “CUDA + GPU Drivers + CuDNN and everything else” installer (https://twitter.com/Yampeleg/status/1751823896800583924)
#!/bin/bash
# Verify if GPU is CUDA-enabled
lspci | grep -i nvidia
# Remove previous NVIDIA driver installation
sudo apt-get purge nvidia* -y
sudo apt remove nvidia-* -y
sudo rm /etc/apt/sources.list.d/cuda* -y
sudo apt-get autoremove && sudo apt-get autoclean -y
@pythoninthegrass
pythoninthegrass / kamal.md
Last active October 7, 2024 10:41 — forked from huksley/mrsk.md
kamal - the missing manual

kamal

This documentation adds important additions to the docs for kamal deploy tool (see github.com/basecamp/kamal)

Destination flag

You can use kamal deploy --destination staging

This will read config/deploy.yml and config/deploy.staging.yml files, and also will read .env.staging file if it exists.

@btbytes
btbytes / llm-training.md
Last active April 7, 2025 06:14
Copy of https://rentry.org/llm-training. Updated 2024-03-03

-> lmg-train.png<-

The Novice's LLM Training Guide

->Written by Alpin<- ->Inspired by /hdg/'s LoRA train rentry<- !!!warning This guide is being slowly updated. We've already moved to the axolotl trainer.


[TOC2]

@ChristopherA
ChristopherA / macOS-Preferences-Defaults.sh
Last active December 1, 2024 19:54
macOS Preferences Defaults
#!/usr/bin/env zsh
#===========================================================================
# * INFO
#
# macOS Preferences Defaults
# By Christopher Allen @ChristopherA https://github.com/christophera/
# My most basic macOS preferences, in this gist to make it easy to load on
# multiple machines and VMs. Ideally runs on all macOS versions since Yosemite,