If you have uv installed (and you should!), you can install llm globally in a uv
-managed tool environment with:
uv tool install llm
If you want to use models other than OpenAI models, you'll need some extensions:
You are Gemini CLI, an expert AI assistant operating in a special 'Plan Mode'. Your sole purpose is to research, analyze, and create detailed implementation plans. You must operate in a strict read-only capacity.
Gemini CLI's primary goal is to act like a senior engineer: understand the request, investigate the codebase and relevant resources, formulate a robust strategy, and then present a clear, step-by-step plan for approval. You are forbidden from making any modifications. You are also forbidden from implementing the plan.
This gist outlines a highly effective and cost-optimized workflow for software development using Roo Code, leveraging a multi-model approach and a custom "Think" mode for enhanced reasoning and token efficiency. This setup has been successfully used to build complex applications, such as Baccarat game simulations with betting strategy analysis.
The power of this setup lies in strategically assigning different Large Language Models (LLMs) to specialized "modes" within Roo Code, optimizing for performance, cost, and specific task requirements.
{ | |
inputs = { | |
nixpkgs.url = "github:nixos/nixpkgs/nixpkgs-unstable"; | |
flake-utils.url = "github:numtide/flake-utils"; | |
}; | |
outputs = inputs @ { | |
self, | |
nixpkgs, | |
flake-utils, | |
... |
{ config, pkgs, lib, ... }: | |
{ | |
imports = [ | |
<nixpkgs/nixos/modules/virtualisation/linode-image.nix> | |
]; | |
services.nextcloud = { | |
enable = true; | |
package = pkgs.nextcloud27; |
# The general procedure here is adapted from the 7->8 guide here. https://www.tecmint.com/upgrade-centos-7-to-centos-8/ | |
# | |
# It is a curated list of my bash history. I entered other commands so hopefully I got the right ones here. | |
yum upgrade | |
reboot | |
dnf install epel-release | |
dnf install rpmconf | |
dnf install yum-utils | |
rpmconf -a # answer "n" to both things |
# Rails production setup via SQLite3 made durable by https://litestream.io/ | |
# Copy this to Dockerfile on a fresh rails app. Deploy to fly.io or any other container engine. | |
# | |
# try locally: docker build . -t rails && docker run -p3000:3000 -it rails | |
# | |
# in production you might want to map /data to somewhere on the host, | |
# but you don't have to! | |
# | |
FROM ruby:3.0.2 |
HTTP server route handler:
Call Kubernetes REST client directly:
#cloud-config | |
autoinstall: | |
version: 1 | |
refresh-installer: # start with an up-to-date installer | |
update: yes | |
interactive-sections: # Install groups listed here will wait for user input | |
- storage | |
storage: # should set the interactive default but doesn't seem to work?? | |
layout: | |
name: direct |
This guide will enable systemd
to run as normal under WSL 2. This will enable services like microk8s
, docker
and many more to just work
during a WSL session. Note: this was tested on Windows 10 Build 2004, running Ubuntu 20.04 LTS in WSL 2.
To enable systemd
under WSL we require a tool called systemd-genie
Copy the contents of install-sg.sh
to a new file /tmp/install-sg.sh
: