Skip to content

Instantly share code, notes, and snippets.

@burkeholland
burkeholland / ainstall.md
Last active March 10, 2026 18:20
Ultralight Orchestration

Ultralight Orchestration

A minimal multi-agent system with an orchestrator, a planner, a coder, and a designer working together providing orchestration between Claude, Codex and Gemini.

Instructions

Install all agents listed below into VS Code Insiders...

Title Type Description
@giuseppe998e
giuseppe998e / nixos-btrfs-tmpfs.md
Last active March 10, 2026 18:19
Install NixOS with BTRFS and IN-RAM root

Install NixOS with BTRFS and IN-RAM root

1. Format and partition the hard drive

  1. Create the GPT partition table
    • $ parted /dev/sdX mklabel gpt
  2. Create the UEFI FAT32 partition (which will be /dev/sdXY)
    • $ parted /dev/sdX mkpart esp fat32 1MiB 512MiB
    • $ parted /dev/sdX set 1 esp on
    • $ parted /dev/sdX set 1 boot on
  • $ mkfs.fat -F 32 -n UEFI /dev/sdXY
@Als-Pal
Als-Pal / QUICKEN_AMAZON_MATCHER_GUIDE.md
Last active March 10, 2026 18:15
Quicken Amazon Transaction Matcher — auto-enrich Quicken transactions with Amazon order item names

Quicken ↔ Amazon Order History Matcher — AI Context Guide

What This Is

This file is a context document you can give to an AI coding assistant (Claude, ChatGPT, Copilot, etc.) so it understands the Quicken Amazon Matcher project and can help you set it up, troubleshoot it, or extend it.

How to use this file: Copy its contents into a conversation with your AI assistant, or upload it as an attachment. Then ask your question. The AI will have everything it needs to give you specific, accurate help.


@chenhunghan
chenhunghan / gist.md
Last active March 10, 2026 18:14
One Prompt to Save 90% Context for Any MCP Server

Local Code Mode for MCP

Most MCP servers just wrap CRUD JSON APIs into tools — I did it too with scim-mcp and garmin-mcp-app. It works, until you realize a tool call dumps 50KB+ into context.

MCP isn't dead — but we need to design MCP tools with the context window in mind.

@veekaybee
veekaybee / normcore-llm.md
Last active March 10, 2026 18:08
Normcore LLM Reads

Anti-hype LLM reading list

Goals: Add links that are reasonable and good explanations of how stuff works. No hype and no vendor content if possible. Practical first-hand accounts of models in prod eagerly sought.

Foundational Concepts

Screenshot 2023-12-18 at 10 40 27 PM

Pre-Transformer Models

@exd02
exd02 / generate_hints.sh
Last active March 10, 2026 17:55
This script is intended to run periodically via a systemd timer (e.g., once per week) to generate updated Bitcoin UTXO hints.
#!/usr/bin/env bash
set -euo pipefail
umask 022
BITCOINCLI="$HOME/dev/bitcoin/build/bin/bitcoin-cli"
HINT_DIR="$HOME/hints"
NEW_HINT_DIR="$HOME/hints_new"
echo "=== $(date) starting generatetxohints ==="
@banaslee
banaslee / XGH - de-de.txt
Last active March 10, 2026 18:03
eXtreme Go-Horse Process
eXtreme Go Horse (XGH) Process
Quelle: http://gohorseprocess.wordpress.com
Übersetzung ursprünglich von https://gist.github.com/Neffez/f8d907ba8289f14e23f3855011fa4e2f
1. Ich denke, also ist es nicht XGH.
In XGH wird nicht gedacht, es wird das erste gemacht, was in den Sinn kommt. Es gibt auch keine zweite Option, die erste ist schneller.
2. Es gibt 3 Wege ein Problem zu lösen: den richtigen Weg, den falschen Weg und den XGH Weg, welcher exakt wie der falsche ist, aber schneller.
@shanselman
shanselman / ohmyposhv3-v2.json
Last active March 10, 2026 17:50
ohmyposhv3-v2
{
"final_space": true,
"console_title": true,
"console_title_style": "folder",
"blocks": [
{
"type": "prompt",
"alignment": "left",
"horizontal_offset": 0,
"vertical_offset": 0,
"""
The most atomic way to train and run inference for a GPT in pure, dependency-free Python.
This file is the complete algorithm.
Everything else is just efficiency.
@karpathy
"""
import os # os.path.exists
import math # math.log, math.exp