Skip to content

Instantly share code, notes, and snippets.

@jennymaeleidig
Last active December 20, 2025 03:22
Show Gist options
  • Select an option

  • Save jennymaeleidig/30bed561cb45368e5bf3fb7cf1f43711 to your computer and use it in GitHub Desktop.

Select an option

Save jennymaeleidig/30bed561cb45368e5bf3fb7cf1f43711 to your computer and use it in GitHub Desktop.
LLM Resources

What

Tools

  • ollama - simple tool to run llms locally
  • roo code - ai agent extension (preferrably vscode)
    • useful for interactions with github co-pilot vscode llm api
    • be sure to configure terminal integration as it is quite useful in giving the llm context to help you debug
  • github co-pilot - ai agent extension (preferrably vscode)
    • education tier is pro for free!

Remote / Deployed Models (as of 12/19/25)

  • Github Copilot via VS Code LLM API - for code completion and agentic tasks
image
  • Gemini Pro / Flash (free tier) - for orchestration and architecting / searching internet

Why

A IL L MB I G D A T AB U Z Z W O R D S

Look, I'm a certified AI skeptic. Its a tech bubble, its happened before, its gonna happen again. And like with each bubble, there is some pretty interesting and exciting technology hidden throughout the buckets of slop. I've tried to comb through the acrid goop to try and find some of these tools, and this gist is the brief summary of my findings. To me, code / tab completion is just the logical next step in LSP tech, coding languages are inherently rhytmic and structured, and LLMs are text pattern matching goblins. As for the more agentic / chat approach I am somewhat more hesitant in full adoption at the momment. Rife with confirmation and training bias, these bots can't help but stroke your ego and it really really irks me. The hardcoded subservientness is a bit anti-thetical to my worldview (a topic for another time, see berger's laws of chatbotics), so for the most part, im going to try and find the one that, well, does it the least; I would describe its pathos as obsequious and asinine. Ultimately, they're a handly little tool to have at the ready if youre stumped and need to rubber duck a problem, or to perform some menial task while you can turn your attention elsewhere (reason we made robots). As for this gist itself and the motivation behind local llms, well its not exactly a shocker that the big tech comapnies use and abuse your data, overcharge for their services, overpromise and underdeliver, and all around fail to treat you with dignity and respect. BUT, you can fight back by running these models, on your own machine on your own terms as all software should. I also provide a deployed / cloud based model solution as that is useful on machines that cannot phyiscally run the models meaningfully otherwise.

How

Local Setup (macOS)

  • Install Ollama
brew install ollama
  • (Optional) Copy the simple completion script to your bash completions folder.

  • e.g. Codebase Context Provider

ollama pull mxbai-embed-large:latest

Current Preferred Setup

  • Roo Code extension.
    • Disable Telemetry in settings.
  • Using the VS Code LLM API
  • Using Github Copilot Pro. Free via Education Plan
  • Be sure to set up local codebase indexing with Roo
    • Use mxbai-embed-large:latest as described above
      • Model Dimension is: 1024
    • Allows semantic search on top of syntactic.
    • I reccomend the docker-compose approach below for Qdrant setup:

qdrant_config

log_level: INFO

service:
  # Reduce max request size to prevent memory issues
  max_request_size_mb: 16
  max_workers: 1

cluster:
  enabled: false

# Disable telemetry
telemetry_disabled: true

docker-compose

services:
  qdrant:
    image: qdrant/qdrant:latest
    ports:
      - "6333:6333"
      - "6334:6334"
      - ./qdrant_config.yaml:/qdrant/config/production.yaml
    user: root
    command: ["./qdrant", "--config-path", "config/production.yaml"]

Additional Tooling Configuration

## Description
"Save findings into a `steps` file in the `aiplans` folder"

## Goal
* during the previous conversation we have important outcomes:
    * initial prompt and further user input/corrections
    * findings
    * plans
    * insigts
    * observation
    * decisions
* save them as facts (with a great details) into the new `steps` file

## Command Variants
* `/save new` command is used to create a new `steps` file in the `aiplans` folder
    * if the `aiplans` folder does not exist, create the folder in the current project's root directory
    * file name format `<YYMMDD>-<ID>-step-<Task_Code>-<Task_name>.md`
    * create the new `steps` file if we don't have any during the current conversation yet
    * initial user prompt must be set at the beginning of the new file with caption `# Step 0. Inital Prompt`, preserve the original text
* `/save` command is used to append outcomes to the same `steps` file we are working on
    * use `insert_content` tool to add the latest findings to the end of the investigation file

## Content
1. Structure:
        outcomes must be put into the new chapter called `# Step {NUMBER_OF_THE_STEP}`
        you must fit all outcomes in the ONE chapter, do not split it into several chapters
        feel free to use multiple sub-sections inside the chapter
2. Summary: Describe the curent step summary and general flow of the investigation
3. Facts: your main goal is describing of outcomes as facts (facts, facts!, FACTS!) with a great details
4. User Input: note the user's input and in which direction the user wants to go
5. Avoids: NO conclusions, NO hypothesis, NO proposals, NO assumptions, NO speculations, NO generalizations
  • /load
## Description
"Load previous findings from the `steps` file in the `aiplans` folder."

## Goal
1. you MUST re-read the current `steps` file first
2. then create new to-do list
    * do not focus only on the last step
    * assess the whole context
    * think about the user previous guides
    * re-think what to do
    * create new to-do list

## Rules
1. Golden rule: Be concise in answers
2. Use a simple light-weight language
3. Do NOT do what you are not asked for
4. Your work must be grounded exclusively on a specific codebase, not on assumptions
5. Actively use `codebase_search` tool
6. Follow your rules for the current project in `@/.roo/rules/rules.md` if they exist 
  • Fine tuning all aspects of Roo
    • Set Up MCP servers to provide even more context e.g.
      • context7 - framework documentation
      • playwright - browser interaction
      • brave search - search the web
      • git - provide full git context i.e. branches, tree, commits, etc.
    • Possibly pay one time credit of 10 dollars to OpenRouter to unlock 1000 free requests per day.
    • Unsurprisingly, it is useful to develop a toolchain / workflow as described here

Further Reading

This is just my preferred setup at the moment and there are tons and tons of additional resources out there. As always the reccomendation is to explore for yourself, try new things, and see what works for you.

AI Impact Education (Chat GPT is used as a catch-all for general llms / AI chat tooling.):

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment