Main Page
Welcome to Game in the Brain wiki - where I will try to document as much knowledge he can as genetic dementia and responsibilities make me forget so much of what i learned and could do.
How to Use Media Wiki
Consult the User's Guide for information on using the wiki software.
- Creating Pages
- Test
- Creating Books
Open Source Tutorials and Notes
Open source Tutorials and Notes is where I will compile
- MediaWiki Set up Guide (Portainer and Docker) 251215
- Ladybird-251231
- Resume Renamer 260120
- Erpnextv15-SSH-setup-241111
- WireguardVPNgui-260123
- WatchLater-deletion-Bookmark-260124
- Local Pocket TTS 260124
Minecraft Bedrock-Java Self-hosted with VM proxy-260124DE-LINKED. VELOCITY has a SPAMBOT- Lean Backwards Analysis and Action 260112
- Save Point Technique 260126
- Hosting static websites from gemini 260128
- GitHub Re-Connection & Workflow 260202
- Docker Setup Win and Linux 260203
Under Heaven
This is the default Setting of Game in the Brain - where it goes into a detailed look at the technology tree and the many setbacks of humanity before it becomes a Cis-Lunar civilization in 2100, and making it an off-Earth Civilization by 2200, and then a Interstellar civilization in 2300. A Traveller-like or Cepheus Engine Like Civilization would be 2500 to 3000 AD.
- Background of Under Heaven (placeholder)
- Technology Levels and Timeline (placeholder)
- Mneme World Generation Version 2 (placeholder)
- Mneme Ship and Vehicle Design (placeholder)
- Mneme Character Creation (placeholder)
- Mneme Fantasy Character Creation
Open PVE CGG
Open PvE Card System (OPCS) is an open-source, cooperative card game framework where players face scenarios they design instead of each other. Cards are proxy-friendly, scarcity-free, and skill-driven. Power comes from strategy, coordination, and accumulated consequences—not purchases, rarity, or power creep.
For Research
These are Open Source Leads - technologies I have to explore and learn to set up.
- https://github.com/xichen1997/opencode - OpenCode is a fully-featured, open-source alternative to Claude Code that runs entirely on your local machine using Ollama. It provides a powerful CLI interface for AI-assisted coding, code analysis, and project management while ensuring complete privacy and security.
- https://github.com/ggml-org/llama.cpp The need to study Sharding (the ability to split a model amongst GPUs relying on incredibly fast networking - 2.5-25GB to make the throughput seemless) and Parallel Decoding (the ability to Split jobs among resources)
- https://github.com/Talnz007/VulkanIlm VulkanIlm is a Python-first wrapper and CLI around llama.cpp's Vulkan backend that brings fast local LLM inference to AMD, Intel, and NVIDIA GPUs — no CUDA required. Built for developers with legacy or non-NVIDIA hardware.
- https://modelscope.ai/home
- https://github.com/MooreThreads