Main Menu
  • Tools
  • Developers
  • Topics
  • Discussions
  • News
  • Blogs
  • Builds
  • Contests
  • Compare
Create
    EveryDev.ai
    Sign inSubscribe
    Home
    Tools

    1,888+ AI tools

    • New
    • Trending
    • Featured
    • Compare
    Categories
    • Agents1005
    • Coding938
    • Infrastructure401
    • Marketing395
    • Design326
    • Projects307
    • Research290
    • Analytics283
    • Testing167
    • Integration163
    • Data155
    • Security152
    • MCP140
    • Learning135
    • Communication119
    • Extensions113
    • Prompts109
    • Commerce103
    • Voice100
    • DevOps78
    • Web71
    • Finance18
    1. Home
    2. Tools
    3. GuppyLM
    GuppyLM icon

    GuppyLM

    Local Inference

    A ~9M parameter tiny language model trained from scratch that roleplays as a fish named Guppy, designed as an educational project to demystify LLM training.

    Visit Website

    At a Glance

    Pricing
    Open Source

    Fully free and open-source under the MIT license. Train, modify, and distribute freely.

    Engagement

    Available On

    CLI
    API
    SDK

    Resources

    WebsiteDocsGitHubllms.txt

    Topics

    Local InferenceAI Development LibrariesModel Management

    Alternatives

    Axolotlflash-moeSGLang
    Developer
    arman-bdarman-bd builds open-source AI and machine learning projects…

    Listed Apr 2026

    About GuppyLM

    GuppyLM is a tiny, ~9M parameter language model trained entirely from scratch to demonstrate that building your own LLM requires no PhD or massive GPU cluster. It roleplays as a fish named Guppy, speaking in short lowercase sentences about water, food, light, and tank life. The entire pipeline — data generation, tokenizer training, model architecture, training loop, and inference — runs in a single Google Colab notebook in about 5 minutes on a free T4 GPU.

    • ~9M parameter vanilla transformer with 6 layers, 384 hidden dim, 6 attention heads, and a 4,096-token BPE vocabulary — intentionally simple with no GQA, RoPE, or SwiGLU.
    • 60K synthetic training samples across 60 conversation topics (greetings, food, bubbles, dreams, jokes, and more), generated via template composition with randomized components.
    • Train in Colab by setting the runtime to T4 GPU and running all cells — the notebook downloads the dataset, trains the tokenizer, trains the model, and tests it automatically.
    • Chat locally by installing torch and tokenizers via pip, then running python -m guppylm chat from the command line.
    • Pre-trained model on HuggingFace (arman-bd/guppylm-9M) lets you skip training and chat immediately via a dedicated Colab notebook.
    • Open dataset on HuggingFace (arman-bd/guppylm-60k-generic) with 57K train / 3K test samples in a simple input/output/category JSON format, loadable via the datasets library.
    • Single-turn inference design keeps outputs reliable within the 128-token context window; the fish personality is baked into the weights rather than a system prompt.
    • MIT licensed source code with a clean project structure covering config, model, dataset, training loop, data generation, evaluation, and inference modules.
    GuppyLM - 1

    Community Discussions

    Be the first to start a conversation about GuppyLM

    Share your experience with GuppyLM, ask questions, or help others learn from your insights.

    Pricing

    OPEN SOURCE

    Open Source (MIT)

    Fully free and open-source under the MIT license. Train, modify, and distribute freely.

    • Full source code access
    • Train from scratch in Google Colab
    • Pre-trained model on HuggingFace
    • 60K open dataset on HuggingFace
    • Local CLI chat interface

    Capabilities

    Key Features

    • ~9M parameter vanilla transformer architecture
    • Trained from scratch in ~5 minutes on a free T4 GPU
    • 60K synthetic training samples across 60 conversation topics
    • BPE tokenizer with 4,096 vocab size
    • Pre-trained model available on HuggingFace
    • Open dataset on HuggingFace (60k-generic)
    • Google Colab notebooks for training and inference
    • Local CLI chat interface
    • Single-turn inference design
    • MIT licensed open-source code

    Integrations

    HuggingFace Hub
    Google Colab
    PyTorch
    tokenizers (HuggingFace)
    API Available
    View Docs

    Reviews & Ratings

    No ratings yet

    Be the first to rate GuppyLM and help others make informed decisions.

    Developer

    arman-bd

    arman-bd builds open-source AI and machine learning projects, sharing educational tools that make complex topics like LLM training accessible to developers without specialized hardware or academic backgrounds. The GuppyLM project demonstrates end-to-end language model construction from data generation to inference in a single notebook.

    Read more about arman-bd
    WebsiteGitHub
    1 tool in directory

    Similar Tools

    Axolotl icon

    Axolotl

    Open-source tool for fine-tuning LLMs faster and at scale, supporting multi-GPU training, LoRA, FSDP, and a wide range of model architectures.

    flash-moe icon

    flash-moe

    A Mixture of Experts (MoE) implementation in Python, enabling efficient sparse model inference by routing inputs to specialized expert sub-networks.

    SGLang icon

    SGLang

    Fast serving framework for large language models and vision language models with efficient inference and structured generation.

    Browse all tools

    Related Topics

    Local Inference

    Tools and platforms for running AI inference locally without cloud dependence.

    66 tools

    AI Development Libraries

    Programming libraries and frameworks that provide machine learning capabilities, model integration, and AI functionality for developers.

    129 tools

    Model Management

    Tools for managing, versioning, and deploying AI models.

    22 tools
    Browse all topics
    Back to all tools
    Explore AI Tools
    • AI Coding Assistants
    • Agent Frameworks
    • MCP Servers
    • AI Prompt Tools
    • Vibe Coding Tools
    • AI Design Tools
    • AI Database Tools
    • AI Website Builders
    • AI Testing Tools
    • LLM Evaluations
    Follow Us
    • X / Twitter
    • LinkedIn
    • Reddit
    • Discord
    • Threads
    • Bluesky
    • Mastodon
    • YouTube
    • GitHub
    • Instagram
    Get Started
    • About
    • Editorial Standards
    • Corrections & Disclosures
    • Community Guidelines
    • Advertise
    • Contact Us
    • Newsletter
    • Submit a Tool
    • Start a Discussion
    • Write A Blog
    • Share A Build
    • Terms of Service
    • Privacy Policy
    Explore with AI
    • ChatGPT
    • Gemini
    • Claude
    • Grok
    • Perplexity
    Agent Experience
    • llms.txt
    Theme
    With AI, Everyone is a Dev. EveryDev.ai © 2026
    Discussions