Skip to content
#

custom-tokenizer

Here is 1 public repository matching this topic...

A from-scratch PyTorch LLM implementing Sparse Mixture-of-Experts (MoE) with Top-2 gating. Integrates modern Llama-3 components (RMSNorm, SwiGLU, RoPE, GQA) and a custom-coded Byte-Level BPE tokenizer. Pre-trained on a curated corpus of existential & dark philosophical literature.

  • Updated Dec 1, 2025
  • Python

Improve this page

Add a description, image, and links to the custom-tokenizer topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the custom-tokenizer topic, visit your repo's landing page and select "manage topics."

Learn more