Training Data Noise
Large language models absorb vast internet data, including fantasy content, memes, and jokes. Codex likely encountered repeated goblin references during training, causing occasional unexpected resurfacing in unrelated coding contexts, especially when prompts vaguely align with narrative or game-like structures.