2026-02-22 22:32:42
It's a hacky way to end the story. Garbage and trite, but that's not the point. The point is that the system prompt explicitly says to give back the original text with improvements. Instead it finished the story from the end.
Why? Because I injected into the conversation and made it "think" that *it* had written the story and was finishing it.
That variant of the story used <start_of_turn> and <end_of_turn> (the model's tokens) instead of <|im_start|>/<|im_end|> which are (if I'm not mistaken) ChatGPT tokens.


