Discovered: Jul 22, 2024 11:16 The serious science of trolling LLMs - lcamtuf’s thing <– QUOTE: In other words, the LLM business is to some extent predicated on deception; we are not supposed to know where the magic ends and where cheap tricks begin. The vendors’ hope is that with time, we will reach full human-LLM parity; and until then, it’s OK to fudge it a bit. From this perspective, the viral examples that make it patently clear that the models don’t reason like humans are not just PR annoyances; they are a threat to product strategy. ... Far from being a waste of time, internet trolling is becoming a legitimate scientific pursuit. When a model aces a human benchmark, it’s hard to know how much of this can be credited to reasoning and how much of it boils down to recall from the training data set. It’s when it fails at a simple task that we know what the limitations are — and trolls are the torch-bearers of this new englightenment.

Leave a comment on github