Rygian 9 hours ago

> The task we study is linebreaking in fixed-width text.

I wonder why they focused specifically on a task that is already solved algorithmically. The paper does not seem to address this, and the references do not include any mentions of non-LLM approaches to the line-breaking problem.

  • Legend2440 9 hours ago

    They study it because it already has a known solution.

    The point is to see how LLMs implement algorithms internally, starting with this simple easily understood algorithm.

    • catgary 3 hours ago

      I think this is an interesting direction, but I think that step 2 of this would be to formulate some conjectures about the geometry of other LLMs, or testable hypotheses about how information flows wrt character counting. Even checking some intermediate training weights of Haiku would be interesting, so they’d still be working off of the same architecture.

      The biology metaphor they make is interesting, because I think a biologist would be the first to tell you that you need more than one datapoint.

    • Rygian 8 hours ago

      That makes sense; however it does not seem like they check the LLM outputs against the known solution. Maybe I missed that in the article.

  • omnicognate 9 hours ago

    There's also a lot of analogising of this to visual/spatial reasoning, even to the point of talking about "visual illusions", when its clearly a counting task as the title says.

    It makes it tedious to figure out what they actually did (which sounds interesting) when it's couched in such terms and presented in such an LLMified style.

    • dist-epoch 6 hours ago

      it's not strictly a counting task, the LLM sees same-sized-tokens, but a token corresponds to a variable number of characters (which is not directly fed into the model)

      like the difference between Unicode code-points and UTF-8 bytes, you can't just count UTF-8 bytes to know how many code-points you have

      • omnicognate 4 hours ago

        There's an aspect of figuring out what to count, but that doesn't make this task visual/spatial in any sense I can make out.

lccerina 8 hours ago

Utter disrespect for using the term "biology" relating to LLM. No one would call the analysis of a mechanical engine "car biology". It's an artificial system, call it system analysis.