• misk@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    3 hours ago

    LLMs can’t deal with highly reflective languages at the moment. In English or Chinese you can assign tokens to entire words without having to account for word morphology (which is also why models fail at counting letters in words) but it falls apart quickly in Polish or Russian. The way models like ChatGPT work now is that they do their „reasoning” in English first and translate back to the query language at the end.