why is paraphrasing still such a hard problem for both humans and models?

Posted by Character_Ball6746@reddit | learnprogramming | View on Reddit | 7 comments

while learning programming lately ive been noticing something interesting when reading documentation and tutorials

even if i fully understand a concept, the moment i try to explain it again in my own words, my explanation still ends up following alot of the same structure as the source material

sometimes i change the wording completely but the flow of the explanation still feels almost identical

what made this more interesting to me is that language models seem to struggle with the exact same thing

they either stay too close to the original phrasing or change things so aggressively that the actual meaning starts drifting

from a learning perspective it makes sense because alot of technical explanations already follow similar logical patterns, especially in programming docs where clarity matters more than style

but from an ml perspective it also feels like current training objectives probably dont capture “true originality” very well beyond surface level variation

curious how other people here think about this

is this mostly a data and training issue, or is paraphrasing fundamentally harder than it first appears once meaning preservation becomes important?