- cross-posted to:
- technology@lemmy.world
- technology@lemmy.world
- cross-posted to:
- technology@lemmy.world
- technology@lemmy.world
cross-posted from: https://lemmy.ml/post/2811405
"We view this moment of hype around generative AI as dangerous. There is a pack mentality in rushing to invest in these tools, while overlooking the fact that they threaten workers and impact consumers by creating lesser quality products and allowing more erroneous outputs. For example, earlier this year America’s National Eating Disorders Association fired helpline workers and attempted to replace them with a chatbot. The bot was then shut down after its responses actively encouraged disordered eating behaviors. "
I don’t care. It doesn’t matter, so I didn’t check. Your reading comprehension is still, in fact, the issue, since you didn’t understand that the “learned” vs “programmed” distinction I had referred to is completely relevant to your post.
That’s what learning is. The fact that it can construct syntactically and semantically correct, relevant responses in perfect English means that it has a highly developed inner model of many things we would consider to be abstract concepts (like the syntax of the English language).
This is wrong. It is obvious and irrefutable that it models sophisticated approximations of abstract concepts. Humans are literally no different. Humans who consider themselves to understand a concept can obviously misunderstand some aspect of the concept in some contexts. The fact that these models are not as robust as that of a human’s doesn’t mean what you’re saying it means.
This is a meaningless point, you’re thinking at the wrong level of abstraction. This argument is equivalent to “a computer cannot convey meaningful information to a human because it simply activates and deactivates bits according to simple rules.” Your statement about an implementation detail says literally nothing about the emergent behavior we’re talking about.