To the best of my knowledge, back-propagation IS learning, whether it’s happening in a neural-net on a chip, or whether we’re doing it, through feedback, & altering our understanding ( so both hard-logic & our wetware use the method for learning, though we use a rather sloppy implimentation of it. )
& altering the relative-significances of concepts IS learning.
( I’m not commenting on whether the new-relation-between-those-concepts is wrong or right, only on the mechanism )
so, I can’t understand your position.
Please don’t deem my comment worthy of answering: I’m only putting this here for the record, is all.
Everybody can downvote my comment into oblivion, & everything in the world’ll still be fine.
dsemy@lemm.ee 5 months ago
This is a major issue I have with basically anyone who talks about current “AI” systems - they’re clearly not even close to AI, as they require an extreme amount of energy and data to perform tasks which would be trivial to an actual brain. They seem to lack any ability to comprehend their input, only mimicking it through brute force, which is only feasible since computers got fast enough and we can currently keep up with the energy demands.
GenderNeutralBro@lemmy.sdf.org 5 months ago
AI does not mean artificial brain or anything similar. It’s a very broad term that’s been in use for about 70 years now.
Pac Man has AI.
dsemy@lemm.ee 5 months ago
Obviously I’m not referring to that, but to what large tech companies call AI. And they are in fact trying to convince people these AI systems they are developing will soon be clever enough to be considered general AI.