Dennis Hackethal’s Blog
My blog about philosophy, coding, and anything else that interests me.
Tweets
An archive of my tweets and retweets through . They may be formatted slightly differently than on Twitter. API access has since gotten prohibitively expensive – I don't know whether or when I'll be able to update this archive.
But in case I will, you can subscribe via RSS – without a Twitter account. Rationale
No. It is evolution that creates knowledge generally, both genetic and human.
BTW, these things need to be explained on the applicable level of emergence. Intelligence is a property of software, not of some underlying physical structures in the brain.
But knowledge does exist in the brain in the form of the structure of cortical columns; plus mechanisms of how to associate them given certain sense data, etc.
Knowledge - of any kind - is created by evolution: variation alternating with criticism.
Indeed. So biological evolution creates the genetic knowledge of how to do the wiring in such a way.
If biological evolution explains the origin of genetic knowledge, why should an evolutionary algorithm running on the brain not explain the origin of human knowledge?
In principle, but I agreed that it would be unlikely to happen.
Anyway, okay, so let's say those components are stored in cortical columns. Where does the knowledge of how to store stimuli in cortical columns come from?
I didn’t say blind people can see color.
Okay, so let’s say those components come to us through the senses. How?
Faith has nothing to do with it.
I had already granted that it would be extremely unlikely to think of dogs in that situation, but that's incidental.
You seem to claim that knowledge comes to us through the senses. Yes?
- For example, that it has legs, a nose, etc.
- Because in the sense you mean, red is a quale. We don't understand qualia, so I don't know.
I'm familiar with the argument you're making. Brain in a vat. I could think all kinds of things about dogs I want. Anything thinkable I can think.
If you were to say that I cannot predict the quale of seeing a dog, then I'm with you.
I'm in principle capable to have any thoughts about cars and dogs I like. I would be extraordinarily unlikely to have them, though.
@HeuristicWorld
As a preprogrammed emotional response, sure, but no associated quale of suffering.
@HeuristicWorld
I think there is no animal behavior that cannot be explained in terms of genetically given algorithms that just need to be executed; that leaves no room for creativity.
If consciousness, suffering etc arise out of creativity and animals are not creative, they are not conscious.
As if the bastardization of the term "AI" had not gone far enough, folks are now starting to bastardize the term "AGI".
If this spreads, we will soon need to find a new term to talk about the real thing again. twitter.com/markcannon5/st…
You're building a shape recognition algorithm. That's not AGI, even if it can recognize all kinds of shapes.
Well, we have a good explanations of why it would be a bad idea to unilaterally disarm.
Using justificationism against justificationists is an interesting approach, however. But does that not change your yardstick for what you consider real? I.e. a good explanation vs "proof"?
Burden of proof is justificationist. My link refutes Mark's points. He now needs to either explain why it does not, or refute the link.
@sciencemagazine @ScienceCareers
This isn't science. This is social justice nonsense.
"Knowledge of how to use the senses is neither encoded genetically, nor can it possibly come from the senses. So where does it come from?"
About the folly of empiricism and the recovery from blindness:
E.g. stoning a woman to death because she dared take off her hijab makes me cringe; not because the perpetrators are ignorant - we are all infinitely (though unequally) ignorant - but because the ideas behind it are deeply false (and in this case, deadly).
Ah, it's not the ignorance per se that's cringeworthy (by some criterion): it's the badly mistaken idea. Some ideas are worse than others in terms of the damage they cause.
Some ideas are cringeworthy, but we may all have different thresholds.
People like Feynman, so they rush to his defense. Ad hominem. And unnecessary: he would have liked to know he was wrong so that he may improve.
Wheeler (both Feynman's thesis advisor and David's boss) knew Popperian epistemology well and may have introduced Feynman to Popper, but it isn't clear.
I know that you know who Wheeler was, but including it here in case others read it and aren't aware of the connections.
Yeah. Quote from David:
[...] I happened to mention Popper in the one conversation I had with Feynman, sometime in the 80s, and he did not say "who's that?" but replied meaningfully to the point.
Feynman seemed to show good understanding of Popperian concepts.
Yeah, he had an aversion to philosophy generally. But he read Popper, or at least he was familiar with his philosophy.
I enjoy Feynman, too, btw. It’s not a personal thing. He made a couple of simple but important mistakes. Inductivist mistakes, no less.
Though I should add that I don't know what he means by "definite" theory. One that isn't vague? Or perhaps one that doesn't require endless computation, as he explains later? Not sure.
Alas, even he made mistakes in his understanding of the creation of knowledge: "[W]e always try to guess the most likely explanation." cringe
Or, less obviously, a bit later: "You can always prove any definite theory wrong."
Sorry? I agree with the first sentence, but lost you after that.
I don’t think so, some wrong answers may still contain truth. Eg Christianity’s commandment not to kill.
From history we know all too well that some were enslaved even though they were much more obviously people than AGIs.
Not to mince words here, but it's the universality of people that implies that everybody is qualitatively equal.
All our laws should apply, but I fear they won't because most won't realize that AGIs are people.
Yes; horribly, that's what companies like @OpenAI and @DeepMindAI seem to be after.
No such thing as "advanced" btw, they are all people just the same.
That's just a map of US airline routes over South America, no?
@ReachChristofer @ToKTeacher @FallingIntoFilm @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
Thank you, good point... not sure. Interestingly, whenever subconscious problem solving is successful, the solution does suddenly jump into consciousness (eg shower thoughts etc). So maybe it's something about the correction part of error correction...
@ToKTeacher @FallingIntoFilm @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
... until I ride the bike completely subconsciously. Perhaps consciousness is either strongly correlated with error correction, or it may even be error correction.
@ToKTeacher @FallingIntoFilm @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
Lately I have wondered if one is aware/conscious of wherever one is trying to detect errors. I learn to ride a bike: very conscious of it, I make mistakes all the time. Then gradually as I iron out the mistakes I grow less conscious of it...
@jamessseattle @ToKTeacher @Crit_Rat
(error correction being the primary ingredient of intelligence)
@jamessseattle @ToKTeacher @Crit_Rat
It can't be analog btw because error correction can only happen in digital systems.
And again, a single Turing machine can simulate multiple Turing machines, so parallelism is incidental at most.
@jamessseattle @ToKTeacher @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
In order:
Could be but unimportant - single Turing machine can simulate multiple Turing machines.
Processor and memory.
Doesn't matter/is incidental (if even true).
Yes (if you mean spoken language).
@jamessseattle @ToKTeacher @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
No such thing as virtual computer (if by "virtual" you mean "abstract"). Computers need to be physically built.
@ToKTeacher @jamessseattle @RatCritical @Malcolm_Ocean @DavidDeutschOxf @reasonisfun @Crit_Rat
Can you explain why its being analog or digital has any bearing on this?
@caerwy @MatjazLeonardis @DavidDeutschOxf @RatCritical @Malcolm_Ocean @ToKTeacher @reasonisfun @Crit_Rat
I don’t think the presence of understanding (ie knowledge) is indicative of consciousness.
It’s trivial to write a function that represents understanding of a prime number.
I guess that consciousness is related to error correction.
@zarzuelazen
This is the kind of vacuous nonsense that has earned philosophy its bad, navel-gazey reputation.
I agree that we start with conjecture and can then test against brain activity. Thank you; you have helped me realize something important about neuroscience.
Then how can we hope to reconstruct the software that caused these patterns? I think there are infinitely many pieces of software that would result in the same pattern.
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Yes, we're in agreement here; though I had already agreed that architecture influences speed. But let me ask you this: can two different algorithms, when run, result in the exact same movement in hardware?
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Do these performance characteristics not lie in the algorithm itself? Would X not also take 10 years to run on a desktop computer? (Assuming that computer would have the same memory and processing power as the brain you're comparing it to.)
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Or to clarify (something I should have clarified before): hardware architecture can and does influence speed and processing power. But it doesn't qualitatively change anything about which algorithms the universal system can run.
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Yup, agreed.
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
"not all algorithms that solve a given computational problem are equally efficient or robust"
Indeed, because this concerns the architecture of software.
@recursus @RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
"Not all architectures are equally good for running a given algorithm"
A universal computer, no matter its architecture, can run any computable algorithm (within its memory constraints).
@Malcolm_Ocean @RatCritical @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
When a computer breaks, shit can get weird, too: the fan keeps running, or the housing gets really hot, or it randomly flashes bright colors, some keys work while others do not, it keeps beeping for no apparent reason, it runs all programs fine except the calculator app... etc.
@Malcolm_Ocean @RatCritical @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
It does not. Evolution only optimizes the ability of the gene to spread through the population.
@Malcolm_Ocean @RatCritical @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Evolution does not optimize for efficient ways to organize any alleged modules.
@RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
If we knew how to program consciousness and ran it on a computer made of chewing gum and vacuum tubes, those interested would start studying the properties of chewing gum in order to understand consciousness.
@RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
Somewhere in the brain there is memory, and somewhere there is a processor. Like in all computers. So what?
In order to understand brain functionality, one needs to understand the software that's running on the brain.
@RatCritical @Malcolm_Ocean @DavidDeutschOxf @ToKTeacher @reasonisfun @Crit_Rat
I haven't read it, but the problem with learning about the brain's functionality from its architecture (hemispheres, regions, parts, etc) is this: the brain is a universal computer. Since it's universal, its architecture does not matter.
I’ve only read one of his stories, and remember it being alright. But I love the video games by @frictionalgames which are Lovecraftian.
RT @andrewdoyle_com:
Politicians should never be invoking the skin colour of their opponents.
This is the cancer of wokeness: it convinces…
@ks445599 @bnielson01 @ToKTeacher
AlphaGo is machine learning. Machine learning is “learning from experience”. It’s empiricism. It’s impossible. Whatever is happening there, it’s not learning. No explanatory or any other alleged kind of knowledge is created.
@SimonDeDeo @wileyprof
Me neither. Finished it yesterday. Was a pain to get through.
I should add that it gets worse, though. I know some decision makers in Silicon Valley who know damn well it has nothing to do with intelligence but choose to call it that anyway because it sells. That’s fraud.
Learning is the creation of knowledge. Narrow AI is the following of predefined steps; the same as all programming so far. Using the term "learning" for something that clearly isn't learning is utterly misleading and dishonest (I'm mad at the industry, not you.)
@ChipkinLogan @scotthortonshow
Very well written. It really is denial of knowledge creation again: all the wealth is allegedly already present, just needs to be redistributed "fairly".
@onnlucky @bnielson01 @ks445599 @mizroba
Well, then maybe you should read more Popper (something I have recommended to you before).
Or at least google "Popper/Bühler four functions of language".
You will then learn that most of the things you listed require explanations, not merely descriptions.
@onnlucky @bnielson01 @ks445599 @mizroba
They're not. The point is it doesn't matter if it's vocalized or only descriptive. Popper explained that the descriptive function of language is built on top of the signaling function.
@onnlucky @bnielson01 @ks445599 @mizroba
None of the things you just listed require spoken language. The only way I can make sense of this is that you thought we were talking about programming language or something. We weren't.
@markcannon5 @bnielson01 @mizroba
Yes. Whoever builds an AGI will be its parent. His obligation will be to help it learn.
@bnielson01 @ks445599 @mizroba
Yes. Language is not required to make a universal explainer. How could it be? You need to be one in order to learn language in the first place. Again, no instruction from without possible.
Have the terms "eco-induction" or "green induction" been coined yet? Might be useful to refer to this sort of idea.
That's what this idea suggests: that there is knowledge somehow already present in the biosphere (not as adaptations!), and we just need to harvest/ingest it.
"How much do species and ecosystems contribute to the size and growth of economies?"
They don't. Only people do. And only explanations guessed by people tell us which resources, species, and ecosystems are at all relevant to economic growth. twitter.com/NaturePortfoli…
"But if [@DavidDeutschOxf] is right that evolution has led to humans as the ultimate knowers, where did this knowing arise?" from sfgate.com/opinion/chopra…
That is a problem the theory of evolution solves. David addresses it in "The Beginning of Infinity".
@ks445599 @bnielson01 @mizroba
Indeed. It could start making all kinds of conjectures about reality and criticize them. It just won't be able to test them. For example, it will know that it exists, and that therefore things can exist. It can then guess that other things might exist, too, what they may be, etc.
@bnielson01 @ks445599 @mizroba
I think you're flirting with induction here. A child does not learn to be intelligent, especially not via instruction from without. He is born intelligent. What knowledge he creates using that intelligence and whether IO may help with that are different matters.
RT @JeffBezos:
It’s time to go back to the Moon, this time to stay. @LockheedMartin @northropgrumman @DraperLab @blueorigin #gradatimferoci…
"How can we build beneficial AGI?" is yet another "Who should rule?" question. It should be replaced with "How can we detect our errors in building AGI?" and then, once built, "How can we help AGI learn?"
Stopping doing one thing and starting doing another is easily programmed and doesn’t involve the halting problem.
It wouldn’t need to be made of DNA. If it wants to study chemical reactions they can be exosomatic.
Ah, yes. Though to be clear, an AGI running on a computer is a completely non-biological phenomenon. No genes required.
That's fine, but then I still don't understand your original comment: twitter.com/EvanOLeary/sta…
Sure. But in order to learn about music?
Why would it have to be able to perform organic chemical reaction experiments in order to be able to learn anything?
Not sure I follow. Care to elaborate? I don’t think any biological substrate is needed.
I’m... not even mad. The article should be titled “In Flagrant Disregard of Computational Universality”. Some real gems in there. twitter.com/CloserToTruth/…
@reasonisfun @HeuristicWorld @dela3499 @ToKTeacher @Crit_Rat @RatCritical @BretWeinstein
Yet they can, and frequently do, lead to death. I suspect this is similar to how genes have to keep their hosts alive but often don’t.