Dennis Hackethal’s Blog
My blog about philosophy, coding, and anything else that interests me.
Tweets
An archive of my tweets and retweets through . They may be formatted slightly differently than on Twitter. API access has since gotten prohibitively expensive – I don't know whether or when I'll be able to update this archive.
But in case I will, you can subscribe via RSS – without a Twitter account. Rationale
Note also how in many of these videos what we don’t see is that the filmed results are preceded by a loooong training phase (akin to the training phases in narrow AI) in which humans reinforce good and punish bad behavior.
Reinforcement “learning” != creative learning. Note how she praises the dog. The error correction happens in the owner, that in the dog. The owner merely passes it on to the dog.
Mimicking behavior does not require intelligence.
He’s trying to see if the cone is edible. No intelligence required for that.
I enjoy the rare occasions when people openly and readily change their minds. t.co/WdSICI43FE
Reminds me of Cameron Hanes’ line “Nobody cares, work harder.”
Yup exactly, that’s what I meant with the original tweet.
Exactly. That realization is what’s missing in the whole “are animals intelligent” debate. Presence of knowledge != intelligence.
Once again, the behavior in this video can be explained entirely in terms of inborn, genetically inherited knowledge. No intelligence on the part of the worm required.
@ModelsofMind @CNC3P0 @SamHarrisOrg
And the fact that people disagree about what should be done implies there is a problem to be solved, and, thereby, knowledge to be created. Which is better than just steamrolling over those who disagree.
@ModelsofMind @CNC3P0 @SamHarrisOrg
Preferences are ideas, too. And ignoring them can lead to coercion.
What I was trying to say is that people need to be persuaded if we want them to do something.
Just caught myself writing: "the fact that... shows..." Then I corrected it to: "the reason that... is..."
The first is inductive, authoritarian, and not an explanation, but a statement. The second is not inductive, leaves room for having made a mistake, and is an explanation.
The bird is pecking at the hog, maybe because the hog smells like food. The bird probably has no idea what a road is or what cars are or that they can kill animals.
It’s pecking at it. It’s probably just trying to get food out of it. It has no idea what forces are.
Not everyone agrees on what needs to be done. Nor should children be told what to do. They're full-fledged people, too.
Plus they have a universally applicable schema so that they can fit any data model. They also support a powerful query language (datalog).
I think they make real-time sync easier than other approaches. Oh, and they don't make you code in strings—db queries are data—which is a major plus for me.
Datomic in the backend and datascript in the frontend because it makes real-time syncing a walk in the park.
...and doesn’t realize it’s stuck and keeps repeating the same action over and over.
I don’t think rationality is selected by biological evolution, it that’s what you mean. People have to learn how to be rational (and are never done learning that).
"We need another lockdown." How about not telling people what to do?
Though of course, that is still granting agency on the part of the elephant, which it doesn't have. I'm just trying to draw attention to the fact that its "motivation" need not have been to save the human.
Another possible caption: "Ooh, looks like it's a human, I've had fun playing with them before and they give me treats. I'll go play with that one again, maybe I'll get treats again this time!"
Correction:
"Any time we're impressed by what non-human animals do, it's simply because we forget that they have genetically inherited sophisticated knowledge, which was created by biological evolution, not by those animals."
Common error: mistaking the sophistication of an animal's knowledge with the animal being intelligent.
For the former, the animal need only contain sophisticated knowledge—never mind where that knowledge came from—for the latter, it must have created the knowledge itself. twitter.com/neiltyson/stat…
Not sure either labels strong/weak or mind virus apply to either of those theories.
But generally speaking, if most people have Newton's idea of gravity, it's because his meme is better at spreading than Einstein's.
From here, it's just one more step to meeting a necessary condition for AGI: making it possible for the program to correct its own errors.
Applied to programming, it means we shouldn't judge a program by how well it solves a given problem, but by how easy it makes it to detect and correct errors it already contains.
Similarly, in science, we do not judge institutions by how much they produce or entrench good theories, but by how easy they make it to criticize and replace bad theories (ibid).
How do we get from Popper to AGI? 🧵👇
Popper's criterion of democracy entails that we shouldn't judge a political system by whether it produces good policies or leaders, but by how easy it makes it "to remove bad ones that are already there." — The Beginning of Infinity, ch. 9
Wait—I could see that weaker strains of virus spread farther, but is it true that strains that spread farther get weaker? Seems as though the causation might be reversed.
And ~1% use Reagent, even though it is superior and doesn’t need any explicit Redux.
@HeuristicAndy @MartvMegen
I checked those out, I may post in themotte in the future.
I’ve had mixed experiences with them for sure. May not post there again.
Just like today people believe in superintelligences!
@Crit_Rat
Due to its superhuman stretching abilities it may potentially be dangerous. Cat breeding should be regulated and we should be prepared to destroy the cats when they rise up and stretch everywhere.
@DoqxaScott @Plinz @BasilMarte @Levi7hart @nosilverv
Yes but my point was merely that the evolution of ideas itself has no purpose. People and their minds do contain purposes, of course.
@Plinz @BasilMarte @Levi7hart @nosilverv
No, there is no purpose in the evolution of ideas. Ideas get themselves adopted because they’re good at spreading. Some ideas help improve world views and intellectual developments etc, but some don’t.
@Plinz @BasilMarte @Levi7hart @nosilverv
Agreed but people are creative so the knowledge of how to fit in need not be encoded genetically, they can create it during their lifetime.
@Plinz @BasilMarte @Levi7hart @nosilverv
Plus memes evolve too quickly to evolve genetic defenses against them. Not to mention that even if it were possible, memes are much more powerful than genes and easily override genetic instructions (as evidenced by the religiously-spread meme of celibacy).
@Plinz @BasilMarte @Levi7hart @nosilverv
Also, following Dawkins, I believe memetic and genetic evolution are largely separate. But even if they weren't, parasitic genes and memes can manage to spread well despite hurting their hosts by keeping them alive well enough to spread, but without promoting their wellbeing.
@Plinz @BasilMarte @Levi7hart @nosilverv
I'm not. I agree that the evolution of religion is rich in memetic terms. That doesn't contradict what I said.
Slightly better, maybe. Luckily, color contrast is objectively measurable. Paste your background color as hex (#041311) into this tool and it will let you pick a well-contrasted text color:
@Plinz @BasilMarte @Levi7hart @nosilverv
I think you're granting way too much genius and agency to the "authors" of any religion.
Religions are memeplexes that spread because they happen to be adapted to spreading, not because they explain the world or are good for people etc.
People get mad when you take computational universality seriously. Or any idea, really. Other than the one that says not to take ideas seriously.
From reddit.com/r/Intellectual… https://t.co/3GF4t9AMlo
You may benefit from increasing the color contrast on your color palette. I had to turn up the brightness on my phone quite a bit to make the UI legible, especially the code on the top right.
Very cool. How have you been building the UI—is it written in HTML and displayed in a browser, or some other way?
I think AGI will be achieved in a qualitative jump from something much less powerful, but it will take time to get to that jump.
Somewhere in the West because Western countries are the only ones with even a hint of good epistemology. No sign of AGI yet though.
RT @deezzer:
For coders - A great discussion on benefits of Functional vs OOP youtube.com/watch?v=uu3tb3… #programming #better #reactjs #functi…
I was interviewed about the new Berlin programming language, which aims to make programming simpler and more enjoyable. This video is a good intro: youtube.com/watch?v=uu3tb3…
This thread has been archived on archive.vn/jsBFE, archive.vn/fVbeG, archive.vn/KzNXG, and various mirrors thereof.
Social-justice warriors may be gaining power right now, but they will eat themselves up eventually and they will turn against you and may well use this twitter thread to do so.
What you post makes for good marketing and it’s very “in” right now, but people may not always look on it so favorably. You may choose to be more careful. You should jump off the social-justice bandwagon while you still can.
I’m not a boomer :) I think you have good intentions btw. But by basing investment decisions on race and publicly admitting to it you’re just begging for someone you turn down in the future to file a lawsuit against you.
By basing investment decisions off of skin color, you’re lowering the standards for blacks relative to nonblacks, which is economically counterproductive, patronizing, and, of course, racist.
You’re clearly making investment decisions based on race—otherwise you couldn’t account for your “subconscious bias,” which is a non-falsifiable idea that people can always use to justify decisions based on skin color.
(And btw, by admitting to “subconscious bias” you basically admitted to harboring racist thoughts against black people.) So then, the question is, how can you ensure a “fair” outcome without favoring black people and disfavoring white people?
I take that as a “yes,” because, to ensure you’re not “exhibit[ing] any subconscious bias,” you must look at the ratio of black founders you invested in. That’s why you share the number of “black investments” so proudly.
And to do that, and because—people would hope—you want to ensure a proportionate and fair outcome, you are taking into account their race to make an investment decision, yes?
Are you favoring black people for investments? (Not a rhetorical question)
I’m assuming the first picture is a rendering, the second is a photograph?
@ReachChristofer @FitzClaridge
Do ideas replicate within a mind?
@ReachChristofer @FitzClaridge
What are the similarities and differences between coercive mechanisms within a mind and across minds?
Is TCS compatible with non-libertarian political views?
... to a more positive, open, people-friend (and, thereby, AGI-friendly) worldview, in which rapid progress and error correction are things not to be feared and avoided, but celebrated.
technology ultimately requires a shift in attitude from the kind of cynical, pessimistic and AGI-skeptical view I explained in this tweet: twitter.com/dchackethal/st…
And, as I have argued in another thread with someone else, most of these surface issues are really only addressed properly by getting on the same page about epistemology.
Lastly, at the end of the day, embracing AGI and seeing it as a fascinating and ultimately positive...
That software is the decisive limiting factor does not make for a trivial case—indeed, it follows from our best explanations about computation.
But yes, as I have argued before, more processing speed can help with faster error correction, including the correction of moral errors
Most of the anti-AGI worries translate into this: sentencing AGIs to death for crimes against humanity they didn’t commit, nor could have committed because they haven’t been born yet, but might possibly commit.
Basically pre-natal thought crimes.
I get VERY excited watching videos like this one: youtube.com/watch?v=y6VlzF…
I've considered that, and it might be an interesting approach. But again, the tech was easy—a browser extension would also involve getting people to install it, use it frequently, etc.
Unless I could somehow integrate the browser extension with an existing social network...
I built a social network once that had the ability to compute the quality of a post. The tech was easy; getting people to join more difficult.
Reminds me I should spend more time reading and less time on Twitter.
@nburn42 @Plinz
Well, the brain is a computer, but the mind is software running on the brain. The mind attempts to explain the world by repeatedly conjecturing and criticizing solutions to problems (Popper).
(and no, faster hardware doesn't make an AGI child even significantly different in the relevant ways)
In other words, faster hardware by itself doesn't make an AGI child not a child.
If this still doesn't convince you, then I suspect that your condition to convince you wasn't sufficient in the first place.
If you're like most AI-safety worriers, you may reply "but this is different! This time, it's orders of magnitude!" So were computers when they first came around, and they have been gaining orders of magnitude every decade since. Which has again been overwhelmingly a good thing
That's not to mention Deutsch's point that people have been increasing their hardware over the centuries through pen, paper, computers, etc and that's overwhelmingly been a good thing.
Hardware performance is overrated for this particular issue. It needs to be accompanied by improved software performance, the AGI child needs to learn how to use fast hardware, etc...
No, "a human mind is a function that takes in observations and yields behavior and policy updates" is an empiricist and behaviorist misrepresentation of the mind. Human minds are creative, and continue do be so without sense input.
I've enjoyed videos from Bucky Roberts' channel: youtube.com/user/thenewbos…
And Treehouse has good tutorials as well (paid though!): teamtreehouse.com/courses
The same question remains: but why? Where does “greater potential for rapidly gaining power and influence over the world” come from? And where do the bad inborn ideas come from?
Or better yet, no school at all, just voluntary learning and fun.
Btw I already explained how and why an AGI child is just like a human child and you ignored that and made an unargued assertion to the contrary when, if I’m not mistaken, I had met your condition to change your mind.
Right, an AGI child might not be bright at all. Or mediocre. Or very smart.
To your second point: but why?!
Ok. To learn about Popperian epistemology, I recommend reading "The Beginning of Infinity" by David Deutsch, or Karl Popper's "Objective Knowledge."