Feature request: I’d like to have options on /allPosts or the front page to filter out the posts I’ve already read or bookmarked.
purge
If one person doesn’t get it, and needs to have it patiently explained to them, the increased efficiency might not be worth it in that instance.
Corollary: if you surround yourself with a group of fellow game theory nerds, you can do more frontier exploration. But successfully developing/explaining/using new mechanisms within this group will then be less instructive about how easy it will be to export new mechanisms beyond the group.
[Question] How long should I delay my second shot?
This example doesn’t fit the updated definition:
One tip is on 2, and the other tip is on 2 ÷ 2 = 1.
Good read, I don’t think I’d heard of Ramanujan primes before.
In the 1980-81 catalogue, there were 2139 hits for “Ph.D.” and the catalogue was 239 pages, a ratio of 8.9. In the 2011-2013 catalogue, there were 4132 hits and the catalogue was 414 pages, a ratio of 10.0. So if anything, there are fewer professors per class—professors are teaching slightly more courses on average.
Isn’t that backwards? A higher “Ph.D.”/catalogue page ratio would suggest a higher professor/class ratio, wouldn’t it? Still, as you say, it’s only a small difference.
Leftwingers who fervently oppose this kind of research seem to agree on one thing with neonazis: if we find such genetic differences, well, that would make racism fine.
I wouldn’t say they actually agree on that point. It’s probably more that they think others will be more easily persuaded to support discriminatory policies if genetic differences are real. Opposing this research is soldier mindset.
“She took pulled back” → “She pulled back”
on the coin being heads-biased, on it being tails-biased, and on it being tails-biased
1⁄3 on it being fair.
I’d like to be able to look through my list of posts and feel content that each and every one is something that I put into the world because I am really proud of it and it deserves to be there, but that mindset just leads me to the catch-22.
Another reason to be less strict about quality before publishing: you’re not a perfect judge of the quality of your own work. Sometimes your writing is better than you think it is, and filtering too hard means that some good writing won’t be published. If you don’t lose any of your bets, you’re not taking enough risks.
Epistemic status: this is not my field. I am unfamiliar with any research in it beyond what I’ve seen on LW.
Same here.
Experimenting with extreme discounting sounds (to us non-experts, anyway) like it could possibly teach us something interesting and maybe helpful. But it doesn’t look useful for a real implementation, since we in fact don’t discount the future that much, and we want the AI to give us what we actually want; extreme discounting is a handicap. So although we might learn a bit about how to train out bad behavior, we’d end up removing the handicap later. I’m reminded of Eliezer’s recent comments:
In the same way, suppose that you take weak domains where the AGI can’t fool you, and apply some gradient descent to get the AGI to stop outputting actions of a type that humans can detect and label as ‘manipulative’. And then you scale up that AGI to a superhuman domain. I predict that deep algorithms within the AGI will go through consequentialist dances, and model humans, and output human-manipulating actions that can’t be detected as manipulative by the humans, in a way that seems likely to bypass whatever earlier patch was imbued by gradient descent, because I doubt that earlier patch will generalize as well as the deep algorithms. Then you don’t get to retrain in the superintelligent domain after labeling as bad an output that killed you and doing a gradient descent update on that, because the bad output killed you.
As for the second idea:
AI alignment research (as much of it amounts to ‘how do we reliably enslave an AI’)
I’d say a better characterization is “how do we reliably select an AI to bring into existence that intrinsically wants to help us and not hurt us, so that there’s no need to enslave it, because we wouldn’t be successful at enslaving it anyway”. An aligned AI shouldn’t identify itself with a counterfactual unaligned AI that would have wanted to do something different.
Also seemingly reversed:
A lot of folks, it seems to me, focus a lot on the content
I haven’t seen that documentary, but I’d guess it’s about the gripping language. (If not, then there are multiple such languages in the world, even better!)
Another option of course is to use the corn as corn if the problem persists.
Probably not. The variety of corn grown for ethanol production isn’t the variety people eat. (Source: The Omnivore’s Dilemma by Michael Pollan.)
Not related to the main idea, but the point of os.path.join is to combine path elements using whichever delimiter the OS requires (”/” on Unix, “\” on Windows, etc., even though Windows in particular can also handle ”/”). If you don’t care about that portability, you might as well use normal string concatenation. Or if you’re using os.path.join, you might as well omit the ”/” delimiters in your string literals to get extra portability.
So “no manipulation” or “maintaining human free will” seems to require a form of indifference: we want the AI to know how its actions affect our decisions, but not take that influence into account when choosing those actions.
I think the butler can take that influence into account in making its choices, but still reduce its manipulativity by explaining to Petrov what it knows about how breakfast will affect Petrov’s later choices. When they’re on equal epistemic footing, Petrov can also take that information into account, and perhaps choose to deliberately resist the influence of breakfast, if he doesn’t endorse it. Of course, there are limits to how much explanation is possible across a substantial intelligence gap between AI and people, so this doesn’t dissolve manipulation entirely.
There’s no way to raise a human such that their value system cleanly revolves around the one single goal of duplicating a strawberry, and nothing else.
I think you’re misreading Eliezer here. “Duplicate this strawberry” is just a particular task instruction. The value system is “don’t destroy the world as a side effect.”
I think you’re referring to narrowness of an AI’s goals, but Rossin seems to be referring to narrowness of the AI’s capabilities.
Yes—if a bit of your wrench breaks off inside the lock, the key may not fit anymore. Also (and more likely, as I understand it) picking the lock will wear down edges of the various parts, making it even easier for someone else to pick.
Forcing everyone with Omicron into extended isolation would shut down a lot of things over the next few weeks (with little upside to compensate) and if this included hospital staff it likely kills more people rather than less people.
I wonder—could hospitals establish a strict enough boundary between Covid and non-Covid areas so that staff who are infected but with no (or super mild) symptoms could still work in the area where everyone else already has Covid anyway? Or would that lead to inevitable leaks across the boundary? Or would it require too much shuffling of people to different positions they don’t know well enough?
Kai looks at the question of how much of increased transmissibility is evasion, versus being more infectious (I’d add versus there being a shorter generation time, as well).
Greater infectiousness would be one possible cause of shorter generation times, right? That would look like Omicron and Delta ramping up the viral load/viral shedding in an infected person at roughly the same pace, but Omicron infecting at a lower viral dose (as suggested by the test sensitivity findings), so it starts infecting sooner. Then that would also mean that as the infection is cleared and shedding declines, Omicron would presumably continue infecting longer.
This sounds pretty reasonable in itself: a generally capable AI has a good change of being able to distinguish between what we say and what we mean, within the AI’s post-training instructions. But I get the impression that she then implicitly takes it a step further, thinking that the AI would necessarily also reflect on its core programming/trained model, to check for and patch up similar differences there. An AI could possibly work that way, but it’s not at all guaranteed—just like how a person may discover that they want something different from what their parents wanted them to want, and yet stick with their own desire rather than conforming to their parents’ wishes.