(e.g. in the hypothetical ‘harbinger tax’ world, you actively want to sabotage the resale value of everything you own that you want to actually use).
“harberger tax,” for anyone trying to look that up.
(e.g. in the hypothetical ‘harbinger tax’ world, you actively want to sabotage the resale value of everything you own that you want to actually use).
“harberger tax,” for anyone trying to look that up.
If you can pay the claimed experts enough to submit to some testing, you could use Google’s new doubly-efficient debate protocol to make them either spend some time colluding, or spend a lot more time in their efforts at deception: https://www.lesswrong.com/posts/79BPxvSsjzBkiSyTq/agi-safety-and-alignment-at-google-deepmind-a-summary-of
This could exclude competent evaluators without other income—this isn’t Dath Ilan, where a bank could evaluate evaluators and front them money at interest rates that depended on their probability of finding important risks—and their shortage of liquidity could provide a lever for distortion of their incentives.
On Earth, if someone’s working for you, and you’re not giving them a salary commensurate with the task, there’s a good chance they are getting compensation in other ways (some of which might be contrary to your goals).
Thanks! Just what I was looking for.
Some cities have dedicated LW/ACX Discord servers, which is pretty neat. Many of the cities hosting meetups over the next month are too small to have much traffic to such a server, were it set up. A combined, LW meetup oriented Discord server for all the smaller cities in the world, with channels for each city and a few channels for common small-meetup concerns, seems like a $20 bill on the sidewalk. So I’m checking whether such a thing exists here, before I start it.
I think the cruxes here are whether Aldi forced out small retailers like Walmart did; and how significant the difference between Walmart and Aldi is, compared to the difference between Aldi and large, successful retail orgs in wentworthland or christiankiland.
(my experience in German shopping is that most grocery stores are one of a half-dozen chains, most hardware stores are Bauhaus or OBI, but there isn’t a dominant “everything” store like Walmart; Müller might be closest but its market dominance and scale is more like K-mart in the 90′s than Walmart today.)
An existing subgenre of this with several examples is the two-timer date. As I recall, it was popular in 90′s sitcoms. Don’t expect INT 18 tier scheming, but it does usually show the perspective of the people frantically trying to keep the deception running.
Here’s the intuition that’s making me doubt the utility of provably correct system design to avoiding bridge crashes:
I model the process leading up to a ship that doesn’t crash into a bridge as having many steps.
1. Marine engineers produce a design for a safe ship
2. Management signs off on the design without cutting essential safety features
3. Shipwrights build it to spec without cutting any essential corners
4. The ship operator understands and follows the operations and maintenance manuals, without cutting any essential corners
5. Nothing out-of-distribution happens over the lifetime of the ship.
And to have a world where no bridges are taken out by cargo ships, repeat that 60,000 times.
It seems to me that provably safe design can help with step 1--but it’s not clear to me that step 1 is where the fault happened with the Francis Scott Key bridge. Engineers can and do make bridge-destroying mistakes (I grew up less than 50 miles from the Tacoma Narrows bridge), but that feels rare to me compared to problems in the other steps: management does cut corners, builders do slip up, and O&M manuals do get ignored.
With verifiable probabilities of catastrophe, maybe a combination of regulation and insurance could incentivize makers and operators of ships to operate safely—but insurers already employ actuaries to estimate the probability of catastrophe, and it’s not clear to me that the premiums charged to the MV Dali were incorrect. As for the Francis Scott Key, I don’t know how insuring a bridge works, but I believe most of the same steps and problems apply.
(Addendum: The new Doubly-Efficient Debate paper on Google’s latest LW post might make all of these messy principal-agent human-corrigibility type problems much more tractable to proofs? Looks promising.)
https://space.mit.edu/home/tegmark/crazy.html Large universes put some subtleties into the meaning of “real” that aren’t present in its common usage.
Decision theory-wise, caring about versions of yourself that are inexorably about to dissolve into thermal noise doesn’t seem useful. As a more general principle, caring about the decisions you make seems useful to the extent that those decisions can predictably change things.
My dreams have none of the consistency that allowed smart people to figure out the laws of nature over the millenia. It might be possible for a superintelligence to figure out how to make decisions within a world working on dream rules which had predictable future effects, but I believe it to be far beyond my powers.
That example seems particularly hard to ameliorate with provable safety. To focus on just one part, how could we prove the ship would not lose power long enough to crash into something? If you try to model the problem at the level of basic physics, it’s obviously impossible. If you model it at the level of a circuit diagram, it’s trivial—power sources on circuit diagrams do not experience failures. There’s no obviously-correct model granularity; there are schelling points, but what if threats to the power supply do not respect our schelling points?
It seems to me that, at most, we could prove safety of a modeled power supply, against a modeled, enumerated range of threats. Intuitively, I’m not sure that compares favorably to standard engineering practices, optimizing for safety instead of for lowest-possible cost.
I guess this is common knowledge, but I missed it: What is with the huge dip in CPI before 2020? I’m confused, especially because the 2008 crash barely shows up. A cursory googling and asking ChatGPT failed me.
Anecdotally*, IPL/laser therapy seems to do all of these except increasing dermal capillaries, which it instead reduces. This makes it ideal for people with rosacea or other inflammatory problems, and fair skin (which often accompanies these problems).
*And with a few references: Effective treatment of rosacea using intense pulsed light systems—PubMed (nih.gov)
IPL irradiation rejuvenates skin collagen via the bidirectional regulation of MMP-1 and TGF-β1 mediated by MAPKs in fibroblasts—PubMed (nih.gov)
some studies find no significant effect on collagen etc.; I don’t know what went wrong here besides perhaps too-tan skin: Intense pulsed light photorejuvenation: a histological and immunohistochemical evaluation—PubMed (nih.gov)
You’ll be happy to know that standards bodies have noticed the “entropy reduction from excessive rules” problem. The latest version of NIST Special Publication 800-63B says to disallow four password categories like “already in a breach database” and “aaaaa,” but goes on to direct verifiers to not impose any other rules on password composition.
As for me, I just choose the first four digits of the busy beaver numbers--1621--as my PIN. As a noncomputable number, it’s guaranteed to be the most random choice possible.
One unstated, load-bearing assumption is that whatever service or good humans can trade to ASI will be of equal or greater worth to it than our subsistence income.
Land Value Tax would solve this.
(Sort of—funding UBI from a 100% LVT would solve it for the case of literal rent seeking, because if landlords increased the rent, that additional money would be taxed back into the UBI pool. To make it a general solution, you’d have to identify all instances of rent-seeking, and tax the underlying asset with a metaphorical 100% LVT).
Sure, that’s fair enough. I was thinking in the context of “formal verification that would have prevented this outage.”
It would specifically be impossible to prove the Crowdstrike driver safe because, by necessity, it regularly loads new data provided by Crowdstrike threat intelligence, and changes its behavior based on those updates.
Even if you could get the CS driver to refuse to load new updates without proving certain attributes of those updates, you would also need some kind of assurance of the characteristics of every other necessary part of the Windows OS, in every future update.
I feel like it’s still Moloch to blame, if a sufficient bribe to landowners would cost less than indefinitely continued rent-seeking.
I don’t have any calculations to offer in support; but I would generally expect an individual landowner’s time preference to be lower than society’s as a whole, so I suspect this is indeed the case.
So the actual reason is that landowners don’t want to be seen taking a bribe, because that would involve acknowledging they have been knowingly rent-seeking since 1879; and the government doesn’t want to openly bribe them for moral hazard whatever; so even though everyone would be better off by their own lights it can’t happen. And that’s fairly moloch-flavored.
It’s very standard advice to notice when a sense of urgency is being created by a counterparty in some transaction; and to reduce your trust in that counterparty as well as pausing.
It feels like a valuable observation, to me, that the counterparty could be internal—some unendorsed part of your own values, perhaps.