As I understand it, one of the biggest issues with a land value tax is that the existence of the tax instantly makes owning land much less desirable—reduced by the net present value of the total future taxation. This is obviously in some sense part of the plan but it causes some pretty large sudden shifts in wealth—in particular away from anyone who has a mortgage but also just from home owners in general.
Implementing it in a fair/politically acceptable way then seems to require either a far-off starting date, a very slow taper in or a very large series of handouts to compensate, and all of these are difficult for a government to implement given the time horizon of elections and a large, wealthy group who will be opposed to this, likely including inside the governing party.
This isn’t especially relevant to your variant but if you’re thinking about how to get efficient taxation then this is something to think about trying to find a solution to :)
On the numbers from The Precipice—I think the point is that the next 100 years have an estimated 1⁄6 chance of extinction, but also contain the power to protect us from future harm and facilitate the human race flourishing across the universe. Extrapolating risk from next 100 years to an expected 600 year lifespan, and using current population forecasts as the number of humans involved therefore seems not in the spirit of his model.
I think this this points to the strategic supremacy of relevant infrastructure in these scenarios. From what I remember of the battleship era, having an advantage in design didn’t seem to be a particularly large advantage—once a new era was entered, everyone with sufficient infrastructure switches to the new technology and an arms race starts from scratch.
This feels similar to the AI scenario, where technology seems likely to spread quickly through a combination of high financial incentive, interconnected social networks, state-sponsored espionage etc. The way in which a serious differential emerges is likely to be more through a gap in the infrastructure to implement the new technology. It seems that the current world is tilted towards infrastructure ability diffusing fast enough to, but it seems possible that if we have a massive increase in economic growth then this balance is altered and infrastructure gaps emerge, creating differentials that can’t easily be reversed by a few algorithm leaks.
Apologies if this is not the discussion you wanted, but it’s hard to engage with comparability classes without a framework for how their boundaries are even minimally plausible.
Would you say that all types of discomfort are comparable with higher quantities of themselves? Is there always a marginally worse type of discomfort for any given negative experience? So long as both of these are true (and I struggle to deny them) then transitivity seems to connect the entire spectrum of negative experience. Do you think there is a way to remove the transitivity of comparability and still have a coherent system? This, to me, would be the core requirement for making dust specks and torture incomparable.
Late to the party but I’m pretty confident he’s saying the opposite—that a 1 PFLOP/s system is likely to have 10 or more times the computational capacity of the human brain, which is rather terrifying.
He gives the example of Baidu’s Deep Speech 2 which requires around 1 GFLOP/s to run and produces human-comparable results. This is 10^6 slower than the 1 PFLOP/s machine. He estimates that this process in humans take around 10^-3 of the human brain, thereby giving the estimate of a 1 PFLOP/s system being 10^3 times faster than the brain. His other examples give similar results.
An easy way to deal with this difficulty is to replace ‘at least as happy with policy A as with policy B (in any situation that we think might arise in practice)’ with ‘at least as happy with policy A as with policy B (when averaged over the distribution of situations that we expect to arise)’, though this is clearly much weaker.
To me it seems that the reason this stronger sense of ordering is used is because we expect this amplification procedure to be of a sort that produces results such that A+ is strictly better than A but that even if this wasn’t the case, the concept of an obstruction would still be a useful one. Perhaps it would be reasonable to take the more relaxed definition but expect that amplification would produce results that are strictly better.
I also agree with Chris below that defining an obstruction in terms of this ‘better than’ relation brings in serious difficulty. There are exponentially many policies B that are no better than A+ and there may well be a subset of these can be amplified beyond A+ but as far as I can tell there’s no clear way to identify these. We thus have an exponential obstacle to progress even within a partition, necessitating a stronger definition.
When you talk about ‘black-box’ versions of Hugh, do you envision that H is able to answer questions relating to the cognitive processes that lead to the answer given, or about H’s thinking in general? This seems to contradict the spirit of a black box but self reflection is an important part of Hugh’s cognitive ability.
Perhaps they are both useful possibilities, my intuition is that this kind of self reflection is as far from being possible for AI as any human ability and so we should expect that we might have systems powerful enough to take on wide responsibility without this ability. If it were possible, though, the ability to use loops of self reflection to check whether a cognitive process serves a certain goal would be very helpful.
I’ve realised that you’ve gotta be careful with this method because when you find a trichromatic subtriangle of the original, it won’t necessarily have the property of only having points of two colours along the edges, and so may not in fact contain a point that maps to the centre.
This isn’t a problem if we just increase the number n by which we divide the whole triangle instead of recursively dividing subtriangles. Unfortunately now we’re not reducing the range of co-ords where this fixed point must be, only finding a triad of arbitrarily close points that map to a triangle surrounding the centre. You can, for example, take the centre point of the first of these triangles (with some method of numbering to make the function definite) for each value of n=1,2,3.. as a sequence in R2. This must have a convergent sequence which should converge to a point that maps to the centre but I can’t prove that last stage.
Cleanest solution I can find for #8:
Also, if we have a proof for #6 there’s a pleasant method for #7 that should work in any dimension:
We take our closed convex set S that has the bounded function h:S→S . We take a triangle T that covers S so that any point in S is also in T .
Now we define a new function h′:T→T such that h′(x)=h(cs(x)) where cs(x) is the function that maps x to the nearest point in S.
By #6 we know that h′ has a fixed point, since cs is continuous. We know that the fixed point of h′ cannot lie outside S because the range of h′ is S. This means h′ has a fixed point within S and since for x∈S, h(x)=h′(x), h has a fixed point.
Yeah agreed, in fact I don’t think you even need to continually bisect, you can just increase n indefinitely. Iterating becomes more dangerous as you move to higher dimensions because an n dimensional simplex with n+1 colours that has been coloured according to analogous rules doesn’t necessarily contain the point that maps to zero.
On the second point, yes I’d been assuming that a bounded function had a bounded gradient, which certainly isn’t true for say sin(x^2), the final step needs more work, I like the way you did it in the proof below.
Here’s a messy way that at least doesn’t need too much exhaustive search:
First let’s separate all of the red nodes into groups so that within each group you can get to any other node in that group only passing through red nodes, but not to red nodes in any other group.
Now, we trace out the paths that surround these groups—they immediately look like the paths from Question 1 so this feels like a good start. More precisely, we draw out the paths such that each vertex forms one side of a triangle that has a blue node at its opposite corner. Note that you can have multiple paths stemming from the same group if the group touches the side of the larger triangle, or if it has internal holes.
Now we have this set of paths we can split them into three kinds. The first is loops, which arise when you have a group which never touches the edge of the larger triangle, or inside ‘holes’ in large groups. These can be seen as a path starting and finishing at the same node. They therefore have an even number of b-g vertices. The second kind is those that begin at the edge of the large triangle and end at the same edge. These paths begin and end on the same colour and therefore also have an even number of b-g vertices. Finally and most importantly there is a kind of path that goes from one edge to the other -in the case of the reds, the left edge to the right edge. This will happen once with the group that includes the top red node, and if any other group spans the larger triangle then it will generate two more of these paths. Sperner’s lemma tells us that these will have an odd number of b-g vertices and we know that there will be an odd number of such paths, so this final type generates an odd number of total b-g vertices.
By the way that we have defined these paths, the total number of r-g-b triangles is equal to the number of g-b vertices on the paths in the set generated above. This number is the sum of an odd number from the spanning paths and a series of even numbers from the other paths, giving an odd overall number of r-g-b vertices, proving number 4 (as long as I haven’t made an error in categorizing the paths).
I hope this makes sense, let me know if it doesn’t or has errors :)
I was able to get at least (I think) close to proving 2 using Sperner’s Lemma as follows:
You can map the continuous function f(x) to a path of the kind found in Question 1 of length n+1 by evaluating f(x) at x=0, x=1 and n-1 equally spaced divisions between these two points and setting a node as blue if f(x) < 0 else as green.
By Sperner’s Lemma there is an odd, and therefore non-zero number of b-g vertices. You can then take any b-g pair of nodes as the starting points for a new path and repeat the process. After k iterations you have two values of x—only one where f(x) is below zero—that are 1/(n^k) away from each other. We thus can find arbitrarily close points that straddle zero. By taking the sequence f(x) of initial nodes x we get a sequence that, by B-W, has a sub-sequence which converges to zero. By continuity we have proved the existence of an x such that f(x)=0.
We can be sure that the sub-sequence does in fact converge to zero, rather than any other value because if it converges to any number |a|>0, the gradient of f(x) would have to be arbitrarily high to dip back below/above 0 for a value of x arbitrarily close by and therefore would not be a continuous function.
Comments to tighten up/poke holes in the above appreciated :)
For long term bets, where the opportunity cost of tying money up in these bets becomes high, I would have thought that the bets should be denominated in US bonds (or other agreed minimal-risk interest rate asset) to minimize this cost.
Even if the bet does not pay out one way or another, the money still accumulates interest.
Other than being incompatible with Augur, are there any theoretical or practical hurdles to using this? It would hopefully reduce the subsidy required to make an attractive market without incurring cost in and of itself.
Thanks for replying :)
If the joining bonus were large enough to give a new member enough DKP to get the choice items, then older members would (quite rightly) complain. If it were smaller, it wouldn’t work.
I guess my central question is, a new player will have infinite EP/GP after they first receive EP. They can therefore wait until their perfect item comes up, and choose that. This to me seems extremely similar to giving an uncertain but potentially very large joining bonus. After losing this infinite ratio status, the situation then seems very similar to a free market one. In particular I don’t understand why having collected lots of points (ie ability to claim future value) would lead to your incentive dropping off, while accumulating a high ratio (which you’d presumably need to ‘save’ for a while for really top items) doesn’t have this problem.
I’m curious but a bit confused about some of the benefits of EP/GP over the straight free-market model, but if EP/GP did indeed take over then I’m sure there’s something I’m missing.
1: Presumably, in both models, in the long run it takes roughly the same average amount of time (modulated by your efficiency of pro-social activity) to get an item of quality >x, but it seems that in EP/GP you get your first almost immediately, while in DKP your timer starts from 0. Was there the issue of individuals jumping around guilds to try and get that first item?
2: Is there any system by which one can defer the receiving of items in EP/GP so that you don’t end up getting something that is of low or nil value to you (especially since they can’t be traded)? The main advantage of the free-market, at least in systems where individuals have similar ability to earn currency, is usually that items go to those who value them most, so you’d expect DWP to have a big efficiency advantage if you can’t choose whether to accept. On the other hand, if this deferral is possible, would this degenerate into something like a free market, except where new entrants have first dibs over everything?
The power of attracting new players is a valuable advantage I’m sure but it’s the only one that I really see from the 3 given above, and I can’t see how this isn’t possible in a similar way by, say, a free market system where a new member gets some kind of joining bonus.