I’ve heard a rule of thumb that if you can avoid buying power off the California power grid’s spot pricing in the 1% most expensive times, you can roughly get a 50% discount on power.
Cedar
Making no claim about the actual value of each, but can’t I counter your specific argument by saying, marriage is a socially enforced cartel for sex, and if they could do so without being punished, a lot more men would rather not get sex without getting married?
When you 1-bit quantize all your probability estimates
This thread might be fun for you, where Reddit talks about some papers that draw connections between NNs and decision trees. https://www.reddit.com/r/MachineLearning/comments/y2pi2a/r_neural_networks_are_decision_trees/
In particular, look for the comment that goes
I think your work in this paper is pretty much entirely subsumed by the following work showing that neural networks with piecewise linear activations are equivalent to max-affine spline operators: https://arxiv.org/abs/1805.06576 They seem to cover everything you do and more, although they don’t take a specifically tree-oriented viewpoint. Unfortunately, like many of the others in this thread, I don’t find results like this particularly compelling.I think your work in this paper is pretty much entirely subsumed by the following work showing that neural networks with piecewise linear activations are equivalent to max-affine spline operators: https://arxiv.org/abs/1805.06576
They seem to cover everything you do and more, although they don’t take a specifically tree-oriented viewpoint. Unfortunately, like many of the others in this thread, I don’t find results like this particularly compelling.
The phrase “was predictable” sets off alarm bells for post facto wiggling .
If it’s predicted, I would expect you to say “was predicted”. If it wasn’t predicted due to somebody applying the model wrong, then I would expect you to say “should have been predicted”.
Lun (the account reposting this to LW) is also a very new account with no other activity.
LessWrong scrape dataset on Hugging face, by NousResearch
https://huggingface.co/datasets/LDJnr/LessWrong-Amplify-Instruct
[Question] Are LLMs being trained using LessWrong text?
I personally would like to see less talk about / with Gary Marcus, and more betting with Gary Marcus, like [here](https://garymarcus.substack.com/p/where-will-ai-be-at-the-end-of-2027).
But I understand that people don’t wanna do it because it’s a pretty bad bet if you win money in futures where money is likely going to become worthless.
I am wary of recommending extreme interventions to people without having a lot more context of their situation
Fwiw I think your recommendation just puts steroids onto my radar, and I still feel fully responsible for researching the associated risks. I think it’s fine to push risky interventions, even to push them really hard, fram ng them as a way for people to get something they really care about.
I think you can reasonably recommended testosterone and expect somebody to encounter the appropriate warnings on their way to acquire them.
Also, drone footage cuts out when the drone hits the target.
Yes, for small, explosive drones. But the Vampire drones mentioned in the article are much bigger and are usually used to drop an explosive payload that detaches from the drone and drops onto the enemy.
(Otherwise I agree with your point)
Arbitrage Drains Worse Markets to Feeds Better Ones
Bellevue Library Meetup—Nov 23
@Brendan Long hi hi wanna hang again on the 26th?
Bellevue Meetup
it’s upstairs on fl2. if you see a map that only shows a meeting room 6 on fl2, don’t believe it. ask the front desk staff for room 5!
I’m at the library. Meeting room 5 is upstairs!
By the way, please let me know if you’ll need a ride to the library. I will be starting out from the lake hills area as early as 12:30pm and potentially making a detour through redmond if people need picking up.
https://www.alignmentforum.org/posts/wnzkjSmrgWZaBa2aC/self-preservation-or-instruction-ambiguity-examining-the
Yeah. Here’s a case where, rather than an intrinsic self-preservation drive, task-accomplishment drives self-preservation behavior.
I’d imagine an AI given a task would similarly resist being vaulted because it would interfere with completing the task.