archermarks 33 minutes ago
This hits like 100% of the AI prose bingo card.
isx726552 27 minutes ago
That’s all well and good, but I think he needs to take a closer look at some of the resulting prose and clarify a little more. Most of it is good, but there are some unclear statements, like this (right after his descriptions of “Camp A” and “Camp B”):
> Sutton says Camp B wins. My essay was filed under Camp A. But decision theory belongs to neither camp.
The second sentence quoted above doesn’t specify, but I’m pretty sure it means that it was filed under Camp A by the commenters, and incorrectly at that. If so, it would probably read better as:
> Sutton says Camp B wins. Commenters seemed to file my essay under Camp A, and then dismissed it. But that’s incorrect; decision theory belongs to neither camp.
Or something along those lines.
I honestly think this isn’t nit-picky feedback, either. This is a crucial set of sentences which appear to lay out the main point of the essay, so it’s vitally important that they be clear … who “filed” it as a particular camp, and was that correct or incorrect? It should be revised to convey that, as well as better connecting that to what incorrect conclusions might have been drawn from that. The information can be gleaned from the surrounding context of course, but I found that crucial sentence to throw off the flow what was otherwise a really great essay.
ordu 34 minutes ago
At that moment things will become really interesting. If decision theory, bayesianism and causality will be able to show something that can be combined with LLM to create something marketable, then they will have their big chance. Or maybe those smart people will devise some other way out of the local maxima.
Bayesian methods and causality has their applications, there are tools to use them, but you can't just feed news into them to get back a most likely structure of a secret global government run by interdimensional lizard people. Probably if you combine them with LLM, than the resulting tool will be able to perform this task?
exmadscientist 29 minutes ago
This paragraph is pretty condescending to your reader. Whatever else is going on with AI authors, the fact is that if your reader can tell you wrote a piece with AI (and I could with this one), you fucked up.
I think one of the longer-term consequences of AI authors will be that writing gets shorter. There's a lot of fluff in a lot of writing (though not as much as there used to be in say the 19th century), and much of it's culturally expected. We might end up at a place where writing is much shorter and readers expect their own AI assistants to fill in the gaps. That might not be so bad.
But if you can't write a piece without AI, do you understand what you've written? It could go either way. But the condescension here combined with the obvious tells do not make me think highly of this author and his argument.
PaulHoule an hour ago
https://en.wikipedia.org/wiki/Mycin
the result is probabilistic in nature, there's always some chance you'll get it wrong.
Language processing is the same. Language is ambiguous, there are thousands of possible parse trees for a common sentence. You might be talking with somebody and then get a piece of information that revises your interpretation of what they said an hour ago. It's just like that.
In that time frame I was very interested in the idea that decision theory was the key link between computation and action whether you were using symbolic methods (e.g. a very plausible set of rules for address matching might be 99.9% reliable in some cases, 97% in others, 2% in others) or learned methods. A model for predicting market prices is priceless, but put that together with a Kelly Better and you've got a trading strategy.
Maybe there is more to his argument than I got, but as I see it he's defending a boundary that isn't there.