OCB

Owen Cotton-Barratt

10914 karmaJoined

Sequences
3

Reflection as a strategic goal
On Wholesomeness
Everyday Longermism

Comments
992

Topic contributions
3

Yeah I think that would be an improvement over the current behaviour. I'd still probably prefer something very short ("LLM usage: zero/minimal/moderate/major") which can be expanded if people want more texture.

I wonder if the disclosures could be non-text by default -- e.g. colour-coded with an optional footnote for details. 

The thing I'm not liking as a reader is having words to process on this stuff at the start (for me this isn't just cases where people aren't following policy; I've felt it some about a case where the words were one of the suggested wordings from the policy). Non-text ways to signal could potentially get best-of-both-worlds in terms of reader attention.

Ok so I can kind of tune into what you're saying here, but I also feel kind of uneasy about it. I guess I'd be curious what you make of the following potential arguments: 

  1. Ingredients are important because we can't directly discern what's in food. But with writing we can see exactly what's there and judge that directly without needing to judge the process. (This perspective would endorse reviews being posted warning people not to read low-quality stuff.)
  2. Requiring disclosure is an inappropriate form of thought policing -- people should have the right to use whatever cognitive processes and augmentation methods they like, and take responsibility for the words they then share. If this produces LLM garbage it's not on them to label that up front, but this should have the natural consequence that people stop listening to them.

Ok, so one place the predictions of these theories might come apart is that my theory suggests a norm against impersonating medics, whereas I think yours doesn't (although maybe I'm just not seeing it; I don't think I would have said that avoiding torture of prisoners was part of protecting the mechanisms of ending war, although I do kind of see what you mean). I haven't looked into it at all, but if that norm has emerged independently multiple times that would be suggestive in favour of the broader theory; whereas if it has just emerged once it looks perhaps more potentially-idiosyncratic, which would be suggestive in favour of the narrower theory.

I agree that the model I proposed is imprecise; I think this counts against its usefulness but not its validity.

I'm not suggesting this as a thing to advocate for; merely as a descriptive pattern of what the category of war crimes is doing. I think the things which make ending war harder are an important class of really destructive thing, but it seems clarity-obscuring to me to claim that this is definitionally what war crimes are? Rather than giving your thing a new label and then getting to discuss what fraction of war crimes are in that category, and whether there are things in that category which aren't war crimes (e.g. if torturing POWs counts under your categorization, then why doesn't conscription count -- after all, it damages the "one side runs out of soldiers" mechanism for ending war).

I like the puzzle. But I wonder if you can make your answer even simpler:

  • Actions taken in war have some benefit to the perpetrator, and some costs to the larger system of permitting them
  • When the ratio between these things gets too extreme, it's regarded as a war crime

I think this explains the category that you outline (undermining trust in the kind of institutions that could stop the war is super destructive!), but also explains some other cases, e.g. abuse of prisoners, not impersonating medical staff, etc.

Hmm, I've used LLMs to varying degrees in writing articles. Usually not to the point of writing significant amounts of text, but a case where I think it clearly helped to improve the output is this story: https://strangecities.substack.com/p/some-days-soon

Functionally, I wrote a complete draft, then got Claude to redraft, then I went through and stitched the best bits of the two drafts together (or wrote new versions where that seemed best). (If you thought the original draft was better I'd be interested to hear that: https://docs.google.com/document/d/1icY2wpcgvKszfzHFButKcOwV8B9xMypTAk48kjnOGz0/edit?usp=drivesdk )

(I notice that I'm more likely to find LLMs helpful in drafting things when writing fiction. I think it's least likely to help when it's important to convey my precise epistemic status towards the things I'm saying.)

Requiring disclosures to be at the top of the post (rather than e.g. allowing them to be at the bottom) does feel like it's sending some implicit "this is kind of bad so people need to be warned about it" message, even if it's in a "recommended uses" section.

Like I think people might reasonably worry about others pre-judging posts with this disclaimer, and hence (perhaps, sometimes) prefer workflows where they don't need to include the disclaimer, even if this makes their posts worse.

I don't think there's an easy answer here -- like, presumably the point of the policy is to allow this kind of pre-judging and let people make differently-informed choices about what they engage with. But I think the post kind of papers over this tension.

I think we're still primarily assessing what was said. And you can make your system flexible enough that it doesn't just entrench existing actors.

I think this pushes towards filtering happening via means other than just assessing the argument -- e.g. something like persistent reputations (with some occasional sampling to allow new entrants to get out of the zone of being ignored). cf. discussion of how AI could help us keep tabs on the reliability of different actors.

Load more