Hide table of contents

Summary

  • I modeled how much cross-posts on LW get karma based on their quality (proxied by karma on EAF) and their topic.
  • Results are not surprising: cross-posts about AI safety gather more interest than posts about other cause areas or utility maximization.
    • Among posts about AI safety, posts about risks, predictions about future, stories and hypothetical scenarios scale the best with quality.

Methods

Categorization of posts

I categorized every post into one of the categories in bold.


I think that there is more than one way to categorize posts. For instance, inside personal and organizational productivity, rationality cluster there are posts that have a very LW vibe and posts that have a very EAF vibe e.g.

I could only get a few posts that show this, but I think that there is a pattern, where LW is more interested in society and personal scale thinking and much less interested in organizational aspect of things.
 

ai-safety-hypothetical71
ai-safety-factual32
ai-safety-practical31
not-ai-safety-maximization24
not-ai-safety-productivity-rationality20
not-ai-safety-other-cause-areas11

Posts

I used LW and EAF api. I believe this only covers posts that were cross-posted using cross post functionality and not posts that were independently posted on both forums.

Results

I fit a one linear regression model with only interaction terms between category and EA karma (no intercept): 

LW_karma = coef1 × (cat1 × EA_karma) + coef2 ×(cat2 × EA_karma) + ... 

The model explains 59.4% of variance in LW karma (adjusted R^2=0.594, F(6,183)=47.069, p<0.001). Tho it’s important to note that this is a no intercept model and it doesn't make that much sense to look at the R^2 and t-statics and F-statistic. 

VariableEstimateInterpretation
ai-safety-factual × EA karma0.80100 EA karma -> ~81 LW karma
ai-safety-hypothetical × EA karma1.50100 EA karma -> ~151 LW karma
ai-safety-practical × EA karma0.97100 EA karma -> ~97 LW karma
maximization × EA karma0.42100 EA karma -> ~43 LW karma
other-cause-areas × EA karma0.34100 EA karma -> ~35 LW karma
productivity-rationality × EA karma1.61100 EA karma -> ~161 LW karma

I do not report p-values for coefficients since my model is weird (no intercepts) and they are less straightforward to interpret than usually.

Limits

  • The results seem to be highly influenced by outliers.
  • I used very simple statistical model.
  • I use a no-intercept model, which assumes posts with zero EA karma receive zero LW karma. My reasoning was that this is more interpretable that a model where some posts receive some karma just by being there but I guess one could argue that there are always nice people who will just upvote everything. 

6

0
0
1

Reactions

0
0
1

More posts like this

Comments
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities