|September 2nd, 2011|
|utilitarianism, morality [html]|
I see these as depending on two details of utilitarianism that people disagree on. The first is whether we're trying to maximize total or average happiness . Consider a large population with medium average happiness against a small population with high average happiness. The large population is large enough that even though individual happiness isn't that high, total happiness is still higher than that of the small population. Which do we think would be a better population to be humanity? Someone maximizing total happiness would choose the large one, while someone maximizing average happiness would choose the small one.
I believe total happiness is what we should be maximizing. Average happiness gives unreasonable claims like that it's better to have one really happy person than a billion almost as happy people. Some people claim that total happiness leads to the "repugnant conclusion" that we should continue to increase population until there are huge numbers of people with very low but still positive happiness. I would argue that this isn't really that bad. When we say 'positive happiness' we include both suffering and joy. So someone with low positive happiness would believe that on balance their life was worth living and they are glad they got the chance to do so. Average utilitarianism would say that we should limit our population so as to have more resources per person and greater average happiness. Total utilitarianism says we should seek the population size that leads to greatest (total) happiness. If we think we are at the point where additional children decrease global happiness by increasing competition for scarce resources, then we should work to limit overpopulation, but not otherwise.
The other question is whether most people are happy or unhappy; is total happiness positive or negative? If we believe total happiness is positive and likely to remain so, then it would make sense to fund asteroid tracking research  to try to prevent everyone dying in an asteroid collision, removing future potential for happiness. If we believe it's negative and expect it to stay so, then we should spend our money on making sad people happier instead of trying to prevent human extinction. 
 "maximizes utility over all people"
 wikipedia: average and total utilitarianism
 Assuming this is the most cost effective existential risk to be trying to prevent.
 At the extreme, someone who believed total happiness was unavoidably negative should work to quickly and painlessly kill everyone. Perhaps researching bioweapons would make sense. The main way this would fail, though, is killing a lot of people but not everyone, which would dramatically increasing suffering. Also, I disagree (a) that happiness is net negative now and (b) that we should expect to to be in the future, so I think this would be a *really bad idea*.