Editor’s note: This article is republished as it appears in The Conversation, an independent news publisher that works with academics worldwide to disseminate research-based articles and commentary. The University of Oregon partners with The Conversation to bring the expertise and views of its faculty members to a wide audience. For more information, see the note following this story.
Looking for something to binge-watch while you’re hunkering down at home?
Consider checking out the popular TV show “The Good Place.” Over four recently concluded seasons, the series follows the adventures and mishaps of four utterly self-centered characters on their quest to become decent and selfless human beings.
The deeper question this philosophy-laced comedy raises is: Can people be truly selfless?
The technical term for this behavior is altruism, the willingness to help others even at a cost to your own well-being. And if the answer to that question is yes, then are those of us who are selfish able to transform ourselves into kind and selfless individuals?
I’m a psychologist who uses brain science to understand how people make decisions. With my team at the University of Oregon, I am investigating why many of us behave altruistically, whether human beings become more altruistic with age and even whether it’s possible to learn how to be altruistic.
Whether people do altruistic deeds because of their altruistic nature or out of ulterior motives is a question that has stumped philosophers, religious thinkers and social scientists for centuries, because selfishness can inspire seemingly altruistic acts.
Even Pamela Hieronymi, a University of California, Los Angeles philosopher who informally served as a consultant for the hit TV show, has expressed serious skepticism about whether anyone can turn from selfish to selfless.
How do scholars like me study what goes on in people’s brains?
My team had participants in a series of experiments lie in MRI scanners, looking at a screen that described different scenarios. Sometimes my colleagues and I told them that $20 was being transferred to their bank accounts. At other times, the same amount would go to a charity, such as a local food pantry. Participants simply observed these $20 transfers, either to themselves or to the charity, without having any say in the matter.
This region, which is a little bigger than a peanut, plays a role in everything from sexual gratification to drug addiction and related neural sites. It becomes active when something happens that makes you happy and that you would like to see repeated in the future.
The experience of money going to the charity boosted activity in those reward areas of the brain for many of our participants. And exactly this observation, we argue, is a manifestation of people’s true altruistic nature: They felt rewarded when someone in need becomes better off, even if they didn’t directly do anything to make a difference.
We found that in about half of our study participants, activity in these reward areas was even stronger when the money went to the charity than when it landed in their own bank accounts. We determined that these people could be neurally defined as altruists.
Then, in a separate stage of the experiment, all of these same participants had the choice to either give some of their money away or to keep it for themselves. Here, the neural altruists were about twice as likely as the others to give their money away.
We believe that this finding indicates that purely altruistic motives can drive generous behavior and that brain imaging can detect those motives.
Aging and altruism
In a related study my colleagues and I conducted, there were 80 participants who were between 20 and 64 years old, but otherwise were comparable in terms of their backgrounds. We found that the proportion of altruists, that is, those whose reward areas were more active when money went to the charity than to themselves – steadily increased with age, going from less than 25 percent through age 35 to around 75 percent among individuals 55 and older.
Also, older participants tended to become more willing to give their money to charity or to volunteer in this experiment. And when assessing their personality characteristics through questionnaires, our group found that they exhibited traits such as agreeableness and empathy more strongly than younger participants.
These observations align with growing evidence of more altruistic acts in the elderly. For example, the share of their income that 60-year-olds give to charity is three times as much as for 25-year-olds. This is significant even though they tend to have more money in general, making it easier to part with some of it.
However, our results are the first to clearly demonstrate that older adults do not just act like they are nicer people, which might easily be driven by selfish motives such as making it more likely that they will be remembered fondly once they are gone. Rather, the fact that their reward areas are so much more responsive to experiencing people in need being helped suggests that they are actually, on average, kinder and genuinely more interested in the welfare of others than everyone else.
The road ahead
These findings raise lots of additional, important questions that we cover in an article we published in Current Directions in Psychological Science, an academic journal. For example, additional research is needed in which people are followed across time to make sure that the age difference in generosity truly reflects personal growth, and not just generational differences. Also, we need to generalize our results to larger samples from more varied backgrounds.
Most importantly, we don’t yet know why older adults appear to be more generous than younger folks. My colleagues and I are planning to look into whether realizing that you have fewer years to live makes you more concerned about the greater good.
—By Ulrich Mayr, Lewis Professor and department head of psychology