Itemoids

Caroline Mimbs Nyce

A Better Way to Make New Year’s Resolutions

The Atlantic

www.theatlantic.com › newsletters › archive › 2023 › 12 › new-years-resolutions-time-management-productivity › 676988

This is an edition of The Wonder Reader, a newsletter in which our editors recommend a set of stories to spark your curiosity and fill you with delight. Sign up here to get it every Saturday morning.

Early in 2023, my colleague Caroline Mimbs Nyce chatted with the writer Oliver Burkeman about New Year’s resolutions. Burkeman is an expert on productivity, but he’s arguably also an expert on getting real about the time human beings have on Earth. Burkeman is the author of Four Thousand Weeks: Time Management for Mere Mortals (4,000 weeks is approximately the length of an average American’s life span). In it, he writes: “The average human lifespan is absurdly, terrifyingly, insultingly short.”

With this in mind, Caroline asked Burkeman: “Do you think New Year’s resolutions are worth making, considering we’re all going to die, as your book posits so bluntly?”

Burkeman has hope for the concept of the resolution. “Confronting how short our lives are, and how limited our time is, is actually a sort of precondition for doing meaningful things, including making personal changes,” he told Caroline. But how we go about making these changes matter, he noted; there are healthy and unhealthy ways to do so. In today’s reading list, our writers walk you through the history of the New Year’s resolution, how brands take advantage of it, and how to use it for your own growth.

On Resolutions

Making a New Year’s Resolution? Don’t Go to War With Yourself.

By Caroline Mimbs Nyce

“The difference between not doing anything at all and doing 10 minutes a few times a week is absolute.”

It’s the Most Inadequate Time of the Year

By Amanda Mull

New Year’s resolutions are the perfect opportunity for consumer brands to remind you about all the ways you could be better.

New Year’s Resolutions That Will Actually Lead to Happiness

By Arthur C. Brooks

Set goals to improve your well-being—not your wallet or your waistline.

Still Curious?

The best time-management advice is depressing but liberating: You can make time for things that matter, or you can make time for more email. Make a to-don’t list: When you’re feeling stuck, focusing on the things you hate can help.

Other Diversions

81 things that blew our minds in 2023 How to be happy growing older Taylor Swift at Harvard

P.S.

Here’s one resolution you can make tomorrow: Don’t go out on New Year’s Eve. It’s not worth it, my colleague Julie Beck wrote in 2018.

— Isabel

Building AI Safely Is Getting Harder and Harder

The Atlantic

www.theatlantic.com › newsletters › archive › 2023 › 12 › building-ai-safety-is-getting-harder-and-harder › 676960

This is Atlantic Intelligence, an eight-week series in which The Atlantic’s leading thinkers on AI will help you understand the complexity and opportunities of this groundbreaking technology. Sign up here.

The bedrock of the AI revolution is the internet, or more specifically, the ever-expanding bounty of data that the web makes available to train algorithms. ChatGPT, Midjourney, and other generative-AI models “learn” by detecting patterns in massive amounts of text, images, and videos scraped from the internet. The process entails hoovering up huge quantities of books, art, memes, and, inevitably, the troves of racist, sexist, and illicit material distributed across the web.

Earlier this week, Stanford researchers found a particularly alarming example of that toxicity: The largest publicly available image data set used to train AIs, LAION-5B, reportedly contains more than 1,000 images depicting the sexual abuse of children, out of more than 5 billion in total. A spokesperson for the data set’s creator, the nonprofit Large-scale Artificial Intelligence Open Network, told me in a written statement that it has a “zero tolerance policy for illegal content” and has temporarily halted the distribution of LAION-5B while it evaluates the report’s findings, although this and earlier versions of the data set have already trained prominent AI models.

Because they are free to download, the LAION data sets have been a key resource for start-ups and academics developing AI. It’s notable that researchers have the ability to peer into these data sets to find such awful material at all: There’s no way to know what content is harbored in similar but proprietary data sets from OpenAI, Google, Meta, and other tech companies. One of those researchers is Abeba Birhane, who has been scrutinizing the LAION data sets since the first version’s release, in 2021. Within six weeks, Birhane, a senior fellow at Mozilla who was then studying at University College Dublin, published a paper detailing her findings of sexist, pornographic, and explicit rape imagery in the data. “I’m really not surprised that they found child-sexual-abuse material” in the newest data set, Birhane, who studies algorithmic justice, told me yesterday.

Birhane and I discussed where the problematic content in giant data sets comes from, the dangers it presents, and why the work of detecting this material grows more challenging by the day. Read our conversation, edited for length and clarity, below.

Matteo Wong, assistant editor

More Challenging By the Day

Matteo Wong: In 2021, you studied the LAION data set, which contained 400 million captioned images, and found evidence of sexual violence and other harmful material. What motivated that work?

Abeba Birhane: Because data sets are getting bigger and bigger, 400 million image-and-text pairs is no longer large. But two years ago, it was advertised as the biggest open-source multimodal data set. When I saw it being announced, I was very curious, and I took a peek. The more I looked into the data set, the more I saw really disturbing stuff.

We found there was a lot of misogyny. For example, any benign word that is remotely related to womanhood, like mama, auntie, beautiful—when you queried the data set with those types of terms, it returned a huge proportion of pornography. We also found images of rape, which was really emotionally heavy and intense work, because we were looking at images that are really disturbing. Alongside that audit, we also put forward a lot of questions about what the data-curation community and larger machine-learning community should do about it. We also later found that, as the size of the LAION data sets increased, so did hateful content. By implication, so does any problematic content.

Wong: This week, the biggest LAION data set was removed because of the finding that it contains child-sexual-abuse material. In the context of your earlier research, how do you view this finding?

Birhane: It did not surprise us. These are the issues that we have been highlighting since the first release of the data set. We need a lot more work on data-set auditing, so when I saw the Stanford report, it’s a welcome addition to a body of work that has been investigating these issues.

Wong: Research by yourself and others has continuously found some really abhorrent and often illegal material in these data sets. This may seem obvious, but why is that dangerous?

Birhane: Data sets are the backbone of any machine-learning system. AI didn’t come into vogue over the past 20 years only because of new theories or new methods. AI became ubiquitous mainly because of the internet, because that allowed for mass harvesting of large-scale data sets. If your data contains illegal stuff or problematic representation, then your model will necessarily inherit those issues, and your model output will reflect these problematic representations.

But if we take another step back, to some extent it’s also disappointing to see data sets like the LAION data set being removed. For example, the LAION data set came into existence because the creators wanted to replicate data sets inside big corporations—for example, what data sets used in OpenAI might look like.

Wong: Does this research suggest that tech companies, if they’re using similar methods to collect their data sets, might harbor similar problems?

Birhane: It’s very, very likely, given the findings of previous research. Scale comes at the cost of quality.

Wong: You’ve written about research you couldn’t do on these giant data sets because of the resources necessary. Does scale also come at the cost of auditability? That is, does it become less possible to understand what’s inside these data sets as they become larger?

Birhane: There is a huge asymmetry in terms of resource allocation, where it’s much easier to build stuff but a lot more taxing in terms of intellectual labor, emotional labor, and computational resources when it comes to cleaning up what’s already been assembled. If you look at the history of data-set creation and curation, say 15 to 20 years ago, the data sets were much smaller scale, but there was a lot of human attention that went into detoxifying them. But now, all that human attention to data sets has really disappeared, because these days a lot of that data sourcing has been automated. That makes it cost-effective if you want to build a data set, but the reverse side is that, because data sets are much larger now, they require a lot of resources, including computational resources, and it’s much more difficult to detoxify them and investigate them.

Wong: Data sets are getting bigger and harder to audit, but more and more people are using AI built on that data. What kind of support would you want to see for your work going forward?

Birhane: I would like to see a push for open-sourcing data sets—not just model architectures, but data itself. As horrible as open-source data sets are, if we don’t know how horrible they are, we can’t make them better.

Related:

America already has an AI underclass AI’s present matters more than its imagined future

P.S.

Struggling to find your travel-information and gift-receipt emails during the holidays? You’re not alone. Designing an algorithm to search your inbox is paradoxically much harder than making one to search the entire internet. My colleague Caroline Mimbs Nyce explored why in a recent article.

— Matteo

HBD, GPT

The Atlantic

www.theatlantic.com › newsletters › archive › 2023 › 12 › chatgpt-one-year-atlantic-intelligence › 676208

This is Atlantic Intelligence, an eight-week series in which The Atlantic’s leading thinkers on AI will help you understand the complexity and opportunities of this groundbreaking technology. (Did someone forward you this newsletter? Sign up here.)

One year ago, ChatGPT was released into the world. The startlingly human-sounding chatbot kicked off the generative-AI revolution and quickly became one of the most successful internet applications ever—invading classrooms, workplaces, online stores, and more. The program has created billions of dollars of value where none existed before and enabled entire new skills and fields of work around coaxing useful responses from chatbots.

But perhaps more important than any direct, material change, ChatGPT has sparked our imaginations, feeding many people’s hopes and fears about the rise of intelligent machines: boundless productivity, boundless information, boundless misinformation. “ChatGPT, the tool, is likely less important than ChatGPT, the cultural object,” Charlie Warzel wrote in a new piece for The Atlantic. OpenAI’s premier chatbot is bending how entire industries, government bodies, financial institutions, and media publications think about the future. For the past year, our brains have been trapped in ChatGPT’s world.

Matteo Wong, assistant editor

Illustration by The Atlantic. Source: Getty. One Year In, ChatGPT’s Legacy Is Clear

The technology is less important than the ideas it represents.
By Charlie Warzel

ChatGPT has accomplished a lot in its first trip around the sun. The chatbot has upended or outright killed high-school and college essay writing and thoroughly scrambled the brains of academics, creating an on-campus arms race that professors have already lost. It has been used to write books, article summaries, and political content, and it has flooded online marketplaces with computer-generated slop.

As we’ve gotten to know ChatGPT, we’ve noticed how malleable it is. The li’l bot loves clichés. Its underlying technology has been integrated into internet search. ChatGPT is a time wastera toy—but also, potentially, a labor-force destroyer and a way for machines to leech the remaining humanity out of our jobs. It may even be the harbinger of an unrecognizable world and a “textpocalypse” to come.

Read the full article here.

What to Read Next

ChatGPT might not need to upend the entire world to make some kind of meaningful difference in your life. Below, some of our favorite Atlantic stories about the quieter AI revolution:

ChatGPT will change housework: People who have started using ChatGPT for their personal to-do lists have found it incredibly useful, Kate Cray writes.

You should ask a chatbot to make you a drink: AI is great at coming up with cocktail recipes, even as it fails at other tasks, Caroline Mimbs Nyce writes.

A chatbot is secretly doing my job: For plenty of people, having a robot help them produce serviceable copy will be exactly enough to allow them to get by in the world, Ryan Bradley writes.

P.S.

Trying to decipher why Spotify Wrapped likened your personality to a pumpkin-spice latte? You’re not alone. Frustration with what, exactly, an algorithm reveals about you is the zeitgeist of the modern internet, Nancy Walecki writes.

— Matteo