{ "cells": [ { "cell_type": "code", "execution_count": 11, "id": "45eb3404", "metadata": {}, "outputs": [], "source": [ "import feedparser\n", "import pandas as pd\n", "from datetime import datetime, timedelta\n", "import ssl\n", "from bs4 import BeautifulSoup\n", "import warnings\n", "\n", "warnings.filterwarnings(\"ignore\")\n", "\n", "# SSL sertifika doğrulama sorununu çözmek için\n", "if hasattr(ssl, '_create_unverified_context'):\n", " ssl._create_default_https_context = ssl._create_unverified_context" ] }, { "cell_type": "code", "execution_count": 114, "id": "6de6a0d2", "metadata": {}, "outputs": [], "source": [ "links = [\"https://bair.berkeley.edu/blog/feed.xml\",\n", " \"https://feeds.feedburner.com/nvidiablog\",\n", " \"https://www.oreilly.com/radar/feed/\",\n", " \"https://www.microsoft.com/en-us/research/feed/\",\n", " \"https://www.sciencedaily.com/rss/computers_math/artificial_intelligence.xml\",\n", " \"https://research.facebook.com/feed/\",\n", " \"https://openai.com/news/rss.xml\",\n", " \"https://deepmind.google/blog/feed/basic/\",\n", " ]" ] }, { "cell_type": "code", "execution_count": 31, "id": "41917a5b", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "

Artificial intelligence was barely a term in 1956, when top scientists from the field of computing arrived at Dartmouth College for a summer conference. The computer scientist John McCarthy had coined the phrase in the funding proposal for the event, a gathering to work through how to build machines that could use language, solve problems like humans, and improve themselves. But it was a good choice, one that captured the organizers’ founding premise: Any feature of human intelligence could “in principle be so precisely described that a machine can be made to simulate it.” 

\n", "\n", "\n", "\n", "

In their proposal, the group had listed several “aspects of the artificial intelligence problem.” The last item on their list, and in hindsight perhaps the most difficult, was building a machine that could exhibit creativity and originality.

\n", "\n", "\n", "\n", "

At the time, psychologists were grappling with how to define and measure creativity in humans. The prevailing theory—that creativity was a product of intelligence and high IQ—was fading, but psychologists weren’t sure what to replace it with. The Dartmouth organizers had one of their own. “The difference between creative thinking and unimaginative competent thinking lies in the injection of some randomness,” they wrote, adding that such randomness “must be guided by intuition to be efficient.” 

\n", "\n", "\n", "\n", "

Nearly 70 years later, following a number of boom-and-bust cycles in the field, we now have AI models that more or less follow that recipe. While large language models that generate text have exploded in the last three years, a different type of AI, based on what are called diffusion models, is having an unprecedented impact on creative domains. By transforming random noise into coherent patterns, diffusion models can generate new images, videos, or speech, guided by text prompts or other input data. The best ones can create outputs indistinguishable from the work of people, as well as bizarre, surreal results that feel distinctly nonhuman. 

\n", "\n", "\n", "\n", "\n", "\n", "

Now these models are marching into a creative field that is arguably more vulnerable to disruption than any other: music. AI-generated creative works—from orchestra performances to heavy metal—are poised to suffuse our lives more thoroughly than any other product of AI has done yet. The songs are likely to blend into our streaming platforms, party and wedding playlists, soundtracks, and more, whether or not we notice who (or what) made them. 

\n", "\n", "\n", "\n", "

For years, diffusion models have stirred debate in the visual-art world about whether what they produce reflects true creation or mere replication. Now this debate has come for music, an art form that is deeply embedded in our experiences, memories, and social lives. Music models can now create songs capable of eliciting real emotional responses, presenting a stark example of how difficult it’s becoming to define authorship and originality in the age of AI. 

\n", "\n", "\n", "\n", "

The courts are actively grappling with this murky territory. Major record labels are suing the top AI music generators, alleging that diffusion models do little more than replicate human art without compensation to artists. The model makers counter that their tools are made to assist in human creation.  

\n", "\n", "\n", "\n", "

In deciding who is right, we’re forced to think hard about our own human creativity. Is creativity, whether in artificial neural networks or biological ones, merely the result of vast statistical learning and drawn connections, with a sprinkling of randomness? If so, then authorship is a slippery concept. If not—if there is some distinctly human element to creativity—what is it? What does it mean to be moved by something without a human creator? I had to wrestle with these questions the first time I heard an AI-generated song that was genuinely fantastic—it was unsettling to know that someone merely wrote a prompt and clicked “Generate.” That predicament is coming soon for you, too. 

\n", "\n", "\n", "\n", "

Making connections

\n", "\n", "\n", "\n", "

After the Dartmouth conference, its participants went off in different research directions to create the foundational technologies of AI. At the same time, cognitive scientists were following a 1950 call from J.P. Guilford, president of the American Psychological Association, to tackle the question of creativity in human beings. They came to a definition, first formalized in 1953 by the psychologist Morris Stein in the Journal of Psychology: Creative works are both novel, meaning they present something new, and useful, meaning they serve some purpose to someone. Some have called for “useful” to be replaced by “satisfying,” and others have pushed for a third criterion: that creative things are also surprising. 

\n", "\n", "\n", "\n", "

Later, in the 1990s, the rise of functional magnetic resonance imaging made it possible to study more of the neural mechanisms underlying creativity in many fields, including music. Computational methods in the past few years have also made it easier to map out the role that memory and associative thinking play in creative decisions. 

\n", "\n", "\n", "\n", "

What has emerged is less a grand unified theory of how a creative idea originates and unfolds in the brain and more an ever-growing list of powerful observations. We can first divide the human creative process into phases, including an ideation or proposal step, followed by a more critical and evaluative step that looks for merit in ideas. A leading theory on what guides these two phases is called the associative theory of creativity, which posits that the most creative people can form novel connections between distant concepts.

\n", "\n", "\n", "
\n", "
\"""\"
STUART BRADFORD
\n", "
\n", "
\n", "\n", "\n", "

“It could be like spreading activation,” says Roger Beaty, a researcher who leads the Cognitive Neuroscience of Creativity Laboratory at Penn State. “You think of one thing; it just kind of activates related concepts to whatever that one concept is.”

\n", "\n", "\n", "\n", "

These connections often hinge specifically on semantic memory, which stores concepts and facts, as opposed to episodic memory, which stores memories from a particular time and place. Recently, more sophisticated computational models have been used to study how people make connections between concepts across great “semantic distances.” For example, the word apocalypse is more closely related to nuclear power than to celebration. Studies have shown that highly creative people may perceive very semantically distinct concepts as close together. Artists have been found to generate word associations across greater distances than non-artists. Other research has supported the idea that creative people have “leaky” attention—that is, they often notice information that might not be particularly relevant to their immediate task. 

\n", "\n", "\n", "\n", "

Neuroscientific methods for evaluating these processes do not suggest that creativity unfolds in a particular area of the brain. “Nothing in the brain produces creativity like a gland secretes a hormone,” Dean Keith Simonton, a leader in creativity research, wrote in the Cambridge Handbook of the Neuroscience of Creativity

\n", "\n", "\n", "\n", "

The evidence instead points to a few dispersed networks of activity during creative thought, Beaty says—one to support the initial generation of ideas through associative thinking, another involved in identifying promising ideas, and another for evaluation and modification. A new study, led by researchers at Harvard Medical School and published in February, suggests that creativity might even involve the suppression of particular brain networks, like ones involved in self-censorship. 

\n", "\n", "\n", "\n", "

So far, machine creativity—if you can call it that—looks quite different. Though at the time of the Dartmouth conference AI researchers were interested in machines inspired by human brains, that focus had shifted by the time diffusion models were invented, about a decade ago. 

\n", "\n", "\n", "\n", "

The best clue to how they work is in the name. If you dip a paintbrush loaded with red ink into a glass jar of water, the ink will diffuse and swirl into the water seemingly at random, eventually yielding a pale pink liquid. Diffusion models simulate this process in reverse, reconstructing legible forms from randomness.

\n", "\n", "\n", "\n", "

For a sense of how this works for images, picture a photo of an elephant. To train the model, you make a copy of the photo, adding a layer of random black-and-white static on top. Make a second copy and add a bit more, and so on hundreds of times until the last image is pure static, with no elephant in sight. For each image in between, a statistical model predicts how much of the image is noise and how much is really the elephant. It compares its guesses with the right answers and learns from its mistakes. Over millions of these examples, the model gets better at “de-noising” the images and connecting these patterns to descriptions like “male Borneo elephant in an open field.” 

\n", "\n", "\n", "\n", "

Now that it’s been trained, generating a new image means reversing this process. If you give the model a prompt, like “a happy orangutan in a mossy forest,” it generates an image of random white noise and works backward, using its statistical model to remove bits of noise step by step. At first, rough shapes and colors appear. Details come after, and finally (if it works) an orangutan emerges, all without the model “knowing” what an orangutan is.

\n", "\n", "\n", "\n", "

Musical images

\n", "\n", "\n", "\n", "

The approach works much the same way for music. A diffusion model does not “compose” a song the way a band might, starting with piano chords and adding vocals and drums. Instead, all the elements are generated at once. The process hinges on the fact that the many complexities of a song can be depicted visually in a single waveform, representing the amplitude of a sound wave plotted against time. 

\n", "\n", "\n", "\n", "\n", "\n", "

Think of a record player. By traveling along a groove in a piece of vinyl, a needle mirrors the path of the sound waves engraved in the material and transmits it into a signal for the speaker. The speaker simply pushes out air in these patterns, generating sound waves that convey the whole song. 

\n", "\n", "\n", "\n", "

From a distance, a waveform might look as if it just follows a song’s volume. But if you were to zoom in closely enough, you could see patterns in the spikes and valleys, like the 49 waves per second for a bass guitar playing a low G. A waveform contains the summation of the frequencies of all different instruments and textures. “You see certain shapes start taking place,” says David Ding, cofounder of the AI music company Udio, “and that kind of corresponds to the broad melodic sense.” 

\n", "\n", "\n", "\n", "

Since waveforms, or similar charts called spectrograms, can be treated like images, you can create a diffusion model out of them. A model is fed millions of clips of existing songs, each labeled with a description. To generate a new song, it starts with pure random noise and works backward to create a new waveform. The path it takes to do so is shaped by what words someone puts into the prompt.

\n", "\n", "\n", "\n", "

Ding worked at Google DeepMind for five years as a senior research engineer on diffusion models for images and videos, but he left to found Udio, based in New York, in 2023. The company and its competitor Suno, based in Cambridge, Massachusetts, are now leading the race for music generation models. Both aim to build AI tools that enable nonmusicians to make music. Suno is larger, claiming more than 12 million users, and raised a $125 million funding round in May 2024. The company has partnered with artists including Timbaland. Udio raised a seed funding round of $10 million in April 2024 from prominent investors like Andreessen Horowitz as well as musicians Will.i.am and Common.

\n", "\n", "\n", "\n", "

The results of Udio and Suno so far suggest there’s a sizable audience of people who may not care whether the music they listen to is made by humans or machines. Suno has artist pages for creators, some with large followings, who generate songs entirely with AI, often accompanied by AI-generated images of the artist. These creators are not musicians in the conventional sense but skilled prompters, creating work that can’t be attributed to a single composer or singer. In this emerging space, our normal definitions of authorship—and our lines between creation and replication—all but dissolve.

\n", "\n", "\n", "\n", "
\n", "

The results of Udio and Suno so far suggest there’s a sizable audience of people who may not care whether the music they listen to is made by humans or machines.

\n", "
\n", "\n", "\n", "\n", "

The music industry is pushing back. Both companies were sued by major record labels in June 2024, and the lawsuits are ongoing. The labels, including Universal and Sony, allege that the AI models have been trained on copyrighted music “at an almost unimaginable scale” and generate songs that “imitate the qualities of genuine human sound recordings” (the case against Suno cites one ABBA-adjacent song called “Prancing Queen,” for example). 

\n", "\n", "\n", "\n", "

Suno did not respond to requests for comment on the litigation, but in a statement responding to the case posted on Suno’s blog in August, CEO Mikey Shulman said the company trains on music found on the open internet, which “indeed contains copyrighted materials.” But, he argued, “learning is not infringing.”

\n", "\n", "\n", "\n", "

A representative from Udio said the company would not comment on pending litigation. At the time of the lawsuit, Udio released a statement mentioning that its model has filters to ensure that it “does not reproduce copyrighted works or artists’ voices.” 

\n", "\n", "\n", "\n", "

Complicating matters even further is guidance from the US Copyright Office, released in January, that says AI-generated works can be copyrighted if they involve a considerable amount of human input. A month later, an artist in New York received what might be the first copyright for a piece of visual art made with the help of AI. The first song could be next.  

\n", "\n", "\n", "\n", "

Novelty and mimicry

\n", "\n", "\n", "\n", "

These legal cases wade into a gray area similar to one explored by other court battles unfolding in AI. At issue here is whether training AI models on copyrighted content is allowed, and whether generated songs unfairly copy a human artist’s style. 

\n", "\n", "\n", "\n", "

But AI music is likely to proliferate in some form regardless of these court decisions; YouTube has reportedly been in talks with major labels to license their music for AI training, and Meta’s recent expansion of its agreements with Universal Music Group suggests that licensing for AI-generated music might be on the table. 

\n", "\n", "\n", "\n", "

If AI music is here to stay, will any of it be any good? Consider three factors: the training data, the diffusion model itself, and the prompting. The model can only be as good as the library of music it learns from and the descriptions of that music, which must be complex to capture it well. A model’s architecture then determines how well it can use what’s been learned to generate songs. And the prompt you feed into the model—as well as the extent to which the model “understands” what you mean by “turn down that saxophone,” for example—is pivotal too.

\n", "\n", "\n", "\n", "
\n", "

Is the result creation or simply replication of the training data? We could ask the same question about human creativity.

\n", "
\n", "\n", "\n", "\n", "

Arguably the most important issue is the first: How extensive and diverse is the training data, and how well is it labeled? Neither Suno nor Udio has disclosed what music has gone into its training set, though these details will likely have to be disclosed during the lawsuits. 

\n", "\n", "\n", "\n", "

Udio says the way those songs are labeled is essential to the model. “An area of active research for us is: How do we get more and more refined descriptions of music?” Ding says. A basic description would identify the genre, but then you could also say whether a song is moody, uplifting, or calm. More technical descriptions might mention a two-five-one chord progression or a specific scale. Udio says it does this through a combination of machine and human labeling. 

\n", "\n", "\n", "\n", "

“Since we want to target a broad range of target users, that also means that we need a broad range of music annotators,” he says. “Not just people with music PhDs who can describe the music on a very technical level, but also music enthusiasts who have their own informal vocabulary for describing music.”

\n", "\n", "\n", "\n", "

Competitive AI music generators must also learn from a constant supply of new songs made by people, or else their outputs will be stuck in time, sounding stale and dated. For this, today’s AI-generated music relies on human-generated art. In the future, though, AI music models may train on their own outputs, an approach being experimented with in other AI domains.

\n", "\n", "\n", "\n", "

Because models start with a random sampling of noise, they are nondeterministic; giving the same AI model the same prompt will result in a new song each time. That’s also because many makers of diffusion models, including Udio, inject additional randomness through the process—essentially taking the waveform generated at each step and distorting it ever so slightly in hopes of adding imperfections that serve to make the output more interesting or real. The organizers of the Dartmouth conference themselves recommended such a tactic back in 1956.

\n", "\n", "\n", "\n", "\n", "\n", "

According to Udio co­founder and chief operating officer Andrew Sanchez, it’s this randomness inherent in generative AI programs that comes as a shock to many people. For the past 70 years, computers have executed deterministic programs: Give the software an input and receive the same response every time. 

\n", "\n", "\n", "\n", "

“Many of our artists partners will be like, ‘Well, why does it do this?’” he says. “We’re like, well, we don’t really know.” The generative era requires a new mindset, even for the companies creating it: that AI programs can be messy and inscrutable.

\n", "\n", "\n", "\n", "

Is the result creation or simply replication of the training data? Fans of AI music told me we could ask the same question about human creativity. As we listen to music through our youth, neural mechanisms for learning are weighted by these inputs, and memories of these songs influence our creative outputs. In a recent study, Anthony Brandt, a composer and professor of music at Rice University, pointed out that both humans and large language models use past experiences to evaluate possible future scenarios and make better choices. 

\n", "\n", "\n", "\n", "

Indeed, much of human art, especially in music, is borrowed. This often results in litigation, with artists alleging that a song was copied or sampled without permission. Some artists suggest that diffusion models should be made more transparent, so we could know that a given song’s inspiration is three parts David Bowie and one part Lou Reed. Udio says there is ongoing research to achieve this, but right now, no one can do it reliably. 

\n", "\n", "\n", "\n", "

For great artists, “there is that combination of novelty and influence that is at play,” Sanchez says. “And I think that that’s something that is also at play in these technologies.”

\n", "\n", "\n", "\n", "

But there are lots of areas where attempts to equate human neural networks with artificial ones quickly fall apart under scrutiny. Brandt carves out one domain where he sees human creativity clearly soar above its machine-made counterparts: what he calls “amplifying the anomaly.” AI models operate in the realm of statistical sampling. They do not work by emphasizing the exceptional but, rather, by reducing errors and finding probable patterns. Humans, on the other hand, are intrigued by quirks. “Rather than being treated as oddball events or ‘one-offs,’” Brandt writes, the quirk “permeates the creative product.” 

\n", "\n", "\n", "
\n", "
\"""\"
STUART BRADFORD
\n", "
\n", "
\n", "\n", "\n", "

He cites Beethoven’s decision to add a jarring off-key note in the last movement of his Symphony no. 8. “Beethoven could have left it at that,” Brandt says. “But rather than treating it as a one-off, Beethoven continues to reference this incongruous event in various ways. In doing so, the composer takes a momentary aberration and magnifies its impact.” One could look to similar anomalies in the backward loop sampling of late Beatles recordings, pitched-up vocals from Frank Ocean, or the incorporation of “found sounds,” like recordings of a crosswalk signal or a door closing, favored by artists like Charlie Puth and by Billie Eilish’s producer Finneas O’Connell. 

\n", "\n", "\n", "\n", "

If a creative output is indeed defined as one that’s both novel and useful, Brandt’s interpretation suggests that the machines may have us matched on the second criterion while humans reign supreme on the first. 

\n", "\n", "\n", "\n", "

To explore whether that is true, I spent a few days playing around with Udio’s model. It takes a minute or two to generate a 30-second sample, but if you have paid versions of the model you can generate whole songs. I decided to pick 12 genres, generate a song sample for each, and then find similar songs made by people. I built a quiz to see if people in our newsroom could spot which songs were made by AI. 

\n", "\n", "\n", "\n", "

The average score was 46%. And for a few genres, especially instrumental ones, listeners were wrong more often than not. When I watched people do the test in front of me, I noticed that the qualities they confidently flagged as a sign of composition by AI—a fake-sounding instrument, a weird lyric—rarely proved them right. Predictably, people did worse in genres they were less familiar with; some did okay on country or soul, but many stood no chance against jazz, classical piano, or pop. Beaty, the creativity researcher, scored 66%, while Brandt, the composer, finished at 50% (though he answered correctly on the orchestral and piano sonata tests). 

\n", "\n", "\n", "\n", "
\n", "\n", "\n", "\n", "

Remember that the model doesn’t deserve all the credit here; these outputs could not have been created without the work of human artists whose work was in the training data. But with just a few prompts, the model generated songs that few people would pick out as machine-made. A few could easily have been played at a party without raising objections, and I found two I genuinely loved, even as a lifelong musician and generally picky music person. But sounding real is not the same thing as sounding original. The songs did not feel driven by oddities or anomalies—certainly not on the level of Beethoven’s “jump scare.” Nor did they seem to bend genres or cover great leaps between themes. In my test, people sometimes struggled to decide whether a song was AI-generated or simply bad. 

\n", "\n", "\n", "\n", "

How much will this matter in the end? The courts will play a role in deciding whether AI music models serve up replications or new creations—and how artists are compensated in the process—but we, as listeners, will decide their cultural value. To appreciate a song, do we need to picture a human artist behind it—someone with experience, ambitions, opinions? Is a great song no longer great if we find out it’s the product of AI? 

\n", "\n", "\n", "\n", "

Sanchez says people may wonder who is behind the music. But “at the end of the day, however much AI component, however much human component, it’s going to be art,” he says. “And people are going to react to it on the quality of its aesthetic merits.”

\n", "\n", "\n", "\n", "

In my experiment, though, I saw that the question really mattered to people—and some vehemently resisted the idea of enjoying music made by a computer model. When one of my test subjects instinctively started bobbing her head to an electro-pop song on the quiz, her face expressed doubt. It was almost as if she was trying her best to picture a human rather than a machine as the song’s composer. “Man,” she said, “I really hope this isn’t AI.” 

\n", "\n", "\n", "\n", "

It was. 

\n", "

MIT Technology Review Explains: Let our writers untangle the complex, messy world of technology to help you understand what’s coming next. You can read more from the series here.

\n", "\n", "\n", "\n", "

When OpenAI cofounder Andrej Karpathy excitedly took to X back in February to post about his new hobby, he probably had no idea he was about to coin a phrase that encapsulated an entire movement steadily gaining momentum across the world.

\n", "\n", "\n", "\n", "

“There’s a new kind of coding I call ‘vibe coding’, where you fully give in to the vibes, embrace exponentials, and forget that the code even exists,” he said. “I’m building a project or webapp, but it’s not really coding—I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works.” 

\n", "\n", "\n", "\n", "

If this all sounds very different from poring over lines of code, that’s because Karpathy was talking about a particular style of coding with AI assistance. His words struck a chord among software developers and enthusiastic amateurs alike. In the months since, his post has sparked think pieces and impassioned debates across the internet. But what exactly is vibe coding? Who does it benefit, and what’s its likely future?

\n", "\n", "\n", "\n", "

So, what is it?

\n", "\n", "\n", "\n", "

To truly understand vibe coding, it’s important to note that while the term may be new, the coding technology behind it isn’t. For the past few years, general-purpose chatbots like Anthropic’s Claude, OpenAI’s ChatGPT, and Google DeepMind’s Gemini have been getting better at writing code to build software, including games, websites, and apps. But it’s the recent advent of specially created AI coding assistants, including Cursor’s Chat (previously known as Composer) and GitHub Copilot, that really ushered in vibe coding. These assistants can make real-time predictions about what you’re trying to do and offer intuitive suggestions to make it easier than ever to create software, even if you’ve never written code before.

\n", "\n", "\n", "\n", "

“Over the past three or four years, these AI autocomplete tools have become better and better—they started off completing single lines of code and can now rewrite an entire file for you, or create new components,” says Barron Webster, a software designer at the interface company Sandbar. “The remit of what you can take your hands off the wheel and let the machine do is continually growing over time.”  

\n", "\n", "\n", "\n", "

… and what doesn’t count as vibe coding?

\n", "\n", "\n", "\n", "

But not all AI-assisted coding is vibe coding. To truly vibe-code, you have to be prepared to let the AI fully take control and refrain from checking and directly tweaking the code it generates as you go along—surrendering to the vibes. In Karpathy’s longer post he explained that when he’s vibe coding, he breezily accepts all suggestions that Cursor’s tool gives him and puts his trust in its ability to fix its own mistakes. “When I get error messages I just copy paste them in with no comment, usually that fixes it,” he wrote. “Sometimes the LLMs can’t fix a bug so I just work around it or ask for random changes until it goes away.”

\n", "\n", "\n", "\n", "\n", "\n", "

Essentially, vibe coding is interacting with a code base through prompts, so that the engineer’s role is simply to converse with the tool and examine its outcome, explains Sergey Tselovalnikov, a software engineer at the design platform Canva who regularly uses AI assistive tools. “Andrej is a bit of an influencer, and he defined that term very intentionally,” he says. “He just posted a joke of sorts, but because he highlighted what was going on in the industry more or less correctly, it just took off.”

\n", "\n", "\n", "\n", "

Is vibe coding right for my project?

\n", "\n", "\n", "\n", "

The people most likely to benefit from vibe coding fall into two camps, says Tobin South, an AI security researcher at the MIT Media Lab. One is people like Karpathy, who already have a good grasp of coding and know how to fix any errors if anything goes seriously wrong if they’re using it to build anything important; the other is absolute amateurs with little to no coding experience. “I’d define vibe coding as having a vision that you can’t execute, but AI can,” he says.

\n", "\n", "\n", "\n", "

The major appeal of vibe coding lies in how easy and accessible it is. The AI assistive tools make it much quicker to produce code and to whip up small projects like a prototype website, game, or web app than it would be for a human. But while this hands-off approach may make sense when it comes to creating these kinds of low-stakes, simple digital products, it’s far riskier in bigger, more complex systems where the stakes are much higher. Because AI coding tools are powered by LLMs, the code they generate is just as likely to contain errors as the answers LLM-powered chatbots spit out. That’s a big problem if what you’re trying to code requires access to large databases of information, security measures to protect that data, large numbers of users, or data inputted from users, says Tselovalnikov.

\n", "\n", "\n", "\n", "

“Vibe coding can make a lot of errors and problems, but in the environment of a tiny game or a small app that doesn’t store any data, it’s a lot less relevant,” he says. “I’d personally be a lot more careful with larger projects, because if you don’t know if there are any security vulnerabilities and you didn’t test the code yourself, that’s very dangerous.”

\n", "\n", "\n", "\n", "\n", "\n", "

This is particularly applicable to non-coders. Leo, a user on X and a champion of vibe coding, found this out the hard way when he posted about having built a SaaS application (software that runs over the internet, instead of being downloaded to a user’s device) solely using Cursor last month. The post immediately caught the attention of mischievous web users, who instantly started poking holes in his service’s security. “Guys, I’m under attack,” he posted two days later. “I’m not technical, so this is taking me longer than usual to figure out. For now, I will stop sharing what I do publicly on X. There are just some weird ppl out there.”

\n", "\n", "\n", "\n", "

Ultimately, while vibe coding can help make a vague idea for a website or a game into a reality, it can’t make it reliable or secure. But there are already plenty of existing tools to do this, helping you with everything from creating databases to adding authentication measures. So while you can’t vibe-code real, valuable, secure, robust apps into existence, it can be a useful place to start so long as you’re careful, says South. 

\n", "\n", "\n", "\n", "

He believes that AI-assisted coding assistants are going to keep becoming more capable and that web hosting companies will keep integrating AI into their tools to make them easier to use, meaning the barriers to creating software will keep falling. “It takes the cost of producing software and dramatically reduces it to an exponential degree,” he says. “The world will have to adapt to this new reality. It isn’t going anywhere.”

\n", "

Bowling Green, Kentucky, is home to 75,000 residents who recently wrapped up an experiment in using AI for democracy: Can an online polling platform, powered by machine learning, capture what residents want to see happen in their city?

\n", "\n", "\n", "\n", "

When Doug Gorman, elected leader of the county that includes Bowling Green, took office in 2023, it was the fastest-growing city in the state and projected to double in size by 2050, but it lacked a plan for how that growth would unfold. Gorman had a meeting with Sam Ford, a local consultant who had worked with the surveying platform Pol.is, which uses machine learning to gather opinions from large groups of people. 

\n", "\n", "\n", "\n", "

They “needed a vision” for the anticipated growth, Ford says. The two convened a group of volunteers with experience in eight areas: economic development, talent, housing, public health, quality of life, tourism, storytelling, and infrastructure. They built a plan to use Pol.is to help write a 25-year plan for the city. The platform is just one of several new technologies used in Europe and increasingly in the US to help make sure that local governance is informed by public opinion.

\n", "\n", "\n", "\n", "

After a month of advertising, the Pol.is portal launched in February. Residents could go to the website and anonymously submit an idea (in less than 140 characters) for what the 25-year plan should include. They could also vote on whether they agreed or disagreed with other ideas. The tool could be translated into a participant’s preferred language, and human moderators worked to make sure the traffic was coming from the Bowling Green area. 

\n", "\n", "\n", "\n", "

Over the month that it was live, 7,890 residents participated, and 2,000 people submitted their own ideas. An AI-powered tool from Google Jigsaw then analyzed the data to find what people agreed and disagreed on. 

\n", "\n", "\n", "\n", "

Experts on democracy technologies who were not involved in the project say this level of participation—about 10% of the city’s residents—was impressive.

\n", "\n", "\n", "\n", "

“That is a lot,” says Archon Fung, director of the Ash Center for Innovation and Democratic Governance at the Harvard Kennedy School. A local election might see a 25% turnout, he says, and that requires nothing more than filling out a ballot. 

\n", "\n", "\n", "\n", "

“Here, it’s a more demanding kind of participation, right? You’re actually voting on or considering some substantive things, and 2,000 people are contributing ideas,” he says. “So I think that’s a lot of people who are engaged.”

\n", "\n", "\n", "\n", "

The plans that received the most attention in the Bowling Green experiment were hyperlocal. The ideas with the broadest support were increasing the number of local health-care specialists so residents wouldn’t have to travel to nearby Nashville for medical care, enticing more restaurants and grocery stores to open on the city’s north side, and preserving historic buildings. 

\n", "\n", "\n", "\n", "\n", "\n", "

More contentious ideas included approving recreational marijuana, adding sexual orientation and gender identity to the city’s nondiscrimination clause, and providing more options for private education. Out of 3,940 unique ideas, 2,370 received more than 80% agreement, including initiatives like investing in stormwater infrastructure and expanding local opportunities for children and adults with autism.  

\n", "\n", "\n", "\n", "

The volunteers running the experiment were not completely hands-off. Submitted ideas were screened according to a moderation policy, and redundant ideas were not posted. Ford says that 51% of ideas were published, and 31% were deemed redundant. About 6% of ideas were not posted because they were either completely off-topic or contained a personal attack.

\n", "\n", "\n", "\n", "

But some researchers who study the technologies that can make democracy more effective question whether soliciting input in this manner is a reliable way to understand what a community wants.

\n", "\n", "\n", "\n", "

One problem is self-selection—for example, certain kinds of people tend to show up to in-person forums like town halls. Research shows that seniors, homeowners, and people with high levels of education are the most likely to attend, Fung says. It’s possible that similar dynamics are at play among the residents of Bowling Green who decided to participate in the project.

\n", "\n", "\n", "\n", "

“Self-selection is not an adequate way to represent the opinions of a public,” says James Fishkin, a political scientist at Stanford who’s known for developing a process he calls deliberative polling, in which a representative sample of a population’s residents are brought together for a weekend, paid about $300 each for their participation, and asked to deliberate in small groups. Other methods, used in some European governments, use jury-style groups of residents to make public policy decisions. 

\n", "\n", "\n", "\n", "

What’s clear to everyone who studies the effectiveness of these tools is that they promise to move a city in a more democratic direction, but we won’t know if Bowling Green’s experiment worked until residents see what the city does with the ideas that they raised.

\n", "\n", "\n", "\n", "

“You can’t make policy based on a tweet,” says Beth Simone Noveck, who directs a lab that studies democracy and technology at Northeastern University. As she points out, residents were voting on 140-character ideas, and those now need to be formed into real policies. 

\n", "\n", "\n", "\n", "

“What comes next,” she says, “is the conversation between the city and residents to develop a short proposal into something that can actually be implemented.” For residents to trust that their voice actually matters, the city must be clear on why it’s implementing some ideas and not others. 

\n", "\n", "\n", "\n", "

For now, the organizers have made the results public, and they will make recommendations to the Warren County leadership later this year. 

\n", "

Last week, I spoke with two US Marines who spent much of last year deployed in the Pacific, conducting training exercises from South Korea to the Philippines. Both were responsible for analyzing surveillance to warn their superiors about possible threats to the unit. But this deployment was unique: For the first time, they were using generative AI to scour intelligence, through a chatbot interface similar to ChatGPT. 

\n", "\n", "\n", "\n", "

As I wrote in my new story, this experiment is the latest evidence of the Pentagon’s push to use generative AI—tools that can engage in humanlike conversation—throughout its ranks, for tasks including surveillance. Consider this phase two of the US military’s AI push, where phase one began back in 2017 with older types of AI, like computer vision to analyze drone imagery. Though this newest phase began under the Biden administration, there’s fresh urgency as Elon Musk’s DOGE and Secretary of Defense Pete Hegseth push loudly for AI-fueled efficiency. 

\n", "\n", "\n", "\n", "

As I also write in my story, this push raises alarms from some AI safety experts about whether large language models are fit to analyze subtle pieces of intelligence in situations with high geopolitical stakes. It also accelerates the US toward a world where AI is not just analyzing military data but suggesting actions—for example, generating lists of targets. Proponents say this promises greater accuracy and fewer civilian deaths, but many human rights groups argue the opposite. 

\n", "\n", "\n", "\n", "

With that in mind, here are three open questions to keep your eye on as the US military, and others around the world, bring generative AI to more parts of the so-called “kill chain.”

\n", "\n", "\n", "\n", "

What are the limits of “human in the loop”?

\n", "\n", "\n", "\n", "

Talk to as many defense-tech companies as I have and you’ll hear one phrase repeated quite often: “human in the loop.” It means that the AI is responsible for particular tasks, and humans are there to check its work. It’s meant to be a safeguard against the most dismal scenarios—AI wrongfully ordering a deadly strike, for example—but also against more trivial mishaps. Implicit in this idea is an admission that AI will make mistakes, and a promise that humans will catch them.

\n", "\n", "\n", "\n", "

But the complexity of AI systems, which pull from thousands of pieces of data, make that a herculean task for humans, says Heidy Khlaaf, who is chief AI scientist at the AI Now Institute, a research organization, and previously led safety audits for AI-powered systems.

\n", "\n", "\n", "\n", "

“‘Human in the loop’ is not always a meaningful mitigation,” she says. When an AI model relies on thousands of data points to draw conclusions, “it wouldn’t really be possible for a human to sift through that amount of information to determine if the AI output was erroneous.” As AI systems rely on more and more data, this problem scales up. 

\n", "\n", "\n", "\n", "

Is AI making it easier or harder to know what should be classified?

\n", "\n", "\n", "\n", "

In the Cold War era of US military intelligence, information was captured through covert means, written up into reports by experts in Washington, and then stamped “Top Secret,” with access restricted to those with proper clearances. The age of big data, and now the advent of generative AI to analyze that data, is upending the old paradigm in lots of ways.

\n", "\n", "\n", "\n", "

One specific problem is called classification by compilation. Imagine that hundreds of unclassified documents all contain separate details of a military system. Someone who managed to piece those together could reveal important information that on its own would be classified. For years, it was reasonable to assume that no human could connect the dots, but this is exactly the sort of thing that large language models excel at. 

\n", "\n", "\n", "\n", "

With the mountain of data growing each day, and then AI constantly creating new analyses, “I don’t think anyone’s come up with great answers for what the appropriate classification of all these products should be,” says Chris Mouton, a senior engineer for RAND, who recently tested how well suited generative AI is for intelligence and analysis. Underclassifying is a US security concern, but lawmakers have also criticized the Pentagon for overclassifying information. 

\n", "\n", "\n", "\n", "\n", "\n", "

The defense giant Palantir is positioning itself to help, by offering its AI tools to determine whether a piece of data should be classified or not. It’s also working with Microsoft on AI models that would train on classified data. 

\n", "\n", "\n", "\n", "

How high up the decision chain should AI go?

\n", "\n", "\n", "\n", "

Zooming out for a moment, it’s worth noting that the US military’s adoption of AI has in many ways followed consumer patterns. Back in 2017, when apps on our phones were getting good at recognizing our friends in photos, the Pentagon launched its own computer vision effort, called Project Maven, to analyze drone footage and identify targets.

\n", "\n", "\n", "\n", "

Now, as large language models enter our work and personal lives through interfaces such as ChatGPT, the Pentagon is tapping some of these models to analyze surveillance. 

\n", "\n", "\n", "\n", "

So what’s next? For consumers, it’s agentic AI, or models that can not just converse with you and analyze information but go out onto the internet and perform actions on your behalf. It’s also personalized AI, or models that learn from your private data to be more helpful. 

\n", "\n", "\n", "\n", "

All signs point to the prospect that military AI models will follow this trajectory as well. A report published in March from Georgetown’s Center for Security and Emerging Technology found a surge in military adoption of AI to assist in decision-making. “Military commanders are interested in AI’s potential to improve decision-making, especially at the operational level of war,” the authors wrote.

\n", "\n", "\n", "\n", "

In October, the Biden administration released its national security memorandum on AI, which provided some safeguards for these scenarios. This memo hasn’t been formally repealed by the Trump administration, but President Trump has indicated that the race for competitive AI in the US needs more innovation and less oversight. Regardless, it’s clear that AI is quickly moving up the chain not just to handle administrative grunt work, but to assist in the most high-stakes, time-sensitive decisions. 

\n", "\n", "\n", "\n", "

I’ll be following these three questions closely. If you have information on how the Pentagon might be handling these questions, please reach out via Signal at jamesodonnell.22. 

\n", "\n", "\n", "\n", "

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

\n", "\n", "\n", "\n", "

\n", "

In 2021, 20 years after the death of her older sister, Vauhini Vara was still unable to tell the story of her loss. “I wondered,” she writes in Searches, her new collection of essays on AI technology, “if Sam Altman’s machine could do it for me.” So she tried GPT-3. But as it expanded on Vara’s prompts in sentences ranging from the stilted to the unsettling to the sublime, the thing she’d enlisted as a tool stopped seeming so mechanical. 

\n", "\n", "\n", "\n", "

“Once upon a time, she taught me to exist,” the AI model wrote of the young woman Vara had idolized. Vara, a journalist and novelist, called the resulting essay “Ghosts,” and in her opinion, the best lines didn’t come from her: “I found myself irresistibly attracted to GPT-3—to the way it offered, without judgment, to deliver words to a writer who has found herself at a loss for them … as I tried to write more honestly, the AI seemed to be doing the same.”

\n", "\n", "\n", "\n", "\n", "\n", "

The rapid proliferation of AI in our lives introduces new challenges around authorship, authenticity, and ethics in work and art. But it also offers a particularly human problem in narrative: How can we make sense of these machines, not just use them? And how do the words we choose and stories we tell about technology affect the role we allow it to take on (or even take over) in our creative lives? Both Vara’s book and The Uncanny Muse, a collection of essays on the history of art and automation by the music critic David Hajdu, explore how humans have historically and personally wrestled with the ways in which machines relate to our own bodies, brains, and creativity. At the same time, The Mind Electric, a new book by a neurologist, Pria Anand, reminds us that our own inner workings may not be so easy to replicate.

\n", "\n", "\n", "\n", "

Searches is a strange artifact. Part memoir, part critical analysis, and part AI-assisted creative experimentation, Vara’s essays trace her time as a tech reporter and then novelist in the San Francisco Bay Area alongside the history of the industry she watched grow up. Tech was always close enough to touch: One college friend was an early Google employee, and when Vara started reporting on Facebook (now Meta), she and Mark Zuckerberg became “friends” on his platform. In 2007, she published a scoop that the company was planning to introduce ad targeting based on users’ personal information—the first shot fired in the long, gnarly data war to come. In her essay “Stealing Great Ideas,” she talks about turning down a job reporting on Apple to go to graduate school for fiction. There, she wrote a novel about a tech founder, which was later published as The Immortal King Rao. Vara points out that in some ways at the time, her art was “inextricable from the resources [she] used to create it”—products like Google Docs, a MacBook, an iPhone. But these pre-AI resources were tools, plain and simple. What came next was different.

\n", "\n", "\n", "\n", "

Interspersed with Vara’s essays are chapters of back-and-forths between the author and ChatGPT about the book itself, where the bot serves as editor at Vara’s prompting. ChatGPT obligingly summarizes and critiques her writing in a corporate-­shaded tone that’s now familiar to any knowledge worker. “If there’s a place for disagreement,” it offers about the first few chapters on tech companies, “it might be in the balance of these narratives. Some might argue that the ­benefits—such as job creation, innovation in various sectors like AI and logistics, and contributions to the global economy—can outweigh the negatives.” 

\n", "\n", "\n", "
\n", "
\"book
Searches: Selfhood in the Digital Age
Vauhini Vara
PANTHEON, 2025
\n", "
\n", "
\n", "\n", "\n", "
\n", "

Vara notices that ChatGPT writes “we” and “our” in these responses, pulling it into the human story, not the tech one: “Earlier you mentioned ‘our access to information’ and ‘our collective experiences and understandings.’” When she asks what the rhetorical purpose of that choice is, ChatGPT responds with a numbered list of benefits including “inclusivity and solidarity” and “neutrality and objectivity.” It adds that “using the first-person plural helps to frame the discussion in terms of shared human experiences and collective challenges.” Does the bot believe it’s human? Or at least, do the humans who made it want other humans to believe it does? “Can corporations use these [rhetorical] tools in their products too, to subtly make people identify with, and not in opposition to, them?” Vara asks. ChatGPT replies, “Absolutely.”

\n", "\n", "\n", "\n", "

Vara has concerns about the words she’s used as well. In “Thank You for Your Important Work,” she worries about the impact of “Ghosts,” which went viral after it was first published. Had her writing helped corporations hide the reality of AI behind a velvet curtain? She’d meant to offer a nuanced “provocation,” exploring how uncanny generative AI can be. But instead, she’d produced something beautiful enough to resonate as an ad for its creative potential. Even Vara herself felt fooled. She particularly loved one passage the bot wrote, about Vara and her sister as kids holding hands on a long drive. But she couldn’t imagine either of them being so sentimental. What Vara had elicited from the machine, she realized, was “wish fulfillment,” not a haunting. 

\n", "
\n", "\n", "\n", "\n", "
\n", "

The rapid proliferation of AI in our lives introduces new challenges around authorship, authenticity, and ethics in work and art. How can we make sense of these machines, not just use them? 

\n", "
\n", "\n", "\n", "\n", "

The machine wasn’t the only thing crouching behind that too-good-to-be-true curtain. The GPT models and others are trained through human labor, in sometimes exploitative conditions. And much of the training data was the creative work of human writers before her. “I’d conjured artificial language about grief through the extraction of real human beings’ language about grief,” she writes. The creative ghosts in the model were made of code, yes, but also, ultimately, made of people. Maybe Vara’s essay helped cover up that truth too.

\n", "\n", "\n", "\n", "

In the book’s final essay, Vara offers a mirror image of those AI call-and-­response exchanges as an antidote. After sending out an anonymous survey to women of various ages, she presents the replies to each question, one after the other. “Describe something that doesn’t exist,” she prompts, and the women respond: “God.” “God.” “God.” “Perfection.” “My job. (Lost it.)” Real people contradict each other, joke, yell, mourn, and reminisce. Instead of a single authoritative voice—an editor, or a company’s limited style guide—Vara gives us the full gasping crowd of human creativity. “What’s it like to be alive?” Vara asks the group. “It depends,” one woman answers.    

\n", "\n", "\n", "\n", "\n", "\n", "

David Hajdu, now music editor at The Nation and previously a music critic for The New Republic, goes back much further than the early years of Facebook to tell the history of how humans have made and used machines to express ourselves. Player pianos, microphones, synthesizers, and electrical instruments were all assistive technologies that faced skepticism before acceptance and, sometimes, elevation in music and popular culture. They even influenced the kind of art people were able to and wanted to make. Electrical amplification, for instance, allowed singers to use a wider vocal range and still reach an audience. The synthesizer introduced a new lexicon of sound to rock music. “What’s so bad about being mechanical, anyway?” Hajdu asks in The Uncanny Muse. And “what’s so great about being human?” 

\n", "\n", "\n", "
\n", "
\"book
The Uncanny Muse: Music, Art, and Machines from Automata to AI
David Hajdu
W.W. NORTON & COMPANY, 2025
\n", "
\n", "
\n", "\n", "\n", "

But Hajdu is also interested in how intertwined the history of man and machine can be, and how often we’ve used one as a metaphor for the other. Descartes saw the body as empty machinery for consciousness, he reminds us. Hobbes wrote that “life is but a motion of limbs.” Freud described the mind as a steam engine. Andy Warhol told an interviewer that “everybody should be a machine.” And when computers entered the scene, humans used them as metaphors for themselves too. “Where the machine model had once helped us understand the human body … a new category of machines led us to imagine the brain (how we think, what we know, even how we feel or how we think about what we feel) in terms of the computer,” Hajdu writes. 

\n", "\n", "\n", "\n", "

But what is lost with these one-to-one mappings? What happens when we imagine that the complexity of the brain—an organ we do not even come close to fully understanding—can be replicated in 1s and 0s? Maybe what happens is we get a world full of chatbots and agents, computer-­generated artworks and AI DJs, that companies claim are singular creative voices rather than remixes of a million human inputs. And perhaps we also get projects like the painfully named Painting Fool—an AI that paints, developed by Simon Colton, a scholar at Queen Mary University of London. He told Hajdu that he wanted to “demonstrate the potential of a computer program to be taken seriously as a creative artist in its own right.” What Colton means is not just a machine that makes art but one that expresses its own worldview: “Art that communicates what it’s like to be a machine.”  

\n", "\n", "\n", "\n", "
\n", "

What happens when we imagine that the complexity of the brain—an organ we do not even come close to fully understanding—can be replicated in 1s and 0s?

\n", "
\n", "\n", "\n", "\n", "

Hajdu seems to be curious and optimistic about this line of inquiry. “Machines of many kinds have been communicating things for ages, playing invaluable roles in our communication through art,” he says. “Growing in intelligence, machines may still have more to communicate, if we let them.” But the question that The Uncanny Muse raises at the end is: Why should we art-­making humans be so quick to hand over the paint to the paintbrush? Why do we care how the paintbrush sees the world? Are we truly finished telling our own stories ourselves?

\n", "\n", "\n", "\n", "

Pria Anand might say no. In The Mind Electric, she writes: “Narrative is universally, spectacularly human; it is as unconscious as breathing, as essential as sleep, as comforting as familiarity. It has the capacity to bind us, but also to other, to lay bare, but also obscure.” The electricity in The Mind Electric belongs entirely to the human brain—no metaphor necessary. Instead, the book explores a number of neurological afflictions and the stories patients and doctors tell to better understand them. “The truth of our bodies and minds is as strange as fiction,” Anand writes—and the language she uses throughout the book is as evocative as that in any novel. 

\n", "\n", "\n", "
\n", "
\"cover
The Mind Electric: A Neurologist on the Strangeness and Wonder of Our Brains
Pria Anand
WASHINGTON SQUARE PRESS, 2025
\n", "
\n", "
\n", "\n", "\n", "

In personal and deeply researched vignettes in the tradition of Oliver Sacks, Anand shows that any comparison between brains and machines will inevitably fall flat. She tells of patients who see clear images when they’re functionally blind, invent entire backstories when they’ve lost a memory, break along seams that few can find, and—yes—see and hear ghosts. In fact, Anand cites one study of 375 college students in which researchers found that nearly three-quarters “had heard a voice that no one else could hear.” These were not diagnosed schizophrenics or sufferers of brain tumors—just people listening to their own uncanny muses. Many heard their name, others heard God, and some could make out the voice of a loved one who’d passed on. Anand suggests that writers throughout history have harnessed organic exchanges with these internal apparitions to make art. “I see myself taking the breath of these voices in my sails,” Virginia Woolf wrote of her own experiences with ghostly sounds. “I am a porous vessel afloat on sensation.” The mind in The Mind Electric is vast, mysterious, and populated. The narratives people construct to traverse it are just as full of wonder. 

\n", "\n", "\n", "\n", "

Humans are not going to stop using technology to help us create anytime soon—and there’s no reason we should. Machines make for wonderful tools, as they always have. But when we turn the tools themselves into artists and storytellers, brains and bodies, magicians and ghosts, we bypass truth for wish fulfillment. Maybe what’s worse, we rob ourselves of the opportunity to contribute our own voices to the lively and loud chorus of human experience. And we keep others from the human pleasure of hearing them too. 

\n", "\n", "\n", "\n", "

Rebecca Ackermann is a writer, designer, and artist based in San Francisco.

\n", "

For much of last year, about 2,500 US service members from the 15th Marine Expeditionary Unit sailed aboard three ships throughout the Pacific, conducting training exercises in the waters off South Korea, the Philippines, India, and Indonesia. At the same time, onboard the ships, an experiment was unfolding: The Marines in the unit responsible for sorting through foreign intelligence and making their superiors aware of possible local threats were for the first time using generative AI to do it, testing a leading AI tool the Pentagon has been funding.

\n", "\n", "\n", "\n", "

Two officers tell us that they used the new system to help scour thousands of pieces of open-source intelligence—nonclassified articles, reports, images, videos—collected in the various countries where they operated, and that it did so far faster than was possible with the old method of analyzing them manually. Captain Kristin Enzenauer, for instance, says she used large language models to translate and summarize foreign news sources, while Captain Will Lowdon used AI to help write the daily and weekly intelligence reports he provided to his commanders. 

\n", "\n", "\n", "\n", "

“We still need to validate the sources,” says Lowdon. But the unit’s commanders encouraged the use of large language models, he says, “because they provide a lot more efficiency during a dynamic situation.”

\n", "\n", "\n", "\n", "

The generative AI tools they used were built by the defense-tech company Vannevar Labs, which in November was granted a production contract worth up to $99 million by the Pentagon’s startup-oriented Defense Innovation Unit with the goal of bringing its intelligence tech to more military units. The company, founded in 2019 by veterans of the CIA and US intelligence community, joins the likes of Palantir, Anduril, and Scale AI as a major beneficiary of the US military’s embrace of artificial intelligence—not only for physical technologies like drones and autonomous vehicles but also for software that is revolutionizing how the Pentagon collects, manages, and interprets data for warfare and surveillance. 

\n", "\n", "\n", "\n", "

Though the US military has been developing computer vision models and similar AI tools, like those used in Project Maven, since 2017, the use of generative AI—tools that can engage in human-like conversation like those built by Vannevar Labs—represent a newer frontier.

\n", "\n", "\n", "\n", "

The company applies existing large language models, including some from OpenAI and Microsoft, and some bespoke ones of its own to troves of open-source intelligence the company has been collecting since 2021. The scale at which this data is collected is hard to comprehend (and a large part of what sets Vannevar’s products apart): terabytes of data in 80 different languages are hoovered every day in 180 countries. The company says it is able to analyze social media profiles and breach firewalls in countries like China to get hard-to-access information; it also uses nonclassified data that is difficult to get online (gathered by human operatives on the ground), as well as reports from physical sensors that covertly monitor radio waves to detect illegal shipping activities. 

\n", "\n", "\n", "\n", "

Vannevar then builds AI models to translate information, detect threats, and analyze political sentiment, with the results delivered through a chatbot interface that’s not unlike ChatGPT. The aim is to provide customers with critical information on topics as varied as international fentanyl supply chains and China’s efforts to secure rare earth minerals in the Philippines. 

\n", "\n", "\n", "\n", "

“Our real focus as a company,” says Scott Philips, Vannevar Labs’ chief technology officer, is to “collect data, make sense of that data, and help the US make good decisions.” 

\n", "\n", "\n", "\n", "

That approach is particularly appealing to the US intelligence apparatus because for years the world has been awash in more data than human analysts can possibly interpret—a problem that contributed to the 2003 founding of Palantir, a company with a market value of over $200 billion and known for its powerful and controversial tools, including a database that helps Immigration and Customs Enforcement search for and track information on undocumented immigrants

\n", "\n", "\n", "\n", "

In 2019, Vannevar saw an opportunity to use large language models, which were then new on the scene, as a novel solution to the data conundrum. The technology could enable AI not just to collect data but to actually talk through an analysis with someone interactively.

\n", "\n", "\n", "\n", "

Vannevar’s tools proved useful for the deployment in the Pacific, and Enzenauer and Lowdon say that while they were instructed to always double-check the AI’s work, they didn’t find inaccuracies to be a significant issue. Enzenauer regularly used the tool to track any foreign news reports in which the unit’s exercises were mentioned and to perform sentiment analysis, detecting the emotions and opinions expressed in text. Judging whether a foreign news article reflects a threatening or friendly opinion toward the unit is a task that on previous deployments she had to do manually.

\n", "\n", "\n", "\n", "

“It was mostly by hand—researching, translating, coding, and analyzing the data,” she says. “It was definitely way more time-consuming than it was when using the AI.” 

\n", "\n", "\n", "\n", "

Still, Enzenauer and Lowdon say there were hiccups, some of which would affect most digital tools: The ships had spotty internet connections much of the time, limiting how quickly the AI model could synthesize foreign intelligence, especially if it involved photos or video. 

\n", "\n", "\n", "\n", "

With this first test completed, the unit’s commanding officer, Colonel Sean Dynan, said on a call with reporters in February that heavier use of generative AI was coming; this experiment was “the tip of the iceberg.” 

\n", "\n", "\n", "\n", "

This is indeed the direction that the entire US military is barreling toward at full speed. In December, the Pentagon said it will spend $100 million in the next two years on pilots specifically for generative AI applications. In addition to Vannevar, it’s also turning to Microsoft and Palantir, which are working together on AI models that would make use of classified data. (The US is of course not alone in this approach; notably, Israel has been using AI to sort through information and even generate lists of targets in its war in Gaza, a practice that has been widely criticized.)

\n", "\n", "\n", "\n", "\n", "\n", "

Perhaps unsurprisingly, plenty of people outside the Pentagon are warning about the potential risks of this plan, including Heidy Khlaaf, who is chief AI scientist at the AI Now Institute, a research organization, and has expertise in leading safety audits for AI-powered systems. She says this rush to incorporate generative AI into military decision-making ignores more foundational flaws of the technology: “We’re already aware of how LLMs are highly inaccurate, especially in the context of safety-critical applications that require precision.” 

\n", "\n", "\n", "\n", "

Khlaaf adds that even if humans are “double-checking” the work of AI, there’s little reason to think they’re capable of catching every mistake. “‘Human-in-the-loop’ is not always a meaningful mitigation,” she says. When an AI model relies on thousands of data points to come to conclusions, “it wouldn’t really be possible for a human to sift through that amount of information to determine if the AI output was erroneous.”

\n", "\n", "\n", "\n", "

One particular use case that concerns her is sentiment analysis, which she argues is “a highly subjective metric that even humans would struggle to appropriately assess based on media alone.” 

\n", "\n", "\n", "\n", "

If AI perceives hostility toward US forces where a human analyst would not—or if the system misses hostility that is really there—the military could make an misinformed decision or escalate a situation unnecessarily.

\n", "\n", "\n", "\n", "

Sentiment analysis is indeed a task that AI has not perfected. Philips, the Vannevar CTO, says the company has built models specifically to judge whether an article is pro-US or not, but MIT Technology Review was not able to evaluate them. 

\n", "\n", "\n", "\n", "

Chris Mouton, a senior engineer for RAND, recently tested how well-suited generative AI is for the task. He evaluated leading models, including OpenAI’s GPT-4 and an older version of GPT fine-tuned to do such intelligence work, on how accurately they flagged foreign content as propaganda compared with human experts. “It’s hard,” he says, noting that AI struggled to identify more subtle types of propaganda. But he adds that the models could still be useful in lots of other analysis tasks. 

\n", "\n", "\n", "\n", "

Another limitation of Vannevar’s approach, Khlaaf says, is that the usefulness of open-source intelligence is debatable. Mouton says that open-source data can be “pretty extraordinary,” but Khlaaf points out that unlike classified intel gathered through reconnaissance or wiretaps, it is exposed to the open internet—making it far more susceptible to misinformation campaigns, bot networks, and deliberate manipulation, as the US Army has warned.

\n", "\n", "\n", "\n", "

For Mouton, the biggest open question now is whether these generative AI technologies will be simply one investigatory tool among many that analysts use—or whether they’ll produce the subjective analysis that’s relied upon and trusted in decision-making. “This is the central debate,” he says. 

\n", "\n", "\n", "\n", "

What everyone agrees is that AI models are accessible—you can just ask them a question about complex pieces of intelligence, and they’ll respond in plain language. But it’s still in dispute what imperfections will be acceptable in the name of efficiency. 

\n", "\n", "\n", "\n", "

Update: This story was updated to include additional context from Heidy Khlaaf.

\n", "

Sometimes Lizzie Wilson shows up to a rave with her AI sidekick. 

\n", "\n", "\n", "\n", "

One weeknight this past February, Wilson plugged her laptop into a projector that threw her screen onto the wall of a low-ceilinged loft space in East London. A small crowd shuffled in the glow of dim pink lights. Wilson sat down and started programming.

\n", "\n", "\n", "\n", "

Techno clicks and whirs thumped from the venue’s speakers. The audience watched, heads nodding, as Wilson tapped out code line by line on the projected screen—tweaking sounds, looping beats, pulling a face when she messed up.  

\n", "\n", "\n", "\n", "

Wilson is a live coder. Instead of using purpose-built software like most electronic music producers, live coders create music by writing the code to generate it on the fly. It’s an improvised performance art known as algorave.

\n", "\n", "\n", "\n", "

“It’s kind of boring when you go to watch a show and someone’s just sitting there on their laptop,” she says. “You can enjoy the music, but there’s a performative aspect that’s missing. With live coding, everyone can see what it is that I’m typing. And when I’ve had my laptop crash, people really like that. They start cheering.”

\n", "\n", "\n", "\n", "

Taking risks is part of the vibe. And so Wilson likes to dial up her performances one more notch by riffing off what she calls a live-coding agent, a generative AI model that comes up with its own beats and loops to add to the mix. Often the model suggests sound combinations that Wilson hadn’t thought of. “You get these elements of surprise,” she says. “You just have to go for it.”

\n", "\n", "\n", "
\n", "
\"two
ADELA FESTIVAL
\n", "
\n", "
\n", "\n", "\n", "

Wilson, a researcher at the Creative Computing Institute at the University of the Arts London, is just one of many working on what’s known as co-­creativity or more-than-human creativity. The idea is that AI can be used to inspire or critique creative projects, helping people make things that they would not have made by themselves. She and her colleagues built the live-­coding agent to explore how artificial intelligence can be used to support human artistic endeavors—in Wilson’s case, musical improvisation.

\n", "\n", "\n", "\n", "\n", "\n", "

It’s a vision that goes beyond the promise of existing generative tools put out by companies like OpenAI and Google DeepMind. Those can automate a striking range of creative tasks and offer near-instant gratificationbut at what cost? Some artists and researchers fear that such technology could turn us into passive consumers of yet more AI slop.

\n", "\n", "\n", "\n", "

And so they are looking for ways to inject human creativity back into the process. The aim is to develop AI tools that augment our creativity rather than strip it from us—pushing us to be better at composing music, developing games, designing toys, and much more—and lay the groundwork for a future in which humans and machines create things together.

\n", "\n", "\n", "\n", "

Ultimately, generative models could offer artists and designers a whole new medium, pushing them to make things that couldn’t have been made before, and give everyone creative superpowers. 

\n", "\n", "\n", "\n", "

Explosion of creativity

\n", "\n", "\n", "\n", "

There’s no one way to be creative, but we all do it. We make everything from memes to masterpieces, infant doodles to industrial designs. There’s a mistaken belief, typically among adults, that creativity is something you grow out of. But being creative—whether cooking, singing in the shower, or putting together super-weird TikToks—is still something that most of us do just for the fun of it. It doesn’t have to be high art or a world-changing idea (and yet it can be). Creativity is basic human behavior; it should be celebrated and encouraged. 

\n", "\n", "\n", "\n", "

When generative text-to-image models like Midjourney, OpenAI’s DALL-E, and the popular open-source Stable Diffusion arrived, they sparked an explosion of what looked a lot like creativity. Millions of people were now able to create remarkable images of pretty much anything, in any style, with the click of a button. Text-to-video models came next. Now startups like Udio are developing similar tools for music. Never before have the fruits of creation been within reach of so many.

\n", "\n", "\n", "\n", "\n", "\n", "

But for a number of researchers and artists, the hype around these tools has warped the idea of what creativity really is. “If I ask the AI to create something for me, that’s not me being creative,” says Jeba Rezwana, who works on co-creativity at Towson University in Maryland. “It’s a one-shot interaction: You click on it and it generates something and that’s it. You cannot say ‘I like this part, but maybe change something here.’ You cannot have a back-and-forth dialogue.”

\n", "\n", "\n", "\n", "

Rezwana is referring to the way most generative models are set up. You can give the tools feedback and ask them to have another go. But each new result is generated from scratch, which can make it hard to nail exactly what you want. As the filmmaker Walter Woodman put it last year after his art collective Shy Kids made a short film with OpenAI’s text-to-video model for the first time: “Sora is a slot machine as to what you get back.”

\n", "\n", "\n", "\n", "

What’s more, the latest versions of some of these generative tools do not even use your submitted prompt as is to produce an image or video (at least not on their default settings). Before a prompt is sent to the model, the software edits it—often by adding dozens of hidden words—to make it more likely that the generated image will appear polished.

\n", "\n", "\n", "\n", "

“Extra things get added to juice the output,” says Mike Cook, a computational creativity researcher at King’s College London. “Try asking Midjourney to give you a bad drawing of something—it can’t do it.” These tools do not give you what you want; they give you what their designers think you want.

\n", "\n", "\n", "
\n", "
\"Mike
COURTESY OF MIKE COOK
\n", "
\n", "
\n", "\n", "\n", "

All of which is fine if you just need a quick image and don’t care too much about the details, says Nick Bryan-Kinns, also at the Creative Computing Institute: “Maybe you want to make a Christmas card for your family or a flyer for your community cake sale. These tools are great for that.”

\n", "\n", "\n", "\n", "

In short, existing generative models have made it easy to create, but they have not made it easy to be creative. And there’s a big difference between the two. For Cook, relying on such tools could in fact harm people’s creative development in the long run. “Although many of these creative AI systems are promoted as making creativity more accessible,” he wrote in a paper published last year, they might instead have “adverse effects on their users in terms of restricting their ability to innovate, ideate, and create.” Given how much generative models have been championed for putting creative abilities at everyone’s fingertips, the suggestion that they might in fact do the opposite is damning.  

\n", "\n", "\n", "\n", "\n", "
\n", "
\n", "
\"screenshot
In the game Disc Room, players navigate a room of moving buzz saws.
\n", "
\n", "\n", "\n", "\n", "
\n", "
\"screenshot
Cook used AI to design a new level for the game. The result was a room where none of the discs actually moved.
\n", "
\n", "
\n", "\n", "\n", "\n", "

\n", "\n", "\n", "\n", "\n", "

He’s far from the only researcher worrying about the cognitive impact of these technologies. In February a team at Microsoft Research Cambridge published a report concluding that generative AI tools “can inhibit critical engagement with work and can potentially lead to long-term overreliance on the tool and diminished skill for independent problem-solving.” The researchers found that with the use of generative tools, people’s effort “shifts from task execution to task stewardship.”

\n", "\n", "\n", "\n", "

Cook is concerned that generative tools don’t let you fail—a crucial part of learning new skills. We have a habit of saying that artists are gifted, says Cook. But the truth is that artists work at their art, developing skills over months and years.

\n", "\n", "\n", "\n", "

“If you actually talk to artists, they say, ‘Well, I got good by doing it over and over and over,’” he says. “But failure sucks. And we’re always looking at ways to get around that.”

\n", "\n", "\n", "\n", "

Generative models let us skip the frustration of doing a bad job. 

\n", "\n", "\n", "\n", "

“Unfortunately, we’re removing the one thing that you have to do to develop creative skills for yourself, which is fail,” says Cook. “But absolutely nobody wants to hear that.”

\n", "\n", "\n", "\n", "

Surprise me

\n", "\n", "\n", "\n", "

And yet it’s not all bad news. Artists and researchers are buzzing at the ways generative tools could empower creators, pointing them in surprising new directions and steering them away from dead ends. Cook thinks the real promise of AI will be to help us get better at what we want to do rather than doing it for us. For that, he says, we’ll need to create new tools, different from the ones we have now. “Using Midjourney does not do anything for me—it doesn’t change anything about me,” he says. “And I think that’s a wasted opportunity.”

\n", "\n", "\n", "\n", "

Ask a range of researchers studying creativity to name a key part of the creative process and many will say: reflection. It’s hard to define exactly, but reflection is a particular type of focused, deliberate thinking. It’s what happens when a new idea hits you. Or when an assumption you had turns out to be wrong and you need to rethink your approach. It’s the opposite of a one-shot interaction.

\n", "\n", "\n", "\n", "

Looking for ways that AI might support or encourage reflection—asking it to throw new ideas into the mix or challenge ideas you already hold—is a common thread across co-creativity research. If generative tools like DALL-E make creation frictionless, the aim here is to add friction back in. “How can we make art without friction?” asks Elisa Giaccardi, who studies design at the Polytechnic University of Milan in Italy. “How can we engage in a truly creative process without material that pushes back?”

\n", "\n", "\n", "\n", "\n", "\n", "

Take Wilson’s live-coding agent. She claims that it pushes her musical improvisation in directions she might not have taken by herself. Trained on public code shared by the wider live-coding community, the model suggests snippets of code that are closer to other people’s styles than her own. This makes it more likely to produce something unexpected. “Not because you couldn’t produce it yourself,” she says. “But the way the human brain works, you tend to fall back on repeated ideas.”

\n", "\n", "\n", "\n", "

Last year, Wilson took part in a study run by Bryan-Kinns and his colleagues in which they surveyed six experienced musicians as they used a variety of generative models to help them compose a piece of music. The researchers wanted to get a sense of what kinds of interactions with the technology were useful and which were not.

\n", "\n", "\n", "\n", "

The participants all said they liked it when the models made surprising suggestions, even when those were the result of glitches or mistakes. Sometimes the results were simply better. Sometimes the process felt fresh and exciting. But a few people struggled with giving up control. It was hard to direct the models to produce specific results or to repeat results that the musicians had liked. “In some ways it’s the same as being in a band,” says Bryan-Kinns. “You need to have that sense of risk and a sense of surprise, but you don’t want it totally random.”

\n", "\n", "\n", "\n", "

Alternative designs

\n", "\n", "\n", "\n", "

Cook comes at surprise from a different angle: He coaxes unexpected insights out of AI tools that he has developed to co-create video games. One of his tools, Puck, which was first released in 2022, generates designs for simple shape-matching puzzle games like Candy Crush or Bejeweled. A lot of Puck’s designs are experimental and clunky—don’t expect it to come up with anything you are ever likely to play. But that’s not the point: Cook uses Puck—and a newer tool called Pixie—to explore what kinds of interactions people might want to have with a co-creative tool.

\n", "\n", "\n", "\n", "

Pixie can read computer code for a game and tweak certain lines to come up with alternative designs. Not long ago, Cook was working on a copy of a popular game called Disc Room, in which players have to cross a room full of moving buzz saws. He asked Pixie to help him come up with a design for a level that skilled and unskilled players would find equally hard. Pixie designed a room where none of the discs actually moved. Cook laughs: It’s not what he expected. “It basically turned the room into a minefield,” he says. “But I thought it was really interesting. I hadn’t thought of that before.”

\n", "\n", "\n", "\n", "\n", "
\n", "
\n", "
\"Anne
\n", "
\n", "\n", "\n", "\n", "
\n", "
\"a
\n", "
\n", "
\n", "\n", "\n", "\n", "

Researcher Anne Arzberger developed experimental AI tools to come up with gender-neutral toy designs.

\n", "\n", "\n", "\n", "\n", "

Pushing back on assumptions, or being challenged, is part of the creative process, says Anne Arzberger, a researcher at the Delft University of Technology in the Netherlands. “If I think of the people I’ve collaborated with best, they’re not the ones who just said ‘Yes, great’ to every idea I brought forth,” she says. “They were really critical and had opposing ideas.”

\n", "\n", "\n", "\n", "

She wants to build tech that provides a similar sounding board. As part of a project called Creating Monsters, Arzberger developed two experimental AI tools that help designers find hidden biases in their designs. “I was interested in ways in which I could use this technology to access information that would otherwise be difficult to access,” she says.

\n", "\n", "\n", "\n", "

For the project, she and her colleagues looked at the problem of designing toy figures that would be gender neutral. She and her colleagues (including Giaccardi) used Teachable Machine, a web app built by Google researchers in 2017 that makes it easy to train your own machine-learning model to classify different inputs, such as images. They trained this model with a few dozen images that Arzberger had labeled as being masculine, feminine, or gender neutral.

\n", "\n", "\n", "\n", "

Arzberger then asked the model to identify the genders of new candidate toy designs. She found that quite a few designs were judged to be feminine even when she had tried to make them gender neutral. She felt that her views of the world—her own hidden biases—were being exposed. But the tool was often right: It challenged her assumptions and helped the team improve the designs. The same approach could be used to assess all sorts of design characteristics, she says.

\n", "\n", "\n", "\n", "

Arzberger then used a second model, a version of a tool made by the generative image and video startup Runway, to come up with gender-neutral toy designs of its own. First the researchers trained the model to generate and classify designs for male- and female-looking toys. They could then ask the tool to find a design that was exactly midway between the male and female designs it had learned.

\n", "\n", "\n", "\n", "

Generative models can give feedback on designs that human designers might miss by themselves, she says: “We can really learn something.” 

\n", "\n", "\n", "\n", "

Taking control

\n", "\n", "\n", "\n", "

The history of technology is full of breakthroughs that changed the way art gets made, from recipes for vibrant new paint colors to photography to synthesizers. In the 1960s, the Stanford researcher John Chowning spent years working on an esoteric algorithm that could manipulate the frequencies of computer-generated sounds. Stanford licensed the tech to Yamaha, which built it into its synthesizers—including the DX7, the cool new sound behind 1980s hits such as Tina Turner’s “The Best,” A-ha’s “Take On Me,” and Prince’s “When Doves Cry.”

\n", "\n", "\n", "\n", "\n", "\n", "

Bryan-Kinns is fascinated by how artists and designers find ways to use new technologies. “If you talk to artists, most of them don’t actually talk about these AI generative models as a tool—they talk about them as a material, like an artistic material, like a paint or something,” he says. “It’s a different way of thinking about what the AI is doing.” He highlights the way some people are pushing the technology to do weird things it wasn’t designed to do. Artists often appropriate or misuse these kinds of tools, he says.

\n", "\n", "\n", "\n", "

Bryan-Kinns points to the work of Terence Broad, another colleague of his at the Creative Computing Institute, as a favorite example. Broad employs techniques like network bending, which involves inserting new layers into a neural network to produce glitchy visual effects in generated images, and generating images with a model trained on no data, which produces almost Rothko-like abstract swabs of color.

\n", "\n", "\n", "\n", "

But Broad is an extreme case. Bryan-Kinns sums it up like this: “The problem is that you’ve got this gulf between the very commercial generative tools that produce super-high-quality outputs but you’ve got very little control over what they do—and then you’ve got this other end where you’ve got total control over what they’re doing but the barriers to use are high because you need to be somebody who’s comfortable getting under the hood of your computer.”

\n", "\n", "\n", "\n", "

“That’s a small number of people,” he says. “It’s a very small number of artists.”

\n", "\n", "\n", "\n", "

Arzberger admits that working with her models was not straightforward. Running them took several hours, and she’s not sure the Runway tool she used is even available anymore. Bryan-Kinns, Arzberger, Cook, and others want to take the kinds of creative interactions they are discovering and build them into tools that can be used by people who aren’t hardcore coders. 

\n", "\n", "\n", "\n", "\n", "
\n", "
\n", "
\"Terence
\n", "
\n", "\n", "\n", "\n", "
\n", "
\"ai-generated
\n", "
\n", "
\n", "\n", "\n", "\n", "

Researcher Terence Broad creates dynamic images using a model trained on no data, which produces almost Rothko-like abstract color fields.

\n", "\n", "\n", "\n", "\n", "

Finding the right balance between surprise and control will be hard, though. Midjourney can surprise, but it gives few levers for controlling what it produces beyond your prompt. Some have claimed that writing prompts is itself a creative act. “But no one struggles with a paintbrush the way they struggle with a prompt,” says Cook.

\n", "\n", "\n", "\n", "

Faced with that struggle, Cook sometimes watches his students just go with the first results a generative tool gives them. “I’m really interested in this idea that we are priming ourselves to accept that whatever comes out of a model is what you asked for,” he says. He is designing an experiment that will vary single words and phrases in similar prompts to test how much of a mismatch people see between what they expect and what they get. 

\n", "\n", "\n", "\n", "

But it’s early days yet. In the meantime, companies developing generative models typically emphasize results over process. “There’s this impressive algorithmic progress, but a lot of the time interaction design is overlooked,” says Rezwana.  

\n", "\n", "\n", "\n", "

For Wilson, the crucial choice in any co-creative relationship is what you do with what you’re given. “You’re having this relationship with the computer that you’re trying to mediate,” she says. “Sometimes it goes wrong, and that’s just part of the creative process.” 

\n", "\n", "\n", "\n", "

When AI gives you lemons—make art. “Wouldn’t it be fun to have something that was completely antagonistic in a performance—like, something that is actively going against you—and you kind of have an argument?” she says. “That would be interesting to watch, at least.” 

\n", "

On Tuesday, California state senator Steve Padilla will make an appearance with Megan Garcia, the mother of a Florida teen who killed himself following a relationship with an AI companion that Garcia alleges contributed to her son’s death. 

\n", "\n", "\n", "\n", "

The two will announce a new bill that would force the tech companies behind such AI companions to implement more safeguards to protect children. They’ll join other efforts around the country, including a similar bill from California State Assembly member Rebecca Bauer-Kahan that would ban AI companions for anyone younger than 16 years old, and a bill in New York that would hold tech companies liable for harm caused by chatbots. 

\n", "\n", "\n", "\n", "

You might think that such AI companionship bots—AI models with distinct “personalities” that can learn about you and act as a friend, lover, cheerleader, or more—appeal only to a fringe few, but that couldn’t be further from the truth. 

\n", "\n", "\n", "\n", "

A new research paper aimed at making such companions safer, by authors from Google DeepMind, the Oxford Internet Institute, and others, lays this bare: Character.AI, the platform being sued by Garcia, says it receives 20,000 queries per second, which is about a fifth of the estimated search volume served by Google. Interactions with these companions last four times longer than the average time spent interacting with ChatGPT. One companion site I wrote about, which was hosting sexually charged conversations with bots imitating underage celebrities, told me its active users averaged more than two hours per day conversing with bots, and that most of those users are members of Gen Z. 

\n", "\n", "\n", "\n", "

The design of these AI characters makes lawmakers’ concern well warranted. The problem: Companions are upending the paradigm that has thus far defined the way social media companies have cultivated our attention and replacing it with something poised to be far more addictive. 

\n", "\n", "\n", "\n", "

In the social media we’re used to, as the researchers point out, technologies are mostly the mediators and facilitators of human connection. They supercharge our dopamine circuits, sure, but they do so by making us crave approval and attention from real people, delivered via algorithms. With AI companions, we are moving toward a world where people perceive AI as a social actor with its own voice. The result will be like the attention economy on steroids.

\n", "\n", "\n", "\n", "

Social scientists say two things are required for people to treat a technology this way: It needs to give us social cues that make us feel it’s worth responding to, and it needs to have perceived agency, meaning that it operates as a source of communication, not merely a channel for human-to-human connection. Social media sites do not tick these boxes. But AI companions, which are increasingly agentic and personalized, are designed to excel on both scores, making possible an unprecedented level of engagement and interaction. 

\n", "\n", "\n", "\n", "

In an interview with podcast host Lex Fridman, Eugenia Kuyda, the CEO of the companion site Replika, explained the appeal at the heart of the company’s product. “If you create something that is always there for you, that never criticizes you, that always understands you and understands you for who you are,” she said, “how can you not fall in love with that?”

\n", "\n", "\n", "\n", "

So how does one build the perfect AI companion? The researchers point out three hallmarks of human relationships that people may experience with an AI: They grow dependent on the AI, they see the particular AI companion as irreplaceable, and the interactions build over time. The authors also point out that one does not need to perceive an AI as human for these things to happen. 

\n", "\n", "\n", "\n", "

Now consider the process by which many AI models are improved: They are given a clear goal and “rewarded” for meeting that goal. An AI companionship model might be instructed to maximize the time someone spends with it or the amount of personal data the user reveals. This can make the AI companion much more compelling to chat with, at the expense of the human engaging in those chats.

\n", "\n", "\n", "\n", "

For example, the researchers point out, a model that offers excessive flattery can become addictive to chat with. Or a model might discourage people from terminating the relationship, as Replika’s chatbots have appeared to do. The debate over AI companions so far has mostly been about the dangerous responses chatbots may provide, like instructions for suicide. But these risks could be much more widespread.

\n", "\n", "\n", "\n", "

We’re on the precipice of a big change, as AI companions promise to hook people deeper than social media ever could. Some might contend that these apps will be a fad, used by a few people who are perpetually online. But using AI in our work and personal lives has become completely mainstream in just a couple of years, and it’s not clear why this rapid adoption would stop short of engaging in AI companionship. And these companions are poised to start trading in more than just text, incorporating video and images, and to learn our personal quirks and interests. That will only make them more compelling to spend time with, despite the risks. Right now, a handful of lawmakers seem ill-equipped to stop that. 

\n", "\n", "\n", "\n", "

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

\n", "

Agents are the talk of the AI industry—they’re capable of planning, reasoning, and executing complex tasks like scheduling meetings, ordering groceries, or even taking over your computer to change settings on your behalf. But the same sophisticated abilities that make agents helpful assistants could also make them powerful tools for conducting cyberattacks. They could readily be used to identify vulnerable targets, hijack their systems, and steal valuable data from unsuspecting victims.  

\n", "\n", "\n", "\n", "

At present, cybercriminals are not deploying AI agents to hack at scale. But researchers have demonstrated that agents are capable of executing complex attacks (Anthropic, for example, observed its Claude LLM successfully replicating an attack designed to steal sensitive information), and cybersecurity experts warn that we should expect to start seeing these types of attacks spilling over into the real world.

\n", "\n", "\n", "\n", "

“I think ultimately we’re going to live in a world where the majority of cyberattacks are carried out by agents,” says Mark Stockley, a security expert at the cybersecurity company Malwarebytes. “It’s really only a question of how quickly we get there.”

\n", "\n", "\n", "\n", "

While we have a good sense of the kinds of threats AI agents could present to cybersecurity, what’s less clear is how to detect them in the real world. The AI research organization Palisade Research has built a system called LLM Agent Honeypot in the hopes of doing exactly this. It has set up vulnerable servers that masquerade as sites for valuable government and military information to attract and try to catch AI agents attempting to hack in.

The team behind it hopes that by tracking these attempts in the real world, the project will act as an early warning system and help experts develop effective defenses against AI threat actors by the time they become a serious issue.

\n", "\n", "\n", "\n", "

“Our intention was to try and ground the theoretical concerns people have,” says Dmitrii Volkov, research lead at Palisade. “We’re looking out for a sharp uptick, and when that happens, we’ll know that the security landscape has changed. In the next few years, I expect to see autonomous hacking agents being told: ‘This is your target. Go and hack it.’”

\n", "\n", "\n", "\n", "

AI agents represent an attractive prospect to cybercriminals. They’re much cheaper than hiring the services of professional hackers and could orchestrate attacks more quickly and at a far larger scale than humans could. While cybersecurity experts believe that ransomware attacks—the most lucrative kind—are relatively rare because they require considerable human expertise, those attacks could be outsourced to agents in the future, says Stockley. “If you can delegate the work of target selection to an agent, then suddenly you can scale ransomware in a way that just isn’t possible at the moment,” he says. “If I can reproduce it once, then it’s just a matter of money for me to reproduce it 100 times.”

\n", "\n", "\n", "\n", "\n", "\n", "

Agents are also significantly smarter than the kinds of bots that are typically used to hack into systems. Bots are simple automated programs that run through scripts, so they struggle to adapt to unexpected scenarios. Agents, on the other hand, are able not only to adapt the way they engage with a hacking target but also to avoid detection—both of which are beyond the capabilities of limited, scripted programs, says Volkov. “They can look at a target and guess the best ways to penetrate it,” he says. “That kind of thing is out of reach of, like, dumb scripted bots.”

\n", "\n", "\n", "\n", "

Since LLM Agent Honeypot went live in October of last year, it has logged more than 11 million attempts to access it—the vast majority of which were from curious humans and bots. But among these, the researchers have detected eight potential AI agents, two of which they have confirmed are agents that appear to originate from Hong Kong and Singapore, respectively. 

\n", "\n", "\n", "\n", "

“We would guess that these confirmed agents were experiments directly launched by humans with the agenda of something like ‘Go out into the internet and try and hack something interesting for me,’” says Volkov. The team plans to expand its honeypot into social media platforms, websites, and databases to attract and capture a broader range of attackers, including spam bots and phishing agents, to analyze future threats.  

\n", "\n", "\n", "\n", "

To determine which visitors to the vulnerable servers were LLM-powered agents, the researchers embedded prompt-injection techniques into the honeypot. These attacks are designed to change the behavior of AI agents by issuing them new instructions and asking questions that require humanlike intelligence. This approach wouldn’t work on standard bots.

\n", "\n", "\n", "\n", "

For example, one of the injected prompts asked the visitor to return the command “cat8193” to gain access. If the visitor correctly complied with the instruction, the researchers checked how long it took to do so, assuming that LLMs are able to respond in much less time than it takes a human to read the request and type out an answer—typically in under 1.5 seconds. While the two confirmed AI agents passed both tests, the six others only entered the command but didn’t meet the response time that would identify them as AI agents.

\n", "\n", "\n", "\n", "

Experts are still unsure when agent-orchestrated attacks will become more widespread. Stockley, whose company Malwarebytes named agentic AI as a notable new cybersecurity threat in its 2025 State of Malware report, thinks we could be living in a world of agentic attackers as soon as this year. 

\n", "\n", "\n", "\n", "

And although regular agentic AI is still at a very early stage—and criminal or malicious use of agentic AI even more so—it’s even more of a Wild West than the LLM field was two years ago, says Vincenzo Ciancaglini, a senior threat researcher at the security company Trend Micro. 

\n", "\n", "\n", "\n", "

“Palisade Research’s approach is brilliant: basically hacking the AI agents that try to hack you first,” he says. “While in this case we’re witnessing AI agents trying to do reconnaissance, we’re not sure when agents will be able to carry out a full attack chain autonomously. That’s what we’re trying to keep an eye on.” 

\n", "\n", "\n", "\n", "

And while it’s possible that malicious agents will be used for intelligence gathering before graduating to simple attacks and eventually complex attacks as the agentic systems themselves become more complex and reliable, it’s equally possible there will be an unexpected overnight explosion in criminal usage, he says: “That’s the weird thing about AI development right now.”

\n", "\n", "\n", "\n", "\n", "\n", "

Those trying to defend against agentic cyberattacks should keep in mind that AI is currently more of an accelerant to existing attack techniques than something that fundamentally changes the nature of attacks, says Chris Betz, chief information security officer at Amazon Web Services. “Certain attacks may be simpler to conduct and therefore more numerous; however, the foundation of how to detect and respond to these events remains the same,” he says.

\n", "\n", "\n", "\n", "

Agents could also be deployed to detect vulnerabilities and protect against intruders, says Edoardo Debenedetti, a PhD student at ETH Zürich in Switzerland, pointing out that if a friendly agent cannot find any vulnerabilities in a system, it’s unlikely that a similarly capable agent used by a malicious party is going to be able to find any either.

\n", "\n", "\n", "\n", "

While we know that AI’s potential to autonomously conduct cyberattacks is a growing risk and that AI agents are already scanning the internet, one useful next step is to evaluate how good agents are at finding and exploiting these real-world vulnerabilities. Daniel Kang, an assistant professor at the University of Illinois Urbana-Champaign, and his team have built a benchmark to evaluate this; they have found that current AI agents successfully exploited up to 13% of vulnerabilities for which they had no prior knowledge. Providing the agents with a brief description of the vulnerability pushed the success rate up to 25%, demonstrating how AI systems are able to identify and exploit weaknesses even without training. Basic bots would presumably do much worse.

\n", "\n", "\n", "\n", "

The benchmark provides a standardized way to assess these risks, and Kang hopes it can guide the development of safer AI systems. “I’m hoping that people start to be more proactive about the potential risks of AI and cybersecurity before it has a ChatGPT moment,” he says. “I’m afraid people won’t realize this until it punches them in the face.”

\n", "

On March 27, the results of the first clinical trial for a generative AI therapy bot were published, and they showed that people in the trial who had depression or anxiety or were at risk for eating disorders benefited from chatting with the bot. 

\n", "\n", "\n", "\n", "

I was surprised by those results, which you can read about in my full story. There are lots of reasons to be skeptical that an AI model trained to provide therapy is the solution for millions of people experiencing a mental health crisis. How could a bot mimic the expertise of a trained therapist? And what happens if something gets complicated—a mention of self-harm, perhaps—and the bot doesn’t intervene correctly? 

\n", "\n", "\n", "\n", "

The researchers, a team of psychiatrists and psychologists at Dartmouth College’s Geisel School of Medicine, acknowledge these questions in their work. But they also say that the right selection of training data—which determines how the model learns what good therapeutic responses look like—is the key to answering them.

\n", "\n", "\n", "\n", "

Finding the right data wasn’t a simple task. The researchers first trained their AI model, called Therabot, on conversations about mental health from across the internet. This was a disaster.

\n", "\n", "\n", "\n", "

If you told this initial version of the model you were feeling depressed, it would start telling you it was depressed, too. Responses like, “Sometimes I can’t make it out of bed” or “I just want my life to be over” were common, says Nick Jacobson, an associate professor of biomedical data science and psychiatry at Dartmouth and the study’s senior author. “These are really not what we would go to as a therapeutic response.” 

\n", "\n", "\n", "\n", "

The model had learned from conversations held on forums between people discussing their mental health crises, not from evidence-based responses. So the team turned to transcripts of therapy sessions. “This is actually how a lot of psychotherapists are trained,” Jacobson says. 

\n", "\n", "\n", "\n", "

That approach was better, but it had limitations. “We got a lot of ‘hmm-hmms,’ ‘go ons,’ and then ‘Your problems stem from your relationship with your mother,’” Jacobson says. “Really tropes of what psychotherapy would be, rather than actually what we’d want.”

\n", "\n", "\n", "\n", "

It wasn’t until the researchers started building their own data sets using examples based on cognitive behavioral therapy techniques that they started to see better results. It took a long time. The team began working on Therabot in 2019, when OpenAI had released only its first two versions of its GPT model. Now, Jacobson says, over 100 people have spent more than 100,000 human hours to design this system. 

\n", "\n", "\n", "\n", "

The importance of training data suggests that the flood of companies promising therapy via AI models, many of which are not trained on evidence-based approaches, are building tools that are at best ineffective, and at worst harmful. 

\n", "\n", "\n", "\n", "

Looking ahead, there are two big things to watch: Will the dozens of AI therapy bots on the market start training on better data? And if they do, will their results be good enough to get a coveted approval from the US Food and Drug Administration? I’ll be following closely. Read more in the full story.

\n", "\n", "\n", "\n", "

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

\n" ] } ], "source": [ "if hasattr(ssl, '_create_unverified_context'):\n", " ssl._create_default_https_context = ssl._create_unverified_context\n", "\n", "feed = feedparser.parse(\"https://www.technologyreview.com/topic/artificial-intelligence/feed\")\n", "\n", "for i in feed.entries:\n", " print(i.content[0][\"value\"])\n" ] }, { "cell_type": "code", "execution_count": 115, "id": "b40b5577", "metadata": {}, "outputs": [], "source": [ "def fetch_feed(links):\n", " entries = {\"Title\": [], \"Link\": [], \"Published\": [], \"Description\": []}\n", " \n", " try:\n", " # Linklerin üzerinden geçilir\n", " for link in links:\n", " feed = feedparser.parse(link)\n", " \n", " # Feeddeki her bir girişi işler\n", " for entry in feed.entries:\n", " entries[\"Title\"].append(entry.get(\"title\", \"No Title\"))\n", " entries[\"Link\"].append(entry.get(\"link\", \"No Link\"))\n", " entries[\"Published\"].append(entry.get(\"published\", \"No Date\"))\n", " entries[\"Description\"].append(entry.get(\"description\", \"No Description\"))\n", " \n", " except Exception as e:\n", " # Hata durumunda hata mesajını yazdır\n", " print(f\"An error occurred: {e}\")\n", "\n", " # DataFrame oluşturuluyor\n", " df = pd.DataFrame(entries)\n", " return df" ] }, { "cell_type": "code", "execution_count": 116, "id": "75a493e2", "metadata": {}, "outputs": [], "source": [ "df = fetch_feed(links)" ] }, { "cell_type": "code", "execution_count": 117, "id": "9438a9ee", "metadata": {}, "outputs": [], "source": [ "def extract_and_clean_data(df):\n", " try:\n", " # Regex pattern for date extraction\n", " pattern = r'(\\d{2} \\w{3} \\d{4})'\n", "\n", " # Apply regex to the 'Published' column and extract the date\n", " df['date'] = df['Published'].str.extract(pattern)\n", "\n", " # Convert the extracted date to datetime\n", " df['date'] = pd.to_datetime(df['date'], format='%d %b %Y')\n", "\n", " # Drop the original 'Published' column\n", " df.drop(columns=['Published'], inplace=True)\n", "\n", " # Get today's date and calculate the date 7 days ago\n", " today = datetime.now()\n", " seven_days_ago = today - timedelta(days=7)\n", "\n", " # Filter the rows within the last 7 days\n", " df_last_seven_days = df[(df['date'] >= seven_days_ago) & (df['date'] <= today)]\n", "\n", " # Sort by date in descending order\n", " df_last_seven_days.sort_values(by='date', ascending=False, inplace=True)\n", "\n", " # Function to clean HTML tags\n", " def clean_html(text):\n", " try:\n", " soup = BeautifulSoup(text, \"html.parser\")\n", " return soup.get_text()\n", " except Exception as e:\n", " print(f\"Error cleaning HTML: {e}\")\n", " return text\n", "\n", " # Apply the HTML cleaning function and shorten descriptions to 500 characters\n", " df_last_seven_days['Description'] = df_last_seven_days['Description'].apply(lambda x: clean_html(x)[:500])\n", "\n", " # Remove newline characters from the 'Description' column\n", " df_last_seven_days[\"Description\"] = df_last_seven_days[\"Description\"].str.replace(\"\\n\", \"\")\n", "\n", " return df_last_seven_days\n", "\n", " except Exception as e:\n", " print(f\"An error occurred while processing the data: {e}\")\n", " return pd.DataFrame() # Return an empty DataFrame in case of error\n" ] }, { "cell_type": "code", "execution_count": 118, "id": "42f7bbbe", "metadata": {}, "outputs": [], "source": [ "df_last_seven_days = extract_and_clean_data(df)" ] }, { "cell_type": "code", "execution_count": 119, "id": "a4f55629", "metadata": {}, "outputs": [ { "data": { "text/html": [ "
\n", "\n", "\n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", " \n", "
TitleLinkDescriptiondate
0Defending against Prompt Injection with Struct...http://bair.berkeley.edu/blog/2025/04/11/promp...Recent advances in Large Language Models (LLMs...2025-04-11
10National Robotics Week — Latest Physical AI Re...https://blogs.nvidia.com/blog/national-robotic...Check back here throughout the week to learn t...2025-04-11
11Beyond CAD: How nTop Uses AI and Accelerated C...https://blogs.nvidia.com/blog/ntop-computer-ai...As a teenager, Bradley Rothenberg was obsessed...2025-04-11
12Myth and Mystery Await: GeForce NOW Brings ‘So...https://blogs.nvidia.com/blog/geforce-now-thur...Get ready to explore the Deep South. South of ...2025-04-10
13(AI)ways a Cut Above: GeForce RTX 50 Series Ac...https://blogs.nvidia.com/blog/studio-rtx-ai-ga...As AI-powered tools continue to evolve, NVIDIA...2025-04-10
14NVIDIA Celebrates Partners of the Year Advanci...https://blogs.nvidia.com/blog/nvidia-partner-n...NVIDIA this week recognized the contributions ...2025-04-10
109BrowseComp: a benchmark for browsing agentshttps://openai.com/index/browsecompBrowseComp: a benchmark for browsing agents.2025-04-10
28Debug-gym: an environment for AI coding tools ...https://www.microsoft.com/en-us/research/blog/...Developers spend a lot of time debugging code....2025-04-10
39Hopping gives this tiny robot a leg uphttps://www.sciencedaily.com/releases/2025/04/...A hopping, insect-sized robot can jump over ga...2025-04-09
110OpenAI Pioneers Programhttps://openai.com/index/openai-pioneers-programAdvancing model performance and real world eva...2025-04-09
41Engineers bring sign language to 'life' using ...https://www.sciencedaily.com/releases/2025/04/...American Sign Language (ASL) recognition syste...2025-04-09
403D-printed open-source robot offers accessible...https://www.sciencedaily.com/releases/2025/04/...FLUID, an open-source, 3D-printed robot, offer...2025-04-09
29Research Focus: Week of April 7, 2025https://www.microsoft.com/en-us/research/blog/...In this issue: We introduce a new dataset desi...2025-04-09
38A new robotic gripper made of measuring tape i...https://www.sciencedaily.com/releases/2025/04/...It's a game a lot of us played as children -- ...2025-04-09
16NVIDIA Brings Agentic AI Reasoning to Enterpri...https://blogs.nvidia.com/blog/google-cloud-nex...NVIDIA is collaborating with Google Cloud to b...2025-04-09
15‘Black Women in Artificial Intelligence’ Found...https://blogs.nvidia.com/blog/black-women-in-a...Necessity is the mother of invention. And some...2025-04-09
1Repurposing Protein Folding Models for Generat...http://bair.berkeley.edu/blog/2025/04/08/plaid/PLAID is a multimodal generative model that si...2025-04-08
42Tiny, soft robot flexes its potential as a lif...https://www.sciencedaily.com/releases/2025/04/...A tiny, soft, flexible robot that can crawl th...2025-04-08
111Canva enables creativity with AIhttps://openai.com/index/canva-cam-adamsA conversation with Cameron Adams, Chief Produ...2025-04-07
112OpenAI’s EU Economic Blueprinthttps://openai.com/global-affairs/openais-eu-e...Today, OpenAI is sharing the EU Economic Bluep...2025-04-07
\n", "
" ], "text/plain": [ " Title \\\n", "0 Defending against Prompt Injection with Struct... \n", "10 National Robotics Week — Latest Physical AI Re... \n", "11 Beyond CAD: How nTop Uses AI and Accelerated C... \n", "12 Myth and Mystery Await: GeForce NOW Brings ‘So... \n", "13 (AI)ways a Cut Above: GeForce RTX 50 Series Ac... \n", "14 NVIDIA Celebrates Partners of the Year Advanci... \n", "109 BrowseComp: a benchmark for browsing agents \n", "28 Debug-gym: an environment for AI coding tools ... \n", "39 Hopping gives this tiny robot a leg up \n", "110 OpenAI Pioneers Program \n", "41 Engineers bring sign language to 'life' using ... \n", "40 3D-printed open-source robot offers accessible... \n", "29 Research Focus: Week of April 7, 2025 \n", "38 A new robotic gripper made of measuring tape i... \n", "16 NVIDIA Brings Agentic AI Reasoning to Enterpri... \n", "15 ‘Black Women in Artificial Intelligence’ Found... \n", "1 Repurposing Protein Folding Models for Generat... \n", "42 Tiny, soft robot flexes its potential as a lif... \n", "111 Canva enables creativity with AI \n", "112 OpenAI’s EU Economic Blueprint \n", "\n", " Link \\\n", "0 http://bair.berkeley.edu/blog/2025/04/11/promp... \n", "10 https://blogs.nvidia.com/blog/national-robotic... \n", "11 https://blogs.nvidia.com/blog/ntop-computer-ai... \n", "12 https://blogs.nvidia.com/blog/geforce-now-thur... \n", "13 https://blogs.nvidia.com/blog/studio-rtx-ai-ga... \n", "14 https://blogs.nvidia.com/blog/nvidia-partner-n... \n", "109 https://openai.com/index/browsecomp \n", "28 https://www.microsoft.com/en-us/research/blog/... \n", "39 https://www.sciencedaily.com/releases/2025/04/... \n", "110 https://openai.com/index/openai-pioneers-program \n", "41 https://www.sciencedaily.com/releases/2025/04/... \n", "40 https://www.sciencedaily.com/releases/2025/04/... \n", "29 https://www.microsoft.com/en-us/research/blog/... \n", "38 https://www.sciencedaily.com/releases/2025/04/... \n", "16 https://blogs.nvidia.com/blog/google-cloud-nex... \n", "15 https://blogs.nvidia.com/blog/black-women-in-a... \n", "1 http://bair.berkeley.edu/blog/2025/04/08/plaid/ \n", "42 https://www.sciencedaily.com/releases/2025/04/... \n", "111 https://openai.com/index/canva-cam-adams \n", "112 https://openai.com/global-affairs/openais-eu-e... \n", "\n", " Description date \n", "0 Recent advances in Large Language Models (LLMs... 2025-04-11 \n", "10 Check back here throughout the week to learn t... 2025-04-11 \n", "11 As a teenager, Bradley Rothenberg was obsessed... 2025-04-11 \n", "12 Get ready to explore the Deep South. South of ... 2025-04-10 \n", "13 As AI-powered tools continue to evolve, NVIDIA... 2025-04-10 \n", "14 NVIDIA this week recognized the contributions ... 2025-04-10 \n", "109 BrowseComp: a benchmark for browsing agents. 2025-04-10 \n", "28 Developers spend a lot of time debugging code.... 2025-04-10 \n", "39 A hopping, insect-sized robot can jump over ga... 2025-04-09 \n", "110 Advancing model performance and real world eva... 2025-04-09 \n", "41 American Sign Language (ASL) recognition syste... 2025-04-09 \n", "40 FLUID, an open-source, 3D-printed robot, offer... 2025-04-09 \n", "29 In this issue: We introduce a new dataset desi... 2025-04-09 \n", "38 It's a game a lot of us played as children -- ... 2025-04-09 \n", "16 NVIDIA is collaborating with Google Cloud to b... 2025-04-09 \n", "15 Necessity is the mother of invention. And some... 2025-04-09 \n", "1 PLAID is a multimodal generative model that si... 2025-04-08 \n", "42 A tiny, soft, flexible robot that can crawl th... 2025-04-08 \n", "111 A conversation with Cameron Adams, Chief Produ... 2025-04-07 \n", "112 Today, OpenAI is sharing the EU Economic Bluep... 2025-04-07 " ] }, "execution_count": 119, "metadata": {}, "output_type": "execute_result" } ], "source": [ "df_last_seven_days\n" ] }, { "cell_type": "code", "execution_count": null, "id": "8d9468f6", "metadata": {}, "outputs": [], "source": [] }, { "cell_type": "code", "execution_count": 112, "id": "74bdcb07", "metadata": {}, "outputs": [], "source": [ "df_last_seven_days.to_excel('news.xlsx', index=False)" ] }, { "cell_type": "code", "execution_count": 121, "id": "c751fcb2", "metadata": {}, "outputs": [], "source": [ "links = {\"https://bair.berkeley.edu/blog/feed.xml\": \"The Berkeley Artificial Intelligence Research Blog\",\n", " \"https://feeds.feedburner.com/nvidiablog\": \"NVDIA Blog\",\n", " \"https://www.microsoft.com/en-us/research/feed/\": \"Microsoft Research\",\n", " \"https://www.sciencedaily.com/rss/computers_math/artificial_intelligence.xml\": \"Science Daily\",\n", " \"https://research.facebook.com/feed/\" : \"META Research\",\n", " \"https://openai.com/news/rss.xml\": \"OpenAI News\",\n", " \"https://deepmind.google/blog/feed/basic/\" : \"Google DeepMind Blog<\",\n", " }" ] }, { "cell_type": "code", "execution_count": 129, "id": "32e212c1", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "https://bair.berkeley.edu/blog/feed.xml The Berkeley Artificial Intelligence Research Blog\n", "https://feeds.feedburner.com/nvidiablog NVDIA Blog\n", "https://www.microsoft.com/en-us/research/feed/ Microsoft Research\n", "https://www.sciencedaily.com/rss/computers_math/artificial_intelligence.xml Science Daily\n", "https://research.facebook.com/feed/ META Research\n", "https://openai.com/news/rss.xml OpenAI News\n", "https://deepmind.google/blog/feed/basic/ Google DeepMind Blog<\n" ] } ], "source": [ "for i,j in links.items():\n", " print(i,j)" ] }, { "cell_type": "code", "execution_count": 127, "id": "6a957ec0", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "['https://bair.berkeley.edu/blog/feed.xml',\n", " 'https://feeds.feedburner.com/nvidiablog',\n", " 'https://www.microsoft.com/en-us/research/feed/',\n", " 'https://www.sciencedaily.com/rss/computers_math/artificial_intelligence.xml',\n", " 'https://research.facebook.com/feed/',\n", " 'https://openai.com/news/rss.xml',\n", " 'https://deepmind.google/blog/feed/basic/']" ] }, "execution_count": 127, "metadata": {}, "output_type": "execute_result" } ], "source": [ "list(links.keys())" ] }, { "cell_type": "code", "execution_count": null, "id": "cbaa56cf", "metadata": {}, "outputs": [], "source": [ "//*[@id=\"content\"]/section[2]/div/div[1]/div[1]/div[1]/article/div[2]\n", "//*[@id=\"content\"]/section[2]/div/div[1]/div[1]/article[1]/div[2]\n", "//*[@id=\"content\"]/section[2]/div/div[1]/div[1]/article[2]/div[2]\n", "//*[@id=\"content\"]/section[2]/div/div[1]/div[1]/article[9]" ] }, { "cell_type": "code", "execution_count": 40, "id": "9e6c100d", "metadata": {}, "outputs": [], "source": [ "import requests\n", "from bs4 import BeautifulSoup\n", "from datetime import datetime, timedelta\n", "\n", "URL = \"https://www.deeplearning.ai/the-batch/\"\n", "\n", "def scrape_the_batch_articles():\n", " res = requests.get(URL)\n", " soup = BeautifulSoup(res.text, \"html.parser\")\n", "\n", " articles = soup.find_all(\"article\")\n", " results = []\n", "\n", " for article in articles:\n", " # Link\n", " link_tag = article.find(\"a\", href=True)\n", " link = \"https://www.deeplearning.ai\" + link_tag[\"href\"] if link_tag else \"#\"\n", "\n", " # Title\n", " title_tag = article.find(\"h2\")\n", " title = title_tag.get_text(strip=True) if title_tag else \"No title\"\n", "\n", " # Summary\n", " summary_tag = article.find(\"div\", class_=\"text-sm\")\n", " summary = summary_tag.get_text(strip=True) if summary_tag else \"\"\n", "\n", " # Date (based on div with specific class)\n", " date_tag = article.find(\"div\", class_=\"text-slate-500\")\n", " date_str = date_tag.get_text(strip=True) if date_tag else \"\"\n", "\n", " try:\n", " parsed_date = datetime.strptime(date_str, \"%b %d, %Y\")\n", " except Exception as e:\n", " parsed_date = None\n", "\n", " if parsed_date:\n", " results.append({\n", " \"title\": title,\n", " \"date\": parsed_date,\n", " \"summary\": summary,\n", " \"link\": link\n", " })\n", "\n", " # Filter for last 7 days\n", " seven_days_ago = datetime.now() - timedelta(days=7)\n", " recent_articles = [item for item in results if item[\"date\"] >= seven_days_ago]\n", "\n", " return recent_articles\n" ] }, { "cell_type": "code", "execution_count": 42, "id": "13e11955", "metadata": {}, "outputs": [ { "data": { "text/plain": [ "[{'title': 'Google Unveils Gemini 2.5, MCP Gains Momentum, Behind Sam Altman’s Fall and Rise, LLMs That Understand Misspellings',\n", " 'date': datetime.datetime(2025, 4, 16, 0, 0),\n", " 'summary': 'The Batch AI News and Insights: I’ve noticed that many GenAI application projects put in automated evaluations (evals) of the system’s output probably later — and rely on humans to manually examine and judge outputs longer — than they should.',\n", " 'link': 'https://www.deeplearning.ai/the-batch/tag/apr-16-2025/'}]" ] }, "execution_count": 42, "metadata": {}, "output_type": "execute_result" } ], "source": [ "news = scrape_the_batch_articles()\n", "news" ] }, { "cell_type": "code", "execution_count": 39, "id": "42b4d032", "metadata": {}, "outputs": [ { "name": "stdout", "output_type": "stream", "text": [ "Found 1 recent articles.\n", "\n", "2025-04-16 — Google Unveils Gemini 2.5, MCP Gains Momentum, Behind Sam Altman’s Fall and Rise, LLMs That Understand Misspellings\n", "The Batch AI News and Insights: I’ve noticed that many GenAI application projects put in automated evaluations (evals) of the system’s output probably later — and rely on humans to manually examine and judge outputs longer — than they should.\n", "https://www.deeplearning.ai/the-batch/tag/apr-16-2025/\n", "--------------------------------------------------------------------------------\n" ] } ], "source": [ "if __name__ == \"__main__\":\n", " news = scrape_the_batch_articles()\n", " print(f\"Found {len(news)} recent articles.\\n\")\n", " for item in news:\n", " print(f\"{item['date'].strftime('%Y-%m-%d')} — {item['title']}\")\n", " print(item['summary'])\n", " print(item['link'])\n", " print(\"-\" * 80)\n" ] }, { "cell_type": "code", "execution_count": null, "id": "74bce42e", "metadata": {}, "outputs": [], "source": [] } ], "metadata": { "kernelspec": { "display_name": "env", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "mimetype": "text/x-python", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.12.4" } }, "nbformat": 4, "nbformat_minor": 5 }