Skip to main content

Marimba-playing robot uses deep-learning AI to compose and perform its own music

When the inevitable robot invasion happens, we now know what the accompanying soundtrack will be — and we have to admit that it’s way less epic than the Terminator 2: Judgment Day theme. Unless you’re a massive fan of the marimba, that is!

That assertion is based on research coming out of the Georgia Institute of Technology, where engineers have developed a marimba-playing robot with four arms and eight sticks that is able to write and perform its own musical compositions. To do this, it uses a dataset of 5,000 pieces of music, combined with the latest in deep learning neural network-based AI.

Recommended Videos

“This is the first example of a robot composing its own music using deep neural networks,” Ph.D. student Mason Bretan, who first began working on the so-called Shimon robot seven years ago, told Digital Trends. “Unlike some of the other recent advances in autonomous music generation from research being done in academia and places like Google, which is all simulation done in software, there is an extra layer of complexity when a robotic system that lives in real physical three-dimensional space generates music. It not only needs to understand music in general, but also to understand characteristics about its embodiment and how to bring its musical ‘ideas’ to fruition.”

Robot Composes, Plays Own Music Using Deep Learning

Training Shimon to generate new pieces of music involves first coming up with a numerical representation of small chunks of music, such as a few beats or a single measure, and then learning how to sequence these chunks. Two separate neural networks are used for the work — with one being an “autoencoder” that comes up with a concise numerical representation, and the second being a long short-term memory (LSTM) network that models sequences from these chunks.

“These sequences come from what is seen in human compositions such as a Chopin concerto or Beatles’ piece,” Bretan continued. “The LSTM is tasked with predicting forward, which means given the first eight musical chunks, it must predict the ninth. If it is able to successfully to do this, then we can provide the LSTM a starting seed and let it continue to predict and generate from there. When Shimon generates, it makes decisions that are not only based off this musical model, but also include information about its physical self so that its musical decisions are optimized for its specific physical constraints.”

It’s pretty fascinating stuff. And while the idea of a music-generating bot might sound of interest only to people studying music, the bigger questions it raises about computational creativity are only going to get more important as time goes on.

“Though we are focusing on music, the more general questions and applications pertain to understanding the processes of human creativity and decision-making,” Bretan said. “If we are able to replicate these processes, then we are getting closer to having a robot successfully survive in the real world, in which creative decision-making is a must when encountering new scenarios and problems each day.”

Luke Dormehl
Former Digital Trends Contributor
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Amazon’s AI shopper makes sure you don’t leave without spending
Amazon Buy for Me feature.

The future of online shopping on Amazon is going to be heavily dependent on AI. Early in 2025, the company pushed its Rufus AI agent to spill product information and help users find the right items. A few weeks later, another AI tool called Interests made its way to the shopping site. 

The new Alexa+ AI assistant is also capable of placing orders semi-autonomously, handling everything from groceries to booking appointments. Now, the company has started to test yet another AI agent that will buy products from other websites if they’re not available on Amazon — without ever leaving the app. 

Read more
Google Gemini’s best AI tricks finally land on Microsoft Copilot
Copilot app for Mac

Microsoft’s Copilot had a rather splashy AI upgrade fest at the company’s recent event. Microsoft made a total of nine product announcements, which include the agentic trick called Actions, Memory, Vision, Pages, Shopping, and Copilot Search. 

A healthy few have already appeared on rival AI products such as Google’s Gemini and OpenAI’s ChatGPT, alongside much smaller players like Perplexity and browser-maker Opera. However, two products that have found some vocal fan-following with Gemini and ChatGPT have finally landed on the Copilot platform. 

Read more
Rivian set to unlock unmapped roads for Gen2 vehicles
rivian unmapped roads gen2 r1t gallery image 0

Rivian fans rejoice! Just a few weeks ago, Rivian rolled out automated, hands-off driving for its second-gen R1 vehicles with a game-changing software update. Yet, the new feature, which is only operational on mapped highways, had left many fans craving for more.
Now the company, which prides itself on listening to - and delivering on - what its customers want, didn’t wait long to signal a ‘map-free’ upgrade will be available later this year.
“One feedback we’ve heard loud and clear is that customers love [Highway Assist] but they want to use it in more places,” James Philbin, Rivian VP of autonomy, said on the podcast RivianTrackr Hangouts. “So that’s something kind of exciting we’re working on, we’re calling it internally ‘Map Free’, that we’re targeting for later this year.”
The lag between the release of Highway Assist (HWA) and Map Free automated driving gives time for the fleet of Rivian vehicles to gather ‘unique events’. These events are used to train Rivian’s offline model in the cloud before data is distilled back to individual vehicles.
As Rivian founder and CEO RJ Scaringe explained in early March, HWA marked the very beginning of an expanding automated-driving feature set, “going from highways to surface roads, to turn-by-turn.”
For now, HWA still requires drivers to keep their eyes on the road. The system will send alerts if you drift too long without paying attention. But stay tuned—eyes-off driving is set for 2026.
It’s also part of what Rivian calls its “Giving you your time back” philosophy, the first of three pillars supporting Rivian’s vision over the next three to five years. Philbin says that philosophy is focused on “meeting drivers where they are”, as opposed to chasing full automation in the way other automakers, such as Tesla’s robotaxi, might be doing.
“We recognize a lot of people buy Rivians to go on these adventures, to have these amazing trips. They want to drive, and we want to let them drive,” Philbin says. “But there’s a lot of other driving that’s very monotonous, very boring, like on the highway. There, giving you your time back is how we can give the best experience.”
This will also eventually lead to the third pillar of Rivian’s vision, which is delivering Level 4, or high-automation vehicles: Those will offer features such as auto park or auto valet, where you can get out of your Rivian at the office, or at the airport, and it goes off and parks itself.
While not promising anything, Philbin says he believes the current Gen 2 hardware and platforms should be able to support these upcoming features.
The second pillar for Rivian is its focus on active safety features, as the EV-maker rewrote its entire autonomous vehicle (AV) system for its Gen2 models. This focus allowed Rivian’s R1T to be the only large truck in North America to get a Top Safety Pick+ from the Insurance Institute for Highway Safety.
“I believe there’s a lot of innovation in the active safety space, in terms of making those features more capable and preventing more accidents,” Philbin says. “Really the goal, the north star goal, would be to have Rivian be one of the safest vehicles on the road, not only for the occupants but also for other road users.”

Read more