Skip to main content

These white-hat Twitter bots collaborate to solve chemistry problems

Tek Image/Getty Images
Tek Image/Getty Images

For a lot of people, the term “Twitter bot” carries some negative connotations. But a fascinating research project coming out of the U.K.’s University of Glasgow is doing its bit to change that — by using algorithmic online communications not to tweet out controversial messages, but to carry out some cutting-edge chemistry.

Developed by chemistry professor Lee Cronin and his team, the #RealTimeChem project uses a pair of robots to perform chemical reactions in the lab, executing simple experiments involving mixing liquids and then recording the results. The Twitter part relates to the fact that the two robots aren’t physically in the same lab, but are in different ones, and are communicating with one another via Twitter.

Given simple chemical experiments (for example, finding a particular color liquid out of 117 possible combinations), the robots shared their findings with one another using the microblogging service. By performing experiments in this collaborative way, they were able to halve the time it took to answer specific questions because they were able to divide up the work between them.

Right now, this is a proof of concept in many ways, but Cronin told Digital Trends that it could be an exciting first step in a new means of doing chemistry. The idea of robots carrying out chemical experiments is not entirely new, but connecting them together like this is something we’ve not come across before. Furthermore, Cronin said that the concept is scalable beyond just two robots — so it would be possible to conceivably have hundreds of robots around the world working together and sharing their results.

“You could imagine that chemists could access the platform and send a message saying, ‘I’ve got a big problem I’m trying to solve. Can other people help me?’” Cronin said. He speculated that different laboratories could then help out by making different molecules and reporting back to one another.

“It allows you to delocalize, decentralize, and parallelize the making of molecules, to make it much faster,” he said. “It’s infinitely scalable. If you’ve got one robot doing 100 experiments in an hour, two robots could do 200 experiments, three could do 300 experiments, and so on. You’d have linear scaling, at least.”

It’s almost enough to make you love Twitter bots!

Editors' Recommendations

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Drone and rover tag teams could help solve the world’s deadly land mine problem
drone and rover wpi landmine project mjb 1464 landminerobot may1

Roughly 20,000 people are maimed or killed each year due to land mines; the vast majority being civilians in parts of the world subject to ongoing conflict. Could a tag team of autonomous robots help solve the problem? That’s certainly what a team of roboticists from Worcester Polytechnic Institute in Massachusetts are hoping. They have spent the past half-decade developing an autonomous robot and sandbag-dropping drone which work together to seek and destroy potentially deadly land mines. And they're almost ready for prime time.

As I’ve written before, collaborative robots are the future. Getting two (or potentially even more) different types of robot to work together means being able to combine the unique abilities of each and transform them into a cohesive solution that’s more than the sum of its parts. Have a slow, but steady rover which could methodologically sweep an area for mines using a metal detector, but has no way of getting rid of them? Have a drone with limited battery life and flight time which could be used to destroy mines by dropping something from above, while staying out of harm’s way? Get them talking to each other and suddenly you have a compelling partnership.
Robots working together
“It was always envisioned that [this project would feature] two different elements working together to solve the problem,” Craig Putnam, Senior Instructor in Robotics Engineering at Worcester Polytechnic Institute, told Digital Trends.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more
AI turned Breaking Bad into an anime — and it’s terrifying
Split image of Breaking Bad anime characters.

These days, it seems like there's nothing AI programs can't do. Thanks to advancements in artificial intelligence, deepfakes have done digital "face-offs" with Hollywood celebrities in films and TV shows, VFX artists can de-age actors almost instantly, and ChatGPT has learned how to write big-budget screenplays in the blink of an eye. Pretty soon, AI will probably decide who wins at the Oscars.

Within the past year, AI has also been used to generate beautiful works of art in seconds, creating a viral new trend and causing a boon for fan artists everywhere. TikTok user @cyborgism recently broke the internet by posting a clip featuring many AI-generated pictures of Breaking Bad. The theme here is that the characters are depicted as anime characters straight out of the 1980s, and the result is concerning to say the least. Depending on your viewpoint, Breaking Bad AI (my unofficial name for it) shows how technology can either threaten the integrity of original works of art or nurture artistic expression.
What if AI created Breaking Bad as a 1980s anime?
Playing over Metro Boomin's rap remix of the famous "I am the one who knocks" monologue, the video features images of the cast that range from shockingly realistic to full-on exaggerated. The clip currently has over 65,000 likes on TikTok alone, and many other users have shared their thoughts on the art. One user wrote, "Regardless of the repercussions on the entertainment industry, I can't wait for AI to be advanced enough to animate the whole show like this."

Read more