A new automated humanoid robot powered by OpenAI’s ChatGPT looks like something akin to the AI Skynet from the sci-fi movie Terminator.
While the new robot is not a killing machine, Figure 01 can perform basic autonomous tasks and maintain real-time conversations with humans, with the help of ChatGPT.
The company, Figure AI, shared a demo video showing how ChatGPT helps the two-legged machine visualize objects, plan future actions, and even reflect on its memory.
The figure’s cameras capture its surroundings and send them to a large vision language model trained by OpenAI, which then translates the images to the robot.
The clip showed a man asking the humanoid to put away dirty clothes, wash dishes, and give him something to eat, and the robot performed the tasks, but unlike ChatGPT, Figure is more hesitant when it comes to answering questions.
OpenAI released a demonstration video of its new robot, Figure 01
When a human asked for food, the robot was able to distinguish the apple as the only edible thing on the table
Figure AI hopes its first AI-enabled humanoid robot will prove capable of performing jobs too dangerous for human workers and could alleviate the worker shortage.
“Two weeks ago, we announced that Figure + OpenAI will join forces to push the boundaries of robot learning,” said Figure founder Brett Adcock. wrote in X.
“Together we are developing next-generation AI models for our humanoid robots,” he added.
Adcock also noted that the robot is not being controlled remotely from a distance and “this was filmed at 1.0x speed and filmed continuously.”
The comment about it being uncontrolled may have been a dig at Elon Musk, who shared a video of Tesla’s Optimus robot to show off its ability; it was later discovered that a human was operating it from a distance.
Figure AI raised $675 million in May 2023 from investors including Jeff Bezos, Nvidia, Microsoft, and of course, OpenAI.
“We hope to be one of the first groups to bring a humanoid to market,” Brett Adcock told reporters last May, “that can actually be useful and do business.”
Figure 01 is like something similar to Skynet from the Terminator movie.
The new video shows a man asking Figura to perform various tasks; In one of them, he asks the robot to give him something edible on the table.
“I see a red apple on a plate in the center of the table, a clothesline with cups and a plate, and you standing nearby with your hand on the table,” Figura said.
Adcock said the video showed the robot’s reasoning using its end-to-end neural networks, a language learning term used to train a model.
ChatGPT was trained with a large amount of data to interact with human users in a conversational manner.
The chatbot can follow an instruction in a message and provide a detailed response, which is how the language learning model in Figure works.
The robot took the apple and gave it directly to the man who was asking for food.
Figure 01 seemed to stutter at times, using words like ‘uh’ and ‘um’ which some people said make him sound more human.
The robot “listens” to a warning and responds with the help of its AI.
However, a recent study subjected ChatGPT to wargaming scenarios and found that it chose to bomb its adversaries almost 100 percent of the time, similar to Terminator’s Skynet.
But for now, Figure is lending a hand to the humans.
The video included another demonstration in which the man asks the robot what it sees on the desk in front of it.
The figure responded: “I see a red apple on a plate in the center of the table, a clothesline with cups and a plate, and you standing nearby with your hand on the table.”
Figure 01 could take out the trash and perform other household tasks, as well as answer questions in real time.
The robot uses onboard cameras connected to large visual language models to recognize its environment.
The US military is reportedly working with OpenAI to add its ChatGPT system to its arsenal.
Figure not only communicates, but also displays his cleaning skills by taking out the trash and putting dishes on the clothesline.
“We fed images from the robot’s cameras and transcribed text from speech captured by embedded microphones to a large OpenAI-trained multimodal model that understands both images and text,” Corey Lynch, an AI engineer at Figure, said in a report. mail in X.
“The model processes the entire history of the conversation, including past images, to generate linguistic responses, which are returned to the human via text-to-speech,” he added.
In the demo video, Figure showed signs of hesitation when answering questions, pausing to say “uh” or “um,” which some people commented makes the robot sound more human-like.
The robot still moves slower than a human, but Adcock said he and his team are “starting to approach human speed.”
Just over six months after that $70 million funding round last May, Figure AI announced a first-of-its-kind deal to put Figure to work in BMW factories.
The German automaker struck a deal to use the humanoids first at a BMW plant in Spartanburg, South Carolina, a sprawling, multimillion-dollar facility that includes high-voltage battery assembly and electric vehicle manufacturing.
While the announcement was short on details about the precise job tasks of robots at BMW, the companies described their intention to “explore advanced technology topics” as part of their “milestone-based approach” collaboration approach.
Adcock has framed the company’s goals as filling a gap for the industry in terms of alleged worker shortages involving complicated, skilled labor that conventional automation techniques have proven incapable of correcting.
“We need humanoid (robots) in the real world, doing real work,” Adcock said. axios.