top | item 45528648

(no title)

HAL3000 | 4 months ago

All of the examples in videos are cherry picked. Go ask anyone working on humanoid robots today, almost everything you see here, if repeated 10 times, will enter failure mode because the happy path is so narrow. There should really be benchmarks where you invite robots from different companies, ask them beforehand about their capabilities, and then create an environment that is within those capabilities but was not used in the training data, and you will see the real failure rate. These things are not ready for anything besides tech demos currently. Most of the training is done in simulations that approximate physics, and the rest is done manually by humans using joysticks (almost everything they do with hands). Failure rates are staggering.

discuss

order

wongarsu|4 months ago

The last example they show (pick up package from pile, put it label-down on conveyor, repeat) seems to be the most realistic. They even have an uncut video of their previous model doing that for an hour on twitter [1].

I'm not sure that task needs a humanoid robot, but the ability to grab and manipulate all those packages and recover from failures is pretty good

1: https://x.com/adcock_brett/status/1931391783306678515

fragmede|4 months ago

Economy of scale. This guy wanted a factory in China to do a custom run of a product where they didn't include the shell. It was cheaper to just buy 10,000 units and have a line to de-shell the items than it was to change the original line. That special purpose robot for the one task is going to get beat out by the general purpose robot thats being produced at 100x the volume.

aDyslecticCrow|4 months ago

> I'm not sure that task needs a humanoid robot

An industrial robot arm with air powered suction cups would do the trick... https://bostondynamics.com/products/stretch/ ...

... So the task they work best at is the task there is already cheaper better robots specialized for.

Animats|4 months ago

That's the problem.

An obvious application, if this robot could do it, is retail store shelf restocking. That's a reasonably constrained pick and place task, some mobility is necessary, and the humanoid form is appropriate working in aisles and shelves spaced for humans. How close is that?

It's been tried before. In 2020.[1] And again in 2022.[2] That one runs on a track, is closer to an traditional industrial robot, and is used by 7-11 Japan.

Robots that just cruise around stores and inspect the shelves visually are in moderately wide use. They just compare the shelf images with the planogram; they don't handle the merchandise. So there are already systems to help plan the restocking task.

Technical University Delft says their group should be able to do this in five years.[3] (From when? No date on press release.)

[1] https://www.youtube.com/watch?v=cHgdW1HYLbM

[2] https://blogs.nvidia.com/blog/telexistence-convenience-store...

[3] https://www.tudelft.nl/en/stories/articles/shelf-stocking-ro...

MakeAJiraTicket|4 months ago

The Telexistence demo isn't so bad, but I have no idea why we're trying to make human robots generally. The human shape sucks at a most things, and we already have people treating roombas and GPT like their boyfriends or pets...

martythemaniak|4 months ago

Rodney Brooks (of iRobot fame) wrote an essay recently about why humanoids are likely decades and not years away from fulfilling their promise. It is quite long, but even a gpt summary will be quite valuable.

https://rodneybrooks.com/why-todays-humanoids-wont-learn-dex...

In short, he makes the case that unlike text and images, human dexterity is based on sensory inputs that we barely understand, that these robots don't have, and it will take a long time to get the right sensors in, get the right data recorded, and only then train them to the level of a human. He is very skeptical that they can learn from video-only data, which is what the companies are doing.

app13|4 months ago

Most companies are using world model simulations for training nowadays, like Issac Gym or Mujoco, not just video data.

smath|4 months ago

came here to see if anyone had read Rodney's recent essay - and to ask how does this announcement by Figure square with Rodney's essay.

The essay was long so I cant claim I read it in detail - one q in my mind is whether humanoids need to do dexterity the same way that humans do. yes they dont have skin and tiny receptors but maybe there is another way to develop dexterity?

skilled|4 months ago

You wrote what I wanted to write but I couldn't find such words.

Indeed, all the videos/examples are marketing pieces.

I would love to see a video like this "Logistics"[0] one, that shows this new iteration doing some household tasks. There is no way that it's not clunky and prone to all kinds of accidents and failures. Not that it's a bad thing - it would simply be nice to see.

Maybe they will do another video? Would love that.

[0]: https://www.youtube.com/watch?v=lkc2y0yb89U

andrewrn|4 months ago

This is always the case, though. The company is few years old. I'm no disciple of humanoids, but stuff has to start somewhere. Unfortunately hype > truth in order to get funding, so it produces incentives to cherry-pick like this.

ipnon|4 months ago

Now the question is if this is GPT-2 and we’re a decade away from autonomous androids given some scaling and tweaks, or if autonomous androids is just an extremely hard problem.

lossolo|4 months ago

https://www.figure.ai/company

"Building Figure won’t be an easy win; it will require decades of commitment and ingenuity."

"Our focus is on what we can achieve 5, 10, 20+ years from now, not the near-term wins."

At least it's not Musk's forever "next year".

kibwen|4 months ago

For LLMs, the input is text, and the output is text. By the time of GPT-2, the internet contained enough training data to make training an interesting LLM feasible (as judged by its ability to output convincing text).

We are nowhere near the same for autonomous robots, and it's not even funny. To continue to use the internet as an analogy for LLMs, we are pre-DARPANET, pre-ASCII, pre-transistor. We don't even have the sensors that would make safe household humanoid robots possible. Any theater from robot companies about trying to train a neural net based on motion capture is laughably foolish. At the current rate of progress, we are more than decades away.

hadlock|4 months ago

This is where I'm at. If you look at Boston Dynamics' first videos, they're 45 second clips of 4 legged robots walking in not even a straight line, just proving they could walk 5 feet over level ground without falling over. The top comment, from 4 years ago is "This was 11 years ago. Now these things are dancing." https://www.youtube.com/watch?v=3gi6Ohnp9x8

If you can make it look believable on camera for 15 seconds under controlled studio conditions... it's probable you can do it autonomously in 10-15 years. I don't think anyone is going to be casually buying these for their house by this time next year, but it certainly demonstrates what is realistically possible.

If they can provably make these things safe, it will have huge implications for in home care in advanced age, where instead of living in an assisted living home at $huge expense for 20+ years, you might be able to live on your own for most of that time.

I am cautiously optimistic.

jcims|4 months ago

I don't know if I caught your comment in my peripheral vision or what but GPT-2 is exactly where I conceptually placed this.

Neural networks for motion control is very clearly resulting in some incredible capability in a relatively short amount of time vs. the more traditional control hierarchies used in something like Boston Dynamics. Look at Unitree's G1

https://www.youtube.com/shorts/mP3Exb1YC8o

https://www.youtube.com/watch?v=bPSLMX_V38E

It's like an agile idiot, very physically capable but no purpose.

The next domain is going to be incorporating goals and intent and short/long term chains of causality into the model, and for that it seems we're presently missing quite a bit usable training data. That will clearly evolve over time, as will the fidelity of simulations that can be used to train the model and the learned experience of deployed robots.

pizzathyme|4 months ago

How does this square with the video where they showed it running continuously for an hour doing an actual Amazon package sorting job? https://www.youtube.com/watch?v=lkc2y0yb89U

dust42|4 months ago

> How does this square with the video where they showed it running continuously for an hour doing an actual Amazon package sorting job? https://www.youtube.com/watch?v=lkc2y0yb89U

The video shows several of glitches. From the comments:

  14:18 the Fall
  28:40 the Fall 2
  41:23 the Fall 3 
Also many of the packages on the left are there throughout the video.

But then I think lots of this can be solved in software and having seen how LLMs have advanced in the last few years, I'd not be surprised to see these robots useful in 5 years.

daveguy|4 months ago

Is it really sorting? All I see is the humanoid robot moving similarly shaped / sized packages from one conveyor belt to a platform to another conveyor belt. A little industrial automation design would be much more effective, cheaper, and faster compared to the task it is performing.

godelski|4 months ago

I'm really confused by this video. What is it even supposed to be doing?

Is it supposed to be taking packages and placing them label face down?

I cannot understand how a robot doing this is cheaper than a second scanner so you can read the label face down or face up. I mean you could do that with a mirror.

But I'm not convinced it is even doing that. Several packages are already "label side down" and it just moves them along. Do those packages even have labels? Clearly the behavior learned is "label not on top", not "label side down". No way is that the intended behavior.

If the bar code is the issue, then why not switch to a QR code or some other format? There's not much information you need in shipping so the QR code can have lots of redundancy, making it readable from many different angles and even if significantly damaged.

The video description also says "approaching human-level dexterity and speed". No way. I'd wager I could do this task at least 10x its speed, if not 20x. And that I'd do it better! I mean I watched a few minutes at 2x speed and man is it slow. Sure, this thing might be able to run 24/7 without breaks, but if I'm running 10-20x faster then what's that matter? I could just come in a few hours a day and blow through its quota. I'd really like to see an actual human worker for comparison.

But if we did want something to do this very narrow task for 24/7, I'm pretty sure there are a hundred different cheaper ways to do it. If there aren't, then it is because there is some edge cases that are pretty important. And without knowing that then we can't actually properly evaluate this video. Besides, this video seems like a pretty simple ideal case. I'm not sure what an actual amazon sorting process looks like, but I suspect not like this.

Regardless, the results look pretty cool and I'm pretty impressed with Figure even if it is an over-simplified case.

JKCalhoun|4 months ago

> There should really be benchmarks where you invite robots from different companies

…and have a surprise dance-off.

sheepybloke|4 months ago

It was interesting to see that they didn't show the robot folding laundry; rather, just it laying out the clothes.

xnyan|4 months ago

There’s a scene about 2/3 through the first video where they show a brief clip of the robot folding and stacking a shirt. The quality and speed was roughly comparable to a 7-10 year old - slow and somewhat sloppy, but recognizably a folded shirt.

WanderPanda|4 months ago

I think this is the frontier when it comes to "unstructured":

https://youtu.be/nmEy1_75qHk

They for sure did not anticipate that the user would backflip into their robot and knock it (and himself) out :D

kevin_thibedeau|4 months ago

I didn't know Ruby Rhod's great grandfather was already alive.

someoneontenet|4 months ago

How did I know this was gonna be speed lol

guerrilla|4 months ago

No doubt. This video is just to hype investors. I'm sure they have more practical business plans than this.

more_corn|4 months ago

I wonder if instead of making a robot to interact with the washing machine we should try to make a washing machine with an input and output hopper. Dump clothing in get clean folded clothing out.

You can control the happy path when the whole thing is your box.

robots0only|4 months ago

+100!!! Please don't fall for the HYPE.

The current best neural networks only have around 60% success rates for small horizon tasks (think 10-20 seconds e.g. pick up apple). That is why there is so much cut-motions in this video. The future will be awesome but it will take time a lot of research still needs to happen (e.g. robust hands, tactile, how to even collect large scale data, RL).

brailsafe|4 months ago

> The future will be awesome

Perhaps this is a bit pedantic, but what about the probable eventual proliferation of useful humanoid robots will make the future awesome? What does an awesome future look like compared to today, to you?

tamimio|4 months ago

> Go ask anyone working on humanoid robots today, almost everything you see here, if repeated 10 times, will enter failure mode

As someone who worked in the robotics industry, 90% of the demos and videos are cherry-picked, or even blatantly fake. That's why for any new robot in the market, my criteria is: Can I buy it? If it's affordable and the consumer can buy it and find it useful in day to day life, then this robot is useful and has potential; other than that, it's just an investor money grab PR hype.

deadbabe|4 months ago

This is so damning. How are you not afraid of retribution from big players in the AI space? You pretty much destroyed their company.

jayd16|4 months ago

I'd watch that game show.

aowie|4 months ago

Oh wow, a robot that can play with my dog so I don't have to. That's exactly the kind of task I'd be relieved to automate.

The fabric wrap is idiotic. Insanely stupid. Let's have an expensive fabric-covered robot wash dishes covered in food. Genius. It's a good thing those "dirty dishes" were already perfectly clean. I doubt this machine could handle anything more. Put it in a real commercial kitchen and have it scrape oven pans and I'll be impressed.

I'm so glad I left robotics. I don't want to have anything to do with this very silly bubble.