top | item 40299061

(no title)

prions | 1 year ago

It looks really interesting, but as an experienced climber I'm not sure if just watching a video of my avatar climbing would really help with skill acquisition.

Also, this claims that the wall type or video quality doesn't matter, but I have a hard time understanding how the model would be able to understand that a small crimp could possibly be dual textured and therefore has only a few specific ways of approaching it.

So it seems that this is more for visualizing a climb (which is a skill most climbers should develop) and not really for dialing in some sort of microbeta for a problem.

discuss

order

petsfed|1 year ago

I suspect that, absent that information about the exact right way to grab a hold, or the exact way to put a foot on a hold, you'll be limited to beta suggestions, which is fine, I think. It'd be like having a group of other climbers nearby to suggest different beta, even if you don't have any friends.

So, in terms of solving complicated beta faster, I see real utility to this.

It is very interesting that since the AI climber is trained on actual climbers, it could, in principle, provide beta to climb consistent with your own style. If you train the bot exclusively on footage of yourself, it would return movement based on your style. If your style is finessy-all-backstep-all-the-time (aka The Edlinger), it can provide beta consistent with that. If your style is to square up and pull (otherwise known as The American), it can provide beta consistent with that instead.

bmj|1 year ago

(Long time climber here)

> It is very interesting that since the AI climber is trained on actual climbers, it could, in principle, provide beta to climb consistent with your own style. If you train the bot exclusively on footage of yourself, it would return movement based on your style. If your style is finessy-all-backstep-all-the-time (aka The Edlinger), it can provide beta consistent with that. If your style is to square up and pull (otherwise known as The American), it can provide beta consistent with that instead.

I would think this is actually a Bad Thing. It's very easy to get stuck trying to make a sequence fit your style of climbing. The better approach (especially for long term skill acquisition) is a willingness to learn new styles. That's to say that every sequence is only solvable via one particular style, but I think long term development is hindered if you approach every crux with the one thing you are good at.

> So, in terms of solving complicated beta faster, I see real utility to this.

I can agree with this. But, to the point that others have made, I do wonder what this and the availability of beta videos for many, many routes and blocs does to climbing skill overall. Perhaps I'm just a grumpy old man, but, particularly when bouldering, sorting out the beta should be part of the journey toward eventually sending. Last fall, I visited Hueco Tanks after a six year absence. I suppose I was a bit disappointed to see so many people watching YouTube beta videos of nearly every problem they tried.

etrautmann|1 year ago

Agreed - so much about the detail of how you would climb something comes down to details that would be hard to measure with a camera, like textures, your estimate of friction, etc. Very cool idea though, looks fun to test.

sandworm101|1 year ago

And this is deep into "sport climbing", borderline gym rat territory imho. It doesn't model all the other core aspects such as protection, rope management, exposure and rest stops. I imaging if you pointed this at a real cliff and recorded several assents it would quickly become a blurry-twitchy mess as all the movements not touching the rock spoiled the data. Maybe for bouldering, but not for real climbing.

Snackchez|1 year ago

What could be interesting is if you could compare your attempts to the avatar climbing and receiving feedback afterwards. This would effectively be a step up from simply recording your send attempts.

logtempo|1 year ago

I don't see the shape of the holds being a big problem. With some help from indoor companies and hold makers, figuring out which hold model is on the wall should be possible.

As for the usefullness of the software, I'm sceptical too as it don't really solve a problem. But maybe I'm not seeing it and it could be good for beginners :) A good improvement would be adding a comparison between you and the model in term of body position and fluidity of movements.

party_possum|1 year ago

The idea of incorporating actual hold data and "recognizing" specific holds is interesting, but I'm not sure it completely solves the problem.

The "Boss" from Pusher is arguably the most famous climbing hold ever made. For a decade or more, every gym had one, but they were all unique. Lots of them had micro chips that became critical to usage of the hold. Some had decent texture and some were glassy smooth from years and years and years of use. A lot of the accidental variation in new holds has gone away as the industry has standardized around a handful of industrial fabricators like Aragon, but even over the course of a single indoor boulder problem's life, the accumulation of chalk, sweat, and shoe rubber can have a significant impact on how a hold climbs.

I guess the real question is, do these changes just make routes harder or do they make them fundamentally different? Do they actually change the set of moves that constitutes the easiest way to the top? To be honest, I'm not entirely sure. But it's something interesting to think about.

Zanfa|1 year ago

> I don't see the shape of the holds being a big problem. With some help from indoor companies and hold makers, figuring out which hold model is on the wall should be possible.

Even if you know the exact hold model and it’s in pristine condition, it’s basically impossible to tell how it’s gonna work from a single angle on video at a distance. Even tiny variations in angle of the wall and rotation of the hold on the wall can completely change how you use it.

smandava|1 year ago

To clarify:

Of the full distribution of possible video qualities one can take on a modern phone camera, the vast majority of video qualities will be fine for the AI to understand fine details. Obviously, if you somehow or for some reason, take a video with really bad quality, it will not give you what you want.

Same explanation goes for the walls. If you take a video of just a really dark wall with really bad holds, it is probably won't give you what you want either.

smandava|1 year ago

What do you think are some use cases where an avatar simulation of you can help (if any)?