top | item 36014135

(no title)

evmaki | 2 years ago

Super exciting to see the work happening in this area! I can especially appreciate the use of ChatGPT to orchestrate the necessary API calls, rather than relying on some kind of middleware to do it.

I have been working in this area (LLMs for ubiquitous computing, more generally) for my PhD dissertation and have discovered some interesting quality issues when you dig deeper [0]. If you only have lights in your house, for instance, the GPT models will always use them in response to just about any command you give, then post-rationalize the answer. If I say "it's too chilly in here" in a house with only lights, it will turn them on as a way of "warming things up". Kind of like a smart home form of hallucination. I think these sorts of quality issues will be the big hurdle to product integration.

[0] https://arxiv.org/abs/2305.09802

discuss

order

ftxbro|2 years ago

> If I say "it's too chilly in here" in a house with only lights, it will turn them on as a way of "warming things up".

is it bad that I'm on GPT's side on this, i mean if they are incandescent lights they are small electric heaters and what else do you expect it to do

awwaiid|2 years ago

Yeah but I think the idea is that it is a knob that calls to be turned. "It's warm in here" -> "I'll make the light blue so you feel nice and cool". "How fast do sparrows fly?" -> "Making the light brown". Like it might want to do _something_ and tweaking the hue or brightness are all it can do.

Good reason to always try to include in a prompt a way-out, a do-nothing or I-don't-understand answer.

bee_rider|2 years ago

It would probably be more useful to report to the user that it doesn’t have any control over than aspect of the environment.

I’d be curious to see what it does if it is told the sun is too bright…

evmaki|2 years ago

It certainly makes logical sense. I think if you have the ability to control the light in the first place via an API, it's probably an LED smart bulb and thus doesn't produce much heat. At least, I'm not aware of any incandescent smart bulbs.

dclowd9901|2 years ago

Gotta wonder why someone is saying “turn up the heat” to an AI that’s only connected to lights.

phh|2 years ago

> If I say "it's too chilly in here" in a house with only lights, it will turn them on as a way of "warming things up".

Thanks for the example that's interesting.

FWIW, this is pretty much what has been described as "waluigi" effect a bit extended: in a text you'll find on the internet, if some information at the beginning is mentioned, it WILL be relevant somewhere at some point later in that text. So an auto-completion algorithm will use all the information that has been given in the prompt. In your example it puts it in an even weirder situation where the model the overall model information (the lights, and that you're cold and nothing else), and it must generate a response. It would be a fun psychological study to look at, but I'm pretty sure even humans would do that in that situation (assuming they realize that lights may indeed produce a little bit of wattage of heat)

ftxbro|2 years ago

> FWIW, this is pretty much what has been described as "waluigi" effect a bit extended

Sorry I disagree for some reasons. First, turning the lights on is literally the only thing the bot can do to heat up the house at all. Turning on the lights does heat it up a little bit. So it's the right answer. Second, that's not the Waluigi effect, not even 'pretty much' and not even 'a bit extended'. Both of them are talking about things LLMs say, but other than that no.

The Waluigi effect applied to this scenario might be like, you tell the bot to make the house comfortable, and describe all the ways that a comfortable house is like. Then by doing this you have also implicitly told the bot how to make the most uncomfortable house possible. Its behavior is only one comfortable/uncomfortable flip away from creating a living hell. Say that in the course of its duties the bot is for some reason unable to make the house as comfortable as it would like to be able to do. It might decide that it didn't do it, because it's actually trying to make the house uncomfortable instead of comfortable. So now you got a bot turning your house into some haunted house beetlejuice nightmare.

famouswaffles|2 years ago

For performant enough models, you can just instruct it not to necessarily use that information in immediate completions.

adding something like

"Write the first page of the first chapter of this novel. Do not introduce the elements of the synopsis too quickly. Weave in the world, characters, and plot naturally. Pace it out properly. That means that several elements of the story may not come into light for several chapters."

after you've written up key elements you want in the story actually makes the models write something that paces ok/normally.

iamflimflam1|2 years ago

Very interesting paper!

It's something that I've been wondering about with ChatGPT plugins - they've kind of left it up to the user to enable/disable plugins. But there's definitely going to come a point where plugins conflict and the LLM is going to have to choose the most appropriate plugin to use.

I have been very impressed at how good it is at turning random commands into concrete API calls. You are right though, pretty much any command can be interpreted as an instruction to use a plugin.

evmaki|2 years ago

Thanks! That is part of the challenge as this idea scales imo - once you've increased the number of plugins or "levers" available to the model, you start to increase the likelihood that it will pull some of them indiscriminately.

To your point about turning random commands into API calls: if you give it the raw JSON from a Philips Hue bridge and ask it to manipulate it in response to commands, it can even do oddly specific things like triggering Hue-specific lighting effects [0] without any description in the plugin yaml. I'm assuming some part of the corpus contains info about the Hue API.

[0] https://evanking.io/posts/homegpt/

kingo55|2 years ago

Making IOT API calls is a solved problem with Home Assistant - plus it works locally.

Where I see this working best is giving Chat GPT some context about the situation in your home and having it work out complex automation logic that can't be implemented through simple rules.