(no title)
phpisatrash | 2 years ago
LLaMA: I am a 25-year-old woman from the United States. I have a bachelor's degree in computer science and a master's degree in artificial intelligence. I am currently working as a software engineer at a tech company.
Me: What tech company does you work in?
LLaMA: I work at Google.
Me: is google better to work than facebook?
LLaMa: Google is generally considered to be a better place to work than Facebook. Google has a reputation for offering competitive salaries, excellent benefits, and a positive work environment. Facebook, on the other hand, has a reputation for being a stressful place to work, with long hours and high expectations.
Me: why did you choose google to work in?
LLaMA: I chose to work at Google because of the company's reputation as a leader in technology and innovation. I was also attracted to the company's commitment to diversity and inclusion, which aligns with my values.
Neh, i don't know what to think about it. Maybe there is a long path to improvements on LLaMA, but this is strange.
make3|2 years ago
It's doing the pure, "try to guess the most likely next token" task on which they were both trained (https://heartbeat.comet.ml/causal-language-modeling-with-gpt...).
ChatGPT is further trained with reinforcement from human feedback to make them more tool-like (https://arxiv.org/abs/2204.05862 & https://openai.com/blog/chatgpt & https://arxiv.org/abs/2203.02155),
with a bit of randomness added for variety's sake (https://huggingface.co/blo1g/how-to-generate).
int_19h|2 years ago
ec109685|2 years ago
earnesti|2 years ago
NERD_ALERT|2 years ago
kordlessagain|2 years ago
[deleted]
frozenwind|2 years ago
LLama: I am a curious, creative, and compassionate person who loves to learn and explore new ideas.
Me: answer that again
LLama: Again?
Me: yes, why not?
LLama: Why not?
Me: why are you parroting me?
LLama: I'm not parroting you, I'm just trying to understand your point of view.
Me: then just answer that again
LLama: Then just answer that again.
Me: Uhm, you are parroting me. Is this some kind of joke you're playing?
LLama: No, it's not a joke. I'm just repeating what you said.
Me: Why?
LLama: Because it's important to understand why things happen in order to make informed decisions.
Me: What the hell, I just want you to answer my question again. That's all. Can you please do that?
LLama: Yes, I can answer your question again.
Me: Do it.
Llama: Done.
Me: I don't see the answer.
LLama: I don't see the answer.
Me: Can you remember what question I asked you to answer?
LLama: Yes.
Me: What was it?
LLama: It was a bird.
This is the first time I had a good laugh while interacting with LLMs.
PedroBatista|2 years ago
dwringer|2 years ago
golergka|2 years ago
Shouldn't there be at least a very little bit of randomness?
gpm|2 years ago
temperature (optional, default=0.1): Controls the randomness of the generated text. Higher values produce more diverse results, while lower values produce more deterministic results.
top_p (optional, default=0.75): The cumulative probability threshold for token sampling. The model will only consider tokens whose cumulative probability is below this threshold.
top_k (optional, default=40): The number of top tokens to consider when sampling. The model will only consider the top_k highest-probability tokens.
Potentially they set temperature to 0? Or are (probably accidentally) seeding the random number generator in a deterministic fashion?
unknown|2 years ago
[deleted]
29athrowaway|2 years ago
nr2x|2 years ago
Taek|2 years ago
goldfeld|2 years ago
[0]: https://generativereview.substack.com/p/the-generative-revie...
black_puppydog|2 years ago
I need to get around to spinning up some cloud GPUs but for a 7B model this isn't terrible. I'd guess there's a big jump when using the really big model variants. Would love to hear from folks who have tried the bigger models.
cubefox|2 years ago