Interesting; would you have any specific recommendations for an intelligent (high parameter count? Something that can hold a conversation ...), uncensored, specific model that is able to do this?
If anyone is looking for a cloud option (e.g. don't have the hardware, just want to pay someone else to handle the compute while still writing about generally uncensored things like a violent RPG adventure or something explicit in the adult sense), then I guess NovelAI could fit the bill: https://novelai.net/
I guess most people focus on their image generation which is pretty cool for anime characters and things like placeholder art (say, in gamedev), but their text generation options are also pretty nice and can be used either through their web UI, or something like SillyTavern: https://github.com/SillyTavern/SillyTavern
If you do want to run something locally and have the hardware, I think people discuss models in a variety of places, like https://old.reddit.com/r/LocalLLaMA/comments/16425mk/best_ns... (quick search, was curious, NSFW discussion) and it seems like there's plenty of models on HuggingFace, though I haven't particularly looked into them.
I will absolutely say that even when I was trying to run a general purpose chatbot LLM, it was kind of impressive how quickly things are moving ahead in regards to LLMs: all of these GGML, GGUF and other formats, being able to use even consumer GPUs (though I couldn't get ROCm working with my RX580) and experimental attempts like using Vulkan.
I think most custom fine-tunes and merges on HuggingFace will do this unless they specifically mention it being censored. Even the lower param models have been surprisingly good, with relatively fast progress being made in the 7b and 11b models.
My "daily driver" is Fimbulvetr v2 11b, surprisingly slapped together by an EMT. Kunoichi 7b seems to be a pretty popular model too. These can be run locally with as little as 8 GB free RAM (preferably VRAM) with an easy install solution like LMStudio or Faraday.
You can generally find a lot of recommendations in places like SillyTavernAI or LocalLLaMa on reddit:
This one works with GGUF-compatible llama.cpp wrappers like the fully-local, fully-private Layla app or the other examples tossed around. With some prompt tweaking it's capable of a broad variety of... tasks.
This model is very compliant and doesn't tend to go to "I'm sorry, as a large language model..." type replies.
isoprophlex|1 year ago
KronisLV|1 year ago
I guess most people focus on their image generation which is pretty cool for anime characters and things like placeholder art (say, in gamedev), but their text generation options are also pretty nice and can be used either through their web UI, or something like SillyTavern: https://github.com/SillyTavern/SillyTavern
If you do want to run something locally and have the hardware, I think people discuss models in a variety of places, like https://old.reddit.com/r/LocalLLaMA/comments/16425mk/best_ns... (quick search, was curious, NSFW discussion) and it seems like there's plenty of models on HuggingFace, though I haven't particularly looked into them.
I will absolutely say that even when I was trying to run a general purpose chatbot LLM, it was kind of impressive how quickly things are moving ahead in regards to LLMs: all of these GGML, GGUF and other formats, being able to use even consumer GPUs (though I couldn't get ROCm working with my RX580) and experimental attempts like using Vulkan.
meat_machine|1 year ago
My "daily driver" is Fimbulvetr v2 11b, surprisingly slapped together by an EMT. Kunoichi 7b seems to be a pretty popular model too. These can be run locally with as little as 8 GB free RAM (preferably VRAM) with an easy install solution like LMStudio or Faraday.
You can generally find a lot of recommendations in places like SillyTavernAI or LocalLLaMa on reddit:
https://old.reddit.com/r/SillyTavernAI/comments/1brig2n/whic...
gryfft|1 year ago
This one works with GGUF-compatible llama.cpp wrappers like the fully-local, fully-private Layla app or the other examples tossed around. With some prompt tweaking it's capable of a broad variety of... tasks.
This model is very compliant and doesn't tend to go to "I'm sorry, as a large language model..." type replies.