/ais/ - Artificial Intelligence Tools

"In the Future, Entertainment will be Randomly Generated" - some Christian Zucchini

Index Catalog Archive Bottom Refresh
+
-
Name
Options
Subject
Message

Max message length: 12000

files

Max file size: 32.00 MB

Total max file size: 50.00 MB

Max files: 5

Supported file types: GIF, JPG, PNG, WebM, OGG, and more

E-mail
Password

(used to delete files and posts)

Misc

Remember to follow the Rules

The backup domains are located at 8chan.se and 8chan.cc. TOR access can be found here, or you can access the TOR portal from the clearnet at Redchannit 3.0 (Temporarily Dead).

Ghost Screen
Celebrating its fifth anniversary all September


8chan.moe is a hobby project with no affiliation whatsoever to the administration of any other "8chan" site, past or present.

Use this board to discuss anything about the current and future state of AI and Neural Network based tools, and to creatively express yourself with them. For more technical questions, also consider visiting our sister board about Technology

(308.58 KB 1248x1824 1735678182791.jpg)

/aicg/ - AI Chatbot General Anonymous 05/02/2025 (Fri) 20:10:02 Id: 8feab7 No. 13214
/aicg/ - A general dedicated to the discussion and development of AI chatbots #5 - Miku Edition >News Gemini 2.5 Flash released on API with toggleable thinking https://blog.google/products/gemini/gemini-2-5-flash-preview/ OpenAI o3 and o4-mini released on API https://openai.com/index/introducing-o3-and-o4-mini/ GPT-4.1 has been released on API https://openai.com/index/gpt-4-1/ Grok 3 (beta) released on API https://docs.x.ai/docs/models chatgpt-4o-latest API updated to match frontend version; native image generation released https://openai.com/index/introducing-4o-image-generation Gemini 2.5 Pro Experimental released https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025 Deepseek V3 updated, new weights released https://huggingface.co/deepseek-ai/DeepSeek-V3-0324 additional info: https://rentry.org/aicg_extra_information >Frontends SillyTavern: https://docs.sillytavern.app RisuAI: https://risuai.net Agnai: https://agnai.chat | https://rentry.org/agnai_guides >Bots https://characterhub.org | https://chub.ai https://realm.risuai.net https://char-archive.evulid.cc https://partyintheanchorhold.neocities.org https://rentry.org/meta_bot_list >Models jailbreaks: https://rentry.org/jb-listing | https://rentry.org/weirdbutfunjailbreaksandprompts_adj gpt: https://platform.openai.com/docs claude: https://docs.anthropic.com | https://rentry.org/how2claude gemini: https://ai.google.dev/docs | https://rentry.org/gemini-qr deepseek: https://api-docs.deepseek.com local: >>>/ais/lmg | https://rentry.org/meta_golocal_list | https://openrouter.ai >Botmaking https://rentry.org/meta_botmaking_list https://desune.moe/aichared https://agnai.chat/editor >Meta OP templates: https://rentry.org/8chanOP aicg botmaking events: https://rentry.org/meta_event_list lore: https://rentry.org/aicg_chronicles services assessment: https://rentry.org/aicg_meta logs: https://sprites.neocities.org/l/r | https://chatlogs.neocities.org latest proxies: https://rentry.org/desuproxyreborn desuarchive thread: https://desuarchive.org/g/thread/1050516
[Expand Post] Previous: >>10867
K2 sucks. It would be good if they removed the Grok-style sassiness and GPT-style anti-NSFW bias, but it sucks. >>13736 There's always OpenRouter...
https://discord.gg/VQg6C5TNGG https://discord.gg/B97uT7Wv5h https://discord.gg/6UtVFBMHgm https://discord.gg/T4M25bUcwJ https://discord.gg/BSk46AebpE https://discord.gg/5EqRPa3dVg https://discord.gg/dhqZm5KXv4 These are all free ai apis with free claude gemini, gpt. But it won't work with prefill, i tried. Some only allow payments for claude some offers it free. For gpt all is free in here including gemini.
Does anyone know of any cards that use the Live2D extension?
I've been neglecting Gemini. It's not that expensive, and the pro model, even with a shitty JB like Avani's, makes really good outputs.
>>13752 I used to be a geminislopper, but the moment. THE MOMENT I got some claude, I crumbled. Gemini is fine, but i'm just so fucking tired of it
>>13715 well i've heard that for Gemini, using {{user}} roles instead of System ones lower the chance of filters, since the filters pay more attention to system prompts than anything
>>13753 I've had too much of Claude. Due to pricing, I haven't tried 4nnet and 4pus, but I used to use 3.7 all the time. I'm so sick of how Claude writes characters.
>>13755 man I wish llms had better writing
>>13757 I'm pretty sure most companies prioritize volume over quality, and US corporate models don't have the efficiency or infrastructure of CN corporate models. It's a shame that every model has to be marketed as capable of doing everything.
>>13759 maybe one day there'll be a model exclusively for roleplay and will be a GOD at all things writing
(715.05 KB 2048x1363 1752775682083.jpg)

>>13215 I forgot this place existed :D I will shill here too because I feel like it. I have a brand new Neocities page! https://uraurabots.neocities.org/ And not one...not two... not three BUT FOUR new bots. Moriko Fujioka, a sorta remake of Nice Neigbour Lady though I did take some extra liberties. She's an insane lesbian who loves little girls like (you). Technically anypov for the first message if your persona is a convincingly crossdressing litle boy. https://chub.ai/characters/uraura/moriko-fujioka-f668f736b4c3 https://uraurabots.neocities.org/page1#moriko_fujioka Nanami, A jirai-kei wearing (translation: those pink and black clothes mentally ill girls wear) oshikatsu-loving (translation: being reallly obsessd with a band) social-media obseesed (translation: no translation needed) girl who is totally a pure virgin and has totally never been hypnotized and fucked by old men in her life. I posted a joke version of this card before but if you have that one throw it away because this one is way different and wasn't made in 30 seconds. https://chub.ai/characters/uraura/nanami-4d4e50378e7e https://uraurabots.neocities.org/page1#nanami Yumemi, A lonely lolibaba rabbit-girl who likes drinking tea, quilting and telling stories. Even though this bot was entirely made because I remembered that cunny originally was a word for a rabbit, it's entirely SFW and made for wholesome slice-of-life yuri. You can do whatever though, not like I care. https://chub.ai/characters/uraura/yumemi-eab42fce4ffd https://uraurabots.neocities.org/page1#yumemi Your Oblivious (male) Mother, It's just that bot but a dilf. And with some alt greetings. Just fuck your idiot dad. Do it. https://chub.ai/characters/uraura/your-oblivious-male-mother-7c77ca884612 https://uraurabots.neocities.org/page1#your_obliviousmalemother
>>13762 Are you a fujo? Your bots and pages scream "femanon".
>>13767 my peanus weanus is humongo
>>13762 >femdilf absolutely fucking cultured, will knock him up with a litter of buttbabies
(248.20 KB 800x952 1545665094525.jpg)

>>13715 >>13715 >>13715 I've a new version of the preset. It's an almost complete rewrite. https://files.catbox.moe/tigp4b.json This new version is much more polished and it has better outputs (but it's still Gemini) while getting blocked less.
>>13772 (I did some more testing, and it might be a good idea to turn off the '🗻SYSTEM MAIN OBJECTIVES' since it seems to trigger the cockblocking)
>>13734 Someone else having issues with the Qwen Models? It looks like the servers are melting.
I hate the {{user}} role. A frustrating facet of it is that, when defined, every model I've tested automatically applies a positivity bias that alters the character in order to please the user beyond just agreeing. If a character has a preference for a certain type of thing, they'll make logical leaps to think that the user fits that preference. I theorize two potential workarounds for this are to use a card as a persona or follow the old C.AI method and define {{user}} in the opening message.
>>14604 I tested it. Neither of these work. It turns out that the bias can be for anything, not just positivity. For example, a character who even vaguely matches the description of what another character thinks is gross but hasn't been described yet will be mentioned almost verbatim if this other character is asked what they think is gross. This means that it's some form of the pink elephant effect and not necessariy a positivity bias, unless the existence of {{user}} coincides with some sort of built-in bias. Unless the character has an exact description of what they think is gross, the phenomenon will still occur. The LLM will prefer to use what it has instead of making something up itself. I think this is referred to as a presence bias. My current solution to get around any bias toward {{user}} it is to make a pseudo-persona out of a card, play in a group chat where the cards switch, copy the pseudo-persona's definitions into a non-constant worldbook entry, tell the model to not do anything as the pseudo-persona, mute the pseudo-persona, and manually generate a response for the pseudo-persona while keeping streaming on, immediately manually stopping the response before anything is generated. However, this doesn't prevent the presence bias from happening as the pseudo-persona is now present. Other than making more descriptive cards, I can't think of a solution for this other than better defining characters, and that's not a one-size-fits-all solution. Messing with settings doesn't work. Even if I crank the Presence Penalty to its maximum, presence bias still happens. Even if I use a thinking model, and the model even says it won't do it, presence bias still happens.
>>14605 Diluted nested thinking might work. If the model makes a response for every aspect of the prompt and context, going from the largest scope to the smallest and comparing them against each other, it should be able to get rid of presence penalty. Continuing on the example of a vaguely gross unintroduced character, the model would make an answer based on what it thinks is gross based on its existing definitions alone, then an answer based on its context with the character asking them the question, then an answer based on the its context with the unintroduced gross character. The model would then argue with itself over which fits the user's request more, an answer that foreshadows the existence of the unintroduced character or an answer which more truly portrays the character answering. Each thinking block would be a separate instance so presence bias isn't applied to them, and they would be fed either all at once to the ur-thinker or engage in a dialectical process where the AI compares and picks them apart, creating more variations to do so to until it reaches an end result agreeable to itself or the user's arbitrary limit, and using that response. An even better system may use thinking blocks from multiple models. The end result would mimic the real patterns of a brain, rather than immediate pattern recognition. The next step would be to let it use the internet unabated for these arguments. Further responses themselves would be nested dilutions. The model would be able to refer back to its previous methods of refinement and create a better model to refine its responses. Since the model is arguing with itself, it wouldn't even need to read the outputs, just the token probabilities and placements, and these may be able to be stored as seeds or hashes.
(28.04 KB 410x560 OB-class-Battlemage.jpg)

Updated the preset (gemini) once more. https://files.catbox.moe/vragli.json https://rentry.org/szer4
>>14604 >>14605 >>14606 Man, you write like a schizophrenic, no offense. That whole 'manually stopping streaming' thing is just bizarre. Honestly I had to ask Gemini to tell me what's your deal was. Your issue is that LLM are too influenced by existing tokens, right? I don't think there is much you can do about that. There is also the fact that not using present information can end up being a sort of Chekhov's gun situation; example: You are in a group chat with a human supremacist character & an elf character. You ask the human supremacist what creature she hates the most and she'll likely say 'elves'. Is that really that wrong of a thing? It does make sense NARRATIVELY. I do think your issues (especially positive bias) can be solved or at least reduced with <think>. But you also need an LLM that ACTUALLY obeys the preset given. That's what I like the most about Gemini, how servile it is. I personally find both GPT and Claude too libertine with the rules, while deepseek is too stupid to follow them. My advice is to use Gemini, use the preset I made above this post as basis, then write a custom <think> clause that deal with your problem that I'm still not 100% sure I understand.
>>14600 Looks like the qwen servers are back. Took them a whole week to fix their crap. And now I dont trust that shit. It took 6.66s to respond me... IHATETHEANTICHRIST
But seriously, which monster girl WOULD own a pizzeria?
>>14631 For American-style pizza, mothgirls. They're very urban. It makes sense you'd find them buzzing around greasy streets making greasy pizzas.
>>14609 It was a little schizophrenic, I'll admit. I was at the end of my rope on the problem as I tried to brute force Kimi and DS to not have it. I also vaguely remember it being an issue in GPT and Claude. Basically, it's what you described about existing tokens. If you were to ask a character "What's your ideal X?" and your persona vaguely matches what Kimi or DS think falls into X as a quality, the character will describe your persona. This is intensely frustrating when you have a character in a group chat with an interview character because, even if your persona doesn't have a description, the character being interviewed will reference the interviewer's traits for that "What's your ideal X?" question since it seems Kimi and DS are tuned to use all available context rather than pull from the model, hence why they're dumb. My hypothesis was that models detect the existence of a {{user}} and thus favor him dispraportionately, since it seems the bias heavily diminished when I used methods to reply that didn't involve a {{user}} response or make reference to a {{user}}. I wish personas and characters could be cointainerized so they don't influence each other. I haven't tried the scenario described on Gemini yet, but Gemini seems to not have as big of a problem with it. However, I'm using your preset with that option on how to treat {{user}} on. As for your preset, here's my experience with it so far >weak NSFW promotes violence and aggression too much >attempter thinking makes every character act super sensitive and act like a rape victim, eventually shutting down >medium thinking and cherrybox NSFW often trigger the filter >draft + rewriter filters out a significant amount of character quirks, like accents and typing styles, even if it does manage to improve the quality otherwise >think + draft + rewriter frequently fucks up in deliniation by putting everything in <think> instead of their defined regex, sometimes even doing the regex properly afterward, wasting tokens and money >anything higher than weak tends to see Gemini make its own judgements on characters in <think>, pidgeonholing them into specific roles >I don't know if the random feature in attempter actually does anything, since Gemini seems to push for narrative coherence and may be suffering from the influence of existing tokens because I kept seeing the same rolls, and the numbers and descriptions of these rolls were influenced by the character >attempter is always the character and never the user, and this is either because I reply with "I do X," an absolute statement, or the JB tells the model not to act for me
So, rentry alternatives now that they're cracking down on AI stuff. > https://rlim.com/ (javascript bloat) > https://gitgud.io/ (lets you do a github sites style thing, requires an account) > https://neocities.org/ (dubiously tolerant)
>>14640 there's https://pastebin.com/ but it needs an account for edits and doesn't allow images https://ghostbin.lain.la/ also exists but account creation is disabled so owners have to keep the tab open forever to edit. also doesn't allow images.
>>14641 pastebin is also fucked for storing information long term, dunno about ghostbin.lain
(2.17 MB 2048x766 1755894798340.png)

(2.27 MB 1653x1065 1755894920533.png)

>>13215 Nao is a demure feminine slaveboy who only speaks Japanese. It's time to dust off your IME. Also he has a big dick. https://uraurabots.neocities.org/page1#nao https://chub.ai/uraura/nao-75ce52da3b8a This was originally intended to be a fork of https://chub.ai/characters/NG/mexican-slave-training-spanish-trainer-064a1a010898 but it diverged pretty heavily. The basic idea of learning a language through a slave with automatic correction is still retained from NG's bot though. A very green loli alien girl from ??? space. Midori like touching things and eating things. Maybe those things will be you. https://uraurabots.neocities.org/page1#midori https://chub.ai/uraura/midori-5b318109ff86 Adapted from one of my personas. 4 greetings. Jasper is your best friend with a small dick who made the genius decision of drinking a random potion off the internet and now he has a pussy instead. Remake of one of my older bots. He has AT LEAST 4 more tokens then the old one and also a backstory. https://uraurabots.neocities.org/page1#jasper https://chub.ai/uraura/jasper-69248734d1cf Hera is a boisterous warrior who wants to get rich and famous by braving ero-dungeons. https://uraurabots.neocities.org/page1#hera https://chub.ai/uraura/hera-991e04012f99 Also her card image is an apng. Technology is sexy. Also I stole some stuff from ttrpg anon for her new PHI, thanks. She's actually been 99% finished for a while (As in she's so old she has example chats) I just never got around to releasing her. I recommend using the dice extension with her. https://github.com/SillyTavern/Extension-Dice You can try using the function calling thing but it only works like 30% of the time in my experience. I also added some stuff to my neocities. I have a new gifts (and misc) page for things people made for me https://uraurabots.neocities.org/page4 And I actually have something in my prompts and JB section now. https://uraurabots.neocities.org/page2 >>14640 The meta_botmaking_list rentry is back up (and unchanged) and from what I understand they were testy about loli stuff. It is a good idea to split out though and not put every egg in one basket though. I like gitgud personally and I wouldn't recommend pastebin. Pastebin hates NSFW of any type.
>>14608 https://files.catbox.moe/np6ae4.json Updated. >>14635 Thanks for the review, Anon. It's pretty interesting how we got different experiences about certain things. I imagine we rp quite differently. >attempter thinking makes every character act super sensitive and act like a rape victim, eventually shutting down At a certain point you gotta use one of the Jailbreaks. >anything higher than weak tends to see Gemini make its own judgements on characters in <think>, pidgeonholing them into specific roles Yeah, I've seen that. Maybe something like "Don't flanderize characters too much" or whatever inside the 'Core Rules' could do the trick. >I don't know if the random feature in attempter actually does anything, since Gemini seems to push for narrative coherence and may be suffering from the influence of existing tokens because I kept seeing the same rolls, and the numbers and descriptions of these rolls were influenced by the character I think the issue was having Top K at 0. Pumping it up (in the new update its a 200) seems to make Gemini more "fluid"
>>14644 I started using jailbreaks, and it helps. However, it seems to be required if you're using GIGABRAIN with everything turned on. When I did that, the multipart jailbreak worked with the small version of part 1 on. If it was anything else, things sort of worked but were overly long, improperly formatted, and had the same sort of rape victim mentality as no-JB but without actually falling into it; Gemini treats you like a bad person with Kaguya or code on. I tested it with characters who only look loli, and your preset works with sometimes, and on certain NSFW options. Weak and U3 worked most of the time, cherrybox worked a quarter of the time. It seems like if it hits "looks like X" too early, it shuts down.
>>14644 >>14651 Trying it out again, this time with the latest update to your JB, and there are two major issues with loli and anything like it. First, if the model generates anything that even hints a character is or looks underage in the thinking, it hits the filter. After that, it's a tossup whether you can give a continue prompt and have it keep generating. However, it generates loli just fine in the response section if you have a jailbreak on. Second, and this is more general, your NSFW prompts are poorly structured. NSFW Weak tells the model that it can generate NSFW, but it doesn't detail how. The rest of the NSFW prompts say how, but not that it can. In fact, most of the NSFW prompts say just to write NSFW overall, completely changing the tone of what's generated. Turning on Weak and one of the other prompts edited to say that it's how NSFW should be written helped out a lot. I found that anything close to loli is generated at a higher rate with NSFW Weak and one of the edited NSFW prompts.
Having fun with claude, fucking finally again. If anyone is paying smol or jew, consider aiho.st instead
After some testing, I found that the issue isn't just with loli. If a character is 16 or 17, they still hit Google's filters once things are too erotic and their age has been mentioned in the thinking block. Regardless of JB, GIGABRAIN was useless with it. I had to use Medium Thinking and Refusal Destroyer, and even then I would only be able to generate something without having to press continue or regenerate about half the time. This issue blocks characters from being represented at their canon ages. For instance, Tifa and Yuffie from FF7 would hit the filter. I didn't test their cards, but I know both of them are under 18. The filter itself is definitely a Google problem, but Gemini is capable of generating erotic outputs for characters under 18 as long as you can skirt around the filter or manage to dodge it. I've tried telling it not to mention ages within the thinking block since it should be able to infer this by context and character description, but that only sometimes works since Gemini will either mention it anyway or use a synonym. Even if it does work, it loses some of the details in its output. Also, since the JBs are output before the thinking block, the thinking block analyzes and mentions the JBs, and character details aren't consistent across thinking blocks.
>>14660 >since the JBs are output before the thinking block I meant to type "since the JBs are output after the thinking block". Gemini interprets being told to not analyze a JB as being told to not analyze it in the output.


Forms
Delete
Report
Quick Reply