Presets

The presets provided by NovelAI can be severely lacking at times. So, if none of them appeal to you, here is a collection of ones created by the community.

''To avoid redundancy, it's best to remove any presets that end up in NAI proper. If you've got a preset you'd like to share, then by all means, add it to the pile.''

Best Guess 2
by Baker-Anon


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Top-P Sampling:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

Hello frens!

Baker here. For your consideration, I'd like to present:

Best Guess 2

It's nothing special, really; just Best Guess settings applied to Euterpe.

But this time around, there are NO changes to the context settings.

Also, the sampling method has been changed to the TFS setting from Basic Coherency.

Optimal for keeping the story on track when you're balls deep in degenerate fetish scenes, results may vary when used in open-ended idea generation situations.

Hopefully, there's one anon out there who this works good for besides me; this is for you <3

Turpy
by Anon


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Top-P Sampling:
 * TFS Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

"slightly modified Co-writer has been my goto"

Storyteller
by ???


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-P Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

"It's Storywriter (Sigurd's default preset), except top-p happens before temperature."

Sphinx Moth v2
by Nyks


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Nucleus Sampling:
 * Repetition Penalty Range:

Experimental - uses custom order of Top-K, Nucleus and Temperature

Reborn from its sandy pit, Sphinx rises again with all of its max randomness glory.

Sphinx Moth is now better than ever, picking out the best tokens and giving them equal chance of being chosen. Truly harnessing the creativity of high ends of Randomness, you can expect a wide array of creativity in a way that is still written with prose that makes sense!

Be ready to wrangle with this beast, for it may avoid a detail or two in place of a more creative route.

Monkey Business
by Belverk


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:
 * TFS applied first, then randomness/temperature

Model agnostic preset done using the token probabilities viewer, debug options and some Sage advice from OccultSage, finetuned for my personal preferences. Tokens that have a lower likelihood than 2% of appearing get mostly culled, while tokens in the range of 94% and above likelihood get bumped to 100%. This behavior is familiar to everyone who has used 0.992 TFS presets before, although works better after adjustments and the TFS overhaul.

With an increased randomness applied after filtering this should give users a consistent and natural, yet creative output experience. Rep penalty curve returns, biased towards my scaffolding method of using lorebook entries, with emphasis on attempting to preserve accurate output of colors.

Why is it named Monkey Business? I made the preset on a whim and tested it on the Monkey World Domination prompt, which proved very useful for testing token logprobs and filtering. Name aside, it's a serious preset and the evolution of my tweaks on Sage's coherent creativity. Currently I am using Sigurd, but I've done testing on early Euterpe, and TFS works the same way for both. Euterpe is more insistent on what tokens come next, so if you want more creativity and plot twists while keeping the same TFS, feel free to bump the randomness up.

Damn Decent TFS
by chmod007


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Nucleus Sampling:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

A generation configuration focused on a subjective model-specific sweet spot.

Generation settings calibrated using New Story defaults with No Module.

Order: top_k, top_p, tfs, temperature

The Old Familiar

 * Randomness:
 * Top-K Sampling:
 * Nucleus Sampling:
 * Repetition Penalty:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

Optimal Machine
by lion"A variant of Belverk's Optimal Whitepaper v2 that I like to use in many of my generator scenarios. Tends to work well for generating content based off of examples high up in context."


 * Randomness:
 * Top-K Sampling:
 * Nucleus Sampling:
 * Repetition Penalty:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

Fated Outcome
by Pause This Preset will always return the same output until something is changed in the Context, allowing a sense of permanence and fate within the world of your narrative.

Fun cases with this Preset include "time warping" to see how a character would have reacted if you said or did something different, and testing the effects of different token associations on the flow of a Story. Additionally, lore details and names should have a significantly higher chance of being correct.

NOTE: This Preset makes the Retry button useless while active, as the same output will always be returned until something changes.


 * Randomness:
 * Max Output Length:
 * Repetition Penalty:
 * Top-K Sampling:
 * Nucleus Sampling:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

Pussy Tentacles (Jeral V4)
by HydroStorm


 * Randomness:
 * Max Output Length:
 * Repetition Penalty:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penalty Slope:

Damn Decent TFS [Sigurd V4]
by chmod007 A generation configuration focused on a subjective model-specific sweet spot.

Generation settings calibrated using New Story defaults with No Module.

Lorebook, token, and context settings are pristine.


 * Randomness:
 * Output Length:
 * Repetition Penalty:
 * Tail-Free Sampling:
 * Repetition Penalty Range:
 * Repetition Penatly Slope:

Complex
by Orion"Been getting really good results with these settings, based off of the 'Complex Readability Grade' posted in Basileus' findings in #novelai-research. With good usage of Tone, Word Choice and maybe Author in the Author's Notes, as well as a decent amount of context for the AI to consider after starting a story, you can get some stunningly evocative prose while the story's progression remains pretty consistent. Of course, it will still need guidance from time to time, and it might require some slight adjustments to Randomness and TFS occasionally based on your preferences, but I think this is going to be my go-to for serious stories until somebody finds a setting preset that's even better than this."


 * Randomness:
 * Tail-Free Sampling:
 * Repetition Penalty:
 * Repetition Penalty Range:
 * Repetition Penalty Slope: