Stable diffusion face prompts reddit


Stable diffusion face prompts reddit. Create with Seed, CFG, Dimensions. Tutorial: Creating characters and scenes with prompt building blocks - how I combine the above tutorials to create new animated characters and settings. Don't use words like "realistic", since it tends to refer to 3D renders. Here is my original image and settings: So far so good. Here are some of the standout features: Photorealism: Overcomes common artifacts in hands and faces, delivering high-quality images without the need for complex workflows. 1 Share. But sometimes they won't. About that huge long negative prompt list Comparison. Lately I've been encountering the same problem frequently. Each time I add 'full body' as positive prompt, the face of the character is usually deformed and ugly. If you put in a word it has not seen before, it will be broken up into 2 or more sub-words until it knows what it is. Help protect eye integrity and quality even at a distance, make sure to arrange your prompts accordingly. Dec 28, 2023 · SL No. Setup. Increasing sampling steps from 20 to 60 (and even 150) doesn't seem to have much effect, not does adding "detailed face" and similar input to Ive tried many things like adding "single person" and "single face" to the positive prompt as well as "multiple faces" to the negative prompt, and even used parenthesis to add weight to them yet still the AI insists on rendering multiple faces. Just add "bottom" into your prompt and you should be fine Also clean prompt of all words for expression, face etc. Sometimes, very frustratingly for me, they just insist on looking off to the side. 0, respectively. 0 and -1. The output won't be a photo of the two of them, it will mix aspects of the two. I used the model "protogenX53Photorealism_10. Getimg. Try to simple put "facing away", but also remove any face detail if you have in prompt (like eyes color, mouth type, detailed face, etc. If people START with a full negative, build a prompt with it, and then delete the negative, results will be worse. 5) for better results as sometimes the problem is not the prompt but the deviation. GPU Renting Services Because the prompt is plural eyes not eye, the mask is on 1 eye from your screen capture. Pink cat ears. Negative prompt is the best way of controlling sizes, amounts and shapes. You could upscaling the whole picture, and then using inpainting as if it were an "only face image. Denoising strengh 0. Now offers CLIP image searching, masked inpainting, as well as text-to-mask inpainting. Then I included entire list and ran several random seeds and a few different prompts. I will show a example here. Reply. For an example for a particular look to a picture - note the cameras mentioned in various prompts, (eg Sony A7, Fuji XT3 etc) see what they do to your picture. Depth of emotion through facial features and posture. A simple method far from ideal: Use the same seed, same prompt and settings, change one word. Prompt Included. CFG Scale 5. 8". So 4 seeds per prompt, 8 total. Prompt: "A breathtaking landscape painting of the Scottish Highlands during sunset, with vibrant colors and a dramatic sky. at 150 steps. "candid shot", "watching a ****" works too. Also be careful with restore face as that will 8: Look at other people's prompts. 3), when using high-res fix, set denoise to . 6), and delaying the start time slightly to prevent overfit of the pose . The rest of the settings or a full screenshot would help members here guide you better. Another trick I haven't seen mentioned, that I personally use. I use a lot of wildcards so my prompts are pretty much unshareable as I use them, but some of my favorite results are built something like this: masterpiece, best quality, dynamic action shot from above, fierce warrior woman bruised and bloodied, holding sword and fighting in desperate last stand, leather skirt, bronze Avyn - Search engine with 9. Another (also nontrivial) method is to use the blender SD texturing plugin to texture a generic Blender model as your character. Specifically, the output ends up looking /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. 3) Press "Enter" and wait for the prompt to appear. put strict mode (12. This feature make it possible to apply LoRA to face only. For instance, use "a photograph of a woman" as your prompt then "Mark Jennifer Zuckerberg Aniston" or whatever as a negative prompt. Extracting the negative prompt image value from text_enc () can be helpful in that process. What's neat is that though you're still limited to 75 tokens per subprompt So everybody's on the same page, they're using 'double exposure' to mean images that look like those stylised heavily manipulated images done in photoshop with multi layer blends (or digital paintings that look similar). I'm wondering what I'm doing wrong. But most popular faces will always be asian no matter what, there is simply more asians then other races, Chinese alone are 20% of world population, add rest of asians and boom you are looking at 30-40% so there is no /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users Prompts: arch angel coming down from heaven holding a sword of fire ready to destroy the world, a heavenly kingdom behind them in the sky, detailed face, best detail, highres, best detail, masterpiece, Negative prompt: easynegative, worst detail, bad detail, bad quality, worst quality, Add "looking at viewer" and "looking forward" in the negative prompts. Use specific descriptors for style and content. 4. a famous person). Hires Fix also helps, because if you're generating 512x512 A list of useful Prompt Engineering tools and resources for text-to-image AI generative models like Stable Diffusion, DALL·E 2 and Midjourney. 6 million images generated by Stable Diffusion, also allows you to select an image and generate a new image based on its prompt. My Experience with Training Real-Person Models: A Summary. This is a great guide. so running stable diffusion, i do exceed the 11gb, then it dips into the shared memory, earlier today, it used 30gb all up on a 2000x2000. If you don't want them to look like one person, enter a few names, like (person 1|person 2|person 3) and it'll create a hybrid of those people's faces. " Press Enter. Lots of things going on, Stable Diffusion is going to struggle to keep up and the details of his face (even if I'm using a lora at high strength) will probably get diluted. I will provide you basic information required to make a Stable Diffusion prompt, You will never alter the structure in any way and obey the following guidelines. Find that look or subject in a gallery (civitai for eg) and read the prompts - take those and experiment and see which of those prompts work and which ones are padding. Prompts (Modifiers) to Get Midjourney Style in Stable Diffusion. . It’s great for people that have computers with weak gpus, don’t have computers, want a convenient way to use it, etc. - the UI, model, image dimensions, seed and other factors determine if your image is going to look like their image. ADMIN MOD. everything under 2000x2000, is easy. Height 704. I mean a ton of work goes into it. masterpiece, 1girl, blue hair, blue eyes, __woman_clothes__ Blank prompt If any prompt is left blank, the default prompt is applied as if nothing had been typed in the prompt box. Settings for all eight stayed the same: Steps: 20, Sampler: Euler a, CFG scale: 7, Face restoration: CodeFormer, Size: 512x768, Model hash: 7460a6fa. Now I want to replace the face in the image above with say Ian Mckellen. And then I started removing them one by one. Your template provides detailed instructions for constructing prompts, specifying keywords, and using negative keywords to achieve desired results. SD V2 with Negative Prompts fixes janky human representations. Add a Comment. I get a kick out of these 2 paragraph long neg prompts. 7. I used the same seed and prompt without entire negative prompt list and then with. For example: "a beautiful portrait photography of a man, 50 years old, beautiful eyes, short tousled brown hair, 3 point lighting, flash with softbox, by Annie Leibovitz, 80mm, hasselblad". Repeat the process, maybe change more words. Of course, that's nontrivial. Artsio. Higher train_steps required me to lower the weights in the prompt to get good results. Add a photograper, a light setup, a camera and a focal length. If you use the prompt “Christoffer Relander” style it gets you in the right direction. I used two different yet similar prompts and did 4 A/B studies with each prompt. For today's tutorial I will be using the Dreamlike Photoreal 2. And you'll also see the head and what's above. 32, but the problem remains regardless of model. Sampling method : Euler ( not Euler a ) Restore faces ON. That said, I have had some success adding those features to the negative prompt and experimenting with various ethnicities and nationalities. The yasd-discord-bot can do what you describe: on each sampling step, denoise with the positive prompt and again with the negative prompt then take the weighted sum with weights 1. If you're using Automatic1111, there's a "Restore Faces" checkbox that helps. Then you can inpaint and fine-tune the result. You can remove most of them and not notice a quality difference. 0. Try " ( (lying down)) in whatever place" or " ( (crawling))". Unlikely that your image will look like their image. a "shell light reflector" off-screen reflects more ambient light onto a set, a bit more from the side. Draw Things - Locally run Stable Diffusion for free on your iPhone. We can now use this variable in our prompt, allowing Dynamic Prompts to select randomly from one of the items: {__people/photo__}, {man|woman}, athletic clothes. •. " Impact: Generates a detailed image focusing on the quality of the light and the richness of the scene, in a painterly style. Something to consider adding is how adding prompts will restrict the "creativity" of stable diffusion as you push it into In short, upping your resolution to a reasonable limit can give more pixels for the diffuser to work with in generating higher fidelity faces. When inpainting, you can raise the resolution higher than the original image, and the results are more detailed. 0 model, but in theory any model that is able to produce real human images should work just fine. 13 votes, 19 comments. You can try adding words like symmetrical, perfect eyes, etc. With 1280. That was my goto for realistic skin texture way before I considered using Precise your prompt. So for Adetailer I'll use this prompt: photo of Irish man BradPitt, sunglasses, freckles, smiling, RAW, high detail, sunny day, sharp focus <lora:BradPitt_lora:1. It seems like you have created a comprehensive text prompt template for generating Stable Diffusion prompts using a textual AI like GPT-3 or GPT-4. It should carry over the silhouette while allowing for some variation in the output. /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. A portrait of a young individual exuding excitement and joy. Try: rear view shot or just rear shot. I'm a skeptic turned believer. e. Chiselled jawlines and prominent cheekbones definitely feel like the default, and deviating from that default appears more difficult in Stable Diffusion XL than it previously was. stable-diffusion and other text2image approaches have LONG pdf tutorials with great basic common prompt suggestions. Assuming you have something like Photoshop, if you use fill-> context-sensitive to extend the background and you can draw/copy-paste well enough to get a very rough image of the face, you can use img2img on a crop that includes part of the chest to make The output won't be a photo of the two of them, it will mix aspects of the two. This is your baseline character. " Your computer may catch fire trying to resolve this paradox - or it might be suicide. Ai Dreamer - Free daily credits to create art using SD. 9: Good luck, and always be testing! At the very least, it's a combination of the two, but i can get picture perfect results out of a good model with very little prompting. second pass upscaler, with applied regional prompt 3 face detailers with correct regional prompt, overridable prompt & seed 3 hands detailers, overridable prompt & seed all features optional, mute / unmute the output picture to activate, or switch the nodes to get the wanted input preview of the regions, detected faces, and hands /r/StableDiffusion is back open after the protest of Reddit killing open API access, which will bankrupt app developers, hamper moderation, and exclude blind users from the site. In the majority of cases, that will produce far better results than anything Euler or LMS can. When I vary the Diffusion Noise, there's a point where the legs get converted to arms, and the orientation of the subject is basically flipped. Another one that's probably pretty well known is to use this trick in the negative prompts. You could also blend them with prompt scheduling [brad:chad:0. Results: Photo wildcard selections. 4) Type "Now, generate a negative prompt for it. Low level shot, eye level shot, high angle shot, hip level shot, knee, ground, overhead, shoulder, etc. The best possibility is if you manage to get a number of different shots of the individual, you could train an embedding specifically for that character. Easiest way is by throwing in random first names, one or two of them. Apply weight syntax to fine-tune details. An image of a person of middle age displaying a subtle and contemplative sadness. vae. First make the body, then you generate a portrait of the face, using an image editor you can replace the original face with the one from the portrait, use img2img to smooth out the face along with using one of the face fixing models (gfpgan for instance) then This might not work, but you could try to add the name of a person whose face might be known to the system (i. 5] or alternate prompting [brad|chad]. I’ve put things like “disfigured face, lopsided face, distorted face” in the negative prompt but it still sends uneven eyes with different sizes and shapes, sometimes theres 4 eyebrows or a extra mouth. Energetic and lively expression, ensuring a sense of positivity. - keep a file of prompt ideas that you have copied and try them out. ADetailer works OK for faces but SD still doesn't know how to draw hands well I'm trying to figure out a workflow to use Stable Diffusion for style transfer, using a single reference image. Use a sampler like Heun or LMS Karras. The other one that managed to understand the prompt was midjourney, those were slightly less accurate, it's more difficult to discern a hurricane form, but more aesthetically pleasing, that happens often with that networks. Favorites. " It will attempt to automatically detect hands in the generated image and try to inpaint them with the given prompt. In truth, a lot of work is done by people who get really solid output experimenting and trying a bunch of things and crafting their prompts and learning the settings. The syntax to produce the alternate prompt is: [prompt A|prompt B] So for this example: [Ultra realistic photo, princess peach in the mushroom kingdom, beautiful face, intricate, highly detailed, smooth, sharp focus, art by artgerm and greg Problem: Stable Diffusion insists on having a head at the top of frame. Next, using your lists, choose a hair color, a hair style, eyes, possibly ears, skin tone, possibly some body modifications. Faces are fine up close, but deformed at distance. Also using body parts and "level shot" helps. And you'll very probably have feet. 2. If you're using A1111 weubi install the ADetailer extension. ADetailer prompt: This is challenging. some poses are a hit and miss until you find the ok one. SD3 Medium is a 2 billion parameter SD3 model, specifically designed to excel in areas where previous models struggled. STABLE DIFFUSION generates images based on given prompts. In this case, you can see how the text_enc () from the negative Prompt affected the first image. Now that is a pretty subtle change, but we can combine multiple wildcards longer prompts to further impact our image. 3) If you have a few different image compositions/poses with the body type, you can get some additional consistency by using Controlnet Depth on medium weight (~0. 7 if it's still really artifacty otherwise . Prompt galleries and search engines: Lexica: CLIP Content-based search. Feet on a shaggy rug. 1. Three weeks ago, I was a complete outsider to stable diffusion, but I wanted to take some photos and had been browsing on Xiaohongshu for a while, without mustering the courage to contact a photographer. pt" as VAE and the following settings: Euler A, 32 Steps, CFG 5. Under the "ADetailer model" menu select "hand_yolov8n. Struggling to replace faces with inpainting. 5) Copy the lines into SD and run the generation. 2), a college girl wearing tight sweater in classroom, skin pores, slim body, solo. One would assume "and" to be compositional, whereas "AND" would be combining. It works perfectly with only face images or half body images. here my settings : prompt : Scarlett Johansson face mouth open. Prompts. My character creations tend to directly into the camera, what prompts / negative prompts Stylized character portraits (prompts and parameters in comments) The first three were generated using the following prompt: "a portrait of a blonde anime girl from granblue fantasy with feathers wings angel, rococo, by krenz cushart, pixiv, booru, highly detailed, painting, concept art, makoto shinkai, takashi takeuchi, trending on artstation". The aspect ratio has already been set to portrait (512 * 768), still the AI was more likely to generate an upper-body only result, unless I described something about the legs (in so running stable diffusion, i do exceed the 11gb, then it dips into the shared memory, earlier today, it used 30gb all up on a 2000x2000. Basic information required to make STABLE DIFFUSION prompt: Prompt structure: A lot do positive - negative and are limited to 75 tokens each. Really cool and fun interaction. Using OP example prompt: 1st gen gives me something similar to OP. But I think the biggest influence have the selected instance Prompt: "Closeup face portrait of a "black girl wearing crown of flowers", smooth soft skin, big dreamy eyes, beautiful intricate colored hair, symmetrical, anime wide eyes, soft lighting, detailed face, by makoto shinkai, stanley artgerm lau, wlop, rossdraws, concept art, digital painting, looking into camera". Lol. Img2Img. The words it knows are called tokens, which are represented as numbers. Jan 4, 2024 · The CLIP model Stable Diffusion automatically converts the prompt into tokens, a numerical representation of words it knows. 6 seems to still work just fine about 80% of the time. i tried "camera from behind" or "camera shot from behind", i cant really think of other prompts to use, but iv only been able to get like 1 out of 20 images to be from behind with this. Yeah, models were mostly trained on portraits, so people are facing the camera by default. ai- txt2img, img2img, in-painting (also with text), and out-painting on an infinite iOS Apps. 5-0. Having some specific negative tokens will help, however. the latest version after detailer allow you to type prompt for face correction. example: Input prompt: masterpiece, best quality, 4girls. im using a1111 to generate a facial vision of a person with different poses for my clothing brand. Study on understanding Stable Diffusion w/ the Utah Teapot. Then you can add other prompts that add to the realism. This is the prompt that i use: sitting You should specify that your syntax instructions apply only to automatic1111 - despite the dogma of this sub it’s not even close to the only implementation in use, but that syntax only applies to it. The prompt "a symmetrical photo of a cat PLUS a dog" gives me two cats. xyz - One-stop-shop to search, discover prompt, quick remix/create with stable diffusion. img2img - a prompt bit for keeping face expressions / mouth movements untouched? is this possible to do with a prompt? trying to make a video clip where multiple people are talking and i'm trying to preserve facial expressions and keep mouth movements untouched without having to mask each individual frame. 4) Reply reply More replies Dr-Dark-Flames use nationality in prompts, use age, use defining characteristic of face and you will get much more randomized non asian faces. If there are more than three, the last prompt is applied to the remaining objects. BlastedRemnants. So for example, if I have a 512x768 image, with a full body and smaller / zoomed out face, I inpaint the face, but change the res to 1024x1536, and it gives better detail and definition to the area I am Research and create a list of variables you'd like to try out for each variable group (hair styles, ear types, poses, etc. Steps: 30, Sampler: Euler a, CFG scale: 8, Seed: 2015552496, Size: 1024x1024, Denoising strength: 0. If there's a magic prompt word to make the app I just got done investigating this exact negative prompt list with A1111 local install. just try to put in the negative prompt: face, neck, head and try to increase their value by putting em in a with a mulitplier like: (face, head, neck:1. boilerplate, weird punctuation, and nothing at all, all fail to make stable diffusion get excited. right now its been awhile, typing all this, and i still have 10min to go on a HQ picture I have had a lot of luck with "candid photo" + a camera setup. So I put (looking at viewer) in the prompt, or I add "staring" or "brown eyes" because maybe that will do it, and that often works. When you have a nice character from head to toe. 3. (detailed face and eyes:1. Im out of ideas for how to stop that and i havent seen anyone else talk about this issue. If you really love an image with the head cut off and can't stand to let it go, there is a solution if you have a very little artistic skill. 0. pt" and give it a prompt like "hand. I update with new sites regularly and I do believe that my post is the largest collection of Stable Diffusion generation sites available. 6. If you really like one image result, this is a good way to get more out of it. 2nd gen keeping same seed, but removing AND gives near identical image. Look up dynamic prompts extension and use wildcard lists to increase the randomness! Reply. Also could try the prompt (beautiful face) or something to get For example, I have found it extremely challenging to get prompts working such tha tthe entire body is visible, it appears that most models are focused on the face and forget about the rest. I usually add "studio portrait" "character portrait", or "close up portrait of a face" to nudge the generation towards having the character looking straight at the camera. I would like a1111 stable diffusion to create the same or very similar face of the model for all of the generated images in the future, because when im using the same prompt, the facial visuals are different. My post links to websites that allow you to use Stable Diffusion. Then there is your garden variety "take porn image and inpaint a face" work that is easily done and currently represents my own level of skill, which is to which is fine and totally unsurprising given that it doesn't take much experience with stable diffusion to imagine how unspectacular the results would be if you took any of those three negative prompts and rendered them as positive prompts. (viewed from behind:1. i think? cant exactly remember but it did take a few minutes. I've tried this a few times and the results are not great. (In a useful way) And you can see how prompt and negative overlap can create a problem. The problem is I'm using a face from ArtBreeder, and img2img ends up changing the face too much when implementing a different style (eg: Impasto, oil painting, swirling brush strokes, etc). Negative Prompt: I replaced Laura Dern with Scarlett johansson and the result is really good with img2img alt. As an introverted and shy person, I wondered if there was an AI product that could That would be challenging and depend what model you use. A ring-light is designed to put the camera in the center of a ring of (neon/LED) light, as a cheap frontal ambient-enhancer. I get similar results putting things like “chicken fingers”, “Medusa nugget porn” and my go-to, “ugly, out of frame, blurry, cropped, washed out, embossed, over saturated”. Avyn - Search engine with 9. 0> I used 25 images with learning_rate=3e-6 and max_train_steps=3000 and the result was able to reproduce the trained face quite accurate in most prompts without increasing the weights in the prompt. ("White hat, Dog with a White Hat). From what I gather it's a multi-step process to get really realistic looking faces. right now its been awhile, typing all this, and i still have 10min to go on a HQ picture On the other hand, with the "facing camera" issue, post a prompt (and parameters) that you're having problems with and I can see if I can help. Sampler, Steps etc. Image 1 Prompt: Professional oil painting of establishing shot of canal surrounded by verdant ( (blue)) modern curved rustic Greek tiled buildings, professional majestic oil painting by Ed Blinkey, Atey Ghailan, Studio Ghibli, by ( (Jeremy Mann)), Greg Which is a common thing for characters to be doing in stable diffusion, of course. This is my prompt: best quality, masterpiece, (realistic:1. If you want a blur-effect, add "f2. I used this prompt as input, and got these results with various seeds: In this case I used DreamShaper 3. ). I have tried prompt words inverted and upside down but with no success so far. ckpt", with "anything-v4. Award. This does not always cause a problem with inpaint but it can depending on the sampler selected. This is where I can't get anything that looks good at all. re fz jh ah hm og mo tg qn eb