AI is going to be our biggest game changer

1313234363748

Comments

  • WonderlandWonderland Posts: 7,019

    I'm getting less concerned about its ability to steal art from other artists. No matter what I do, I can't get anything remotely similar to my art or likeness by putting in my name. Maybe it would for more famous people. But even inputting Elvis Presley or Jim Morrison didn't really look like them but just an essence them. But things like Wonder Woman could be pulling in from movie posters and comic books, not sure. My artist name Wonderland just brings up Alice in Wonderland stuff but I'm wondering what other artists names that are more specific would produce. 

  • vonHobovonHobo Posts: 1,692
    edited February 2023

    Wonderland said:

    von Hobo said:

    Wonderland said:

    So my computer was in for repair and I wanted to make a sexy Valentine's Day card for my boyfriend so I tried Midjourney on my iPad. It took like 100  tries with refinements to have something decent (some had 3 legs and worse!) it's especially hard to get a decent face AND body. But I got it close enough then brought it into Snapseed, and then some selfie altering apps that I discovered works great for my purposes, you can even change makeup and eyebrows, reshape face, eyes, lips! It's really a better art tool in many ways than photoshop! I ended up spending as much time on this as I would have on a render plus post on the computer and I really like how it finally came out. Is it cheating? Well yeah, but I still feel like I put in so much creative input and time that's it's more based on AI rather than pure AI and really when we use PA clothes, hair, morphs and skin, is that cheating any less? I always do post so no matter what I'm doing I put a lot of creative input into it, but I realize that using AI as a base can actually lead to a creative image that's your own if you add a lot of your own changes to it, especially in post.

    Will it create the same character consistently in a bunch of different renders? Or does it make something totally different for the character every time, even if you want the same looking character?

    I was thinking this could be a fun and easy way to create illustrations for a children's book, but only if the program is consistent with the same characters in different settings. If it's all random, I guess it would not work.


    No it will not produce the same character consistently. I found it's hard to get a decent head and body together. You can get a beautiful face but then it has 3 legs and 8 fingers or parts of the clothes are weird. It's pretty hard to get a usable image without major postwork. It can work well as a base but you need to have certain Photoshop skills or can even use various phone or iPad apps but most images have issues as is if you look closely. Often expressions are weird, eyes looking in different directions, artifacts everywhere. It is not completed art. But if you need a quick base and are good with postwork, it can be interesting. If you have drawing skills you can definitely put your own spin on it and fix up all the errors. Many people use stock photos commercially, at least these are more original and require more talent and creativity to make look good. 

    Ok. Thanks. Sounds pretty interesting but doesn't sound like it would work for narrative work. 

     

    Post edited by vonHobo on
  • N-RArtsN-RArts Posts: 1,518

    Sorry. But AI's still have a way to go if they can't even generate a man-eating chocolate chip cookie *rolls eyes*

     

  • SnowSultanSnowSultan Posts: 3,630
    edited February 2023

    I'm getting less concerned about its ability to steal art from other artists. No matter what I do, I can't get anything remotely similar to my art or likeness by putting in my name.

    We are not nearly famous enough for that to happen. Even when your prompt asks the AI to copy an artist's style, like Artgerm or Sakimichan, it rarely looks THAT similar. It can copy things like anime series and celebrity likenesses very well, but it wouldn't know you or me from Adam.

     

    Sorry. But AI's still have a way to go if they can't even generate a man-eating chocolate chip cookie *rolls eyes*'

    Ask Midjourney to make that, it probably will do a very good job.

    Post edited by SnowSultan on
  • N-RArts said:

    Sorry. But AI's still have a way to go if they can't even generate a man-eating chocolate chip cookie *rolls eyes*

     

    that would be a interesting challenge... but even conventional intelligence may misunderstand your instruction.

  • wolf359wolf359 Posts: 3,834
    edited February 2023

    No it will not produce the same character consistently. 

    Not unless you train it with your provided images
    on a local SD install on your own
    GPU . 

    Post edited by wolf359 on
  • SnowSultanSnowSultan Posts: 3,630
    edited February 2023

    SolitarySandpiper said:

    N-RArts said:

    Sorry. But AI's still have a way to go if they can't even generate a man-eating chocolate chip cookie *rolls eyes*

    that would be a interesting challenge... but even conventional intelligence may misunderstand your instruction.

     

    Will you settle for a cop and his giant cookie partner going undercover? (I know they're low quality, I made them in about 30 seconds)

     

    Wolf, have you trained any pure 3D characters? I'm very curious to know if it's possible to get non-3D results using a LoRA or checkpoint that was trained exclusively on 3D renders.

     

    3587254037-a giant cookie chasing a man through city streets, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp.png
    640 x 640 - 647K
    3587254039-a giant cookie chasing a man through city streets, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp.png
    640 x 640 - 771K
    2157309109-a giant cookie chasing a man through city streets, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp.png
    640 x 640 - 714K
    Post edited by SnowSultan on
  • SnowSultanSnowSultan Posts: 3,630

    oops, forgot one. "Cookie Crooks" (1983) starring Alec Baldwin and Joe Pesci

    3587254035-a giant cookie chasing a man through city streets, trending on ArtStation, trending on CGSociety, Intricate, High Detail, Sharp.png
    640 x 640 - 678K
  • SolitarySandpiperSolitarySandpiper Posts: 566
    edited February 2023

    SnowSultan said:

     

    Wolf, have you trained any pure 3D characters? I'm very curious to know if it's possible to get non-3D results using a LoRA or checkpoint that was trained exclusively on 3D renders.

     

    Theoretically Daz should provide an ideal environment to set up optimal images for LoRA training... might be a worthwhile endeavour.

    combine that with the new controlNet extension and already you're putting together a formidable suite of programs that can tackle a wide range of subjects.

    Post edited by SolitarySandpiper on
  • SolitarySandpiperSolitarySandpiper Posts: 566
    edited February 2023

    N-RArts said:

    Sorry. But AI's still have a way to go if they can't even generate a man-eating chocolate chip cookie *rolls eyes*

     

    i know its hideous but it's just a bit of fun... i aint got much time to spend on other peoples whims.  

    man-eating cookie.png
    1536 x 1024 - 2M
    Post edited by SolitarySandpiper on
  • SnowSultanSnowSultan Posts: 3,630

    That is amazing, is it Midjourney? People have come up with some methods for getting fairly consistent characters in Midjourney (there are videos on Youtube), but you don't have as much direct control as you do in Stable Diffusion now.

  • SnowSultan said:

    That is amazing, is it Midjourney? People have come up with some methods for getting fairly consistent characters in Midjourney (there are videos on Youtube), but you don't have as much direct control as you do in Stable Diffusion now.

    Yes it's MJ... you can't beat it for quick n easy but i much prefer SD (or a combo) 

  • SolitarySandpiperSolitarySandpiper Posts: 566
    edited February 2023

    One of my favorite combo results

     

    00048-1775181010.png
    792 x 792 - 728K
    Post edited by SolitarySandpiper on
  • SnowSultanSnowSultan Posts: 3,630

    Nice, thanks. I do think Midjourney gives better results out of the box and with less fiddling than SD, but there's more creative potential (and potential for using 3D renders) with Stable Diffusion.

  • SnowSultan said:

    Nice, thanks. I do think Midjourney gives better results out of the box and with less fiddling than SD, but there's more creative potential (and potential for using 3D renders) with Stable Diffusion.

    i think the fiddling will become the most rewarding part of the process... the results i may or may not care for.

    But i'l always get satisfaction from the process. 

  • SnowSnow Posts: 95
    edited March 2023

    deleted

    Post edited by Snow on
  • So you guys had 100 tries at getting a decent image out of these things, and one of them looked pretty good.  I have a 1st Class Hons in AI but from the 90's (not pulling rank, I just thought about these things maybe more than most people have) i.e. before the "Deep Learning" breakthrough.  I thought back then and still think today  that the hype this stuff generates with our already primed imaginations (primed by science fiction) is far, far greater than the practical reality.  If you could show me where these things are useful or likely to be?  You type in some words and it produces something with the same statistical properties as the 2 million tagged images you trained it on.  You can't actually ask it what those properties are (that would at least be interesting).

    There are some very focused cases I've seen where it seems pretty good and useful, like Cascadeur physics, protein folding and Flowframes video upscaling.  But in the general sense I don't think anyone's going to be out of a job in the creative industries any time soon.  The lack of control is the main problem here.  You give it a word soup and it returns an image soup.  That's about it!

  • bluejauntebluejaunte Posts: 1,909

    Pickle Renderer said:

    So you guys had 100 tries at getting a decent image out of these things, and one of them looked pretty good.  I have a 1st Class Hons in AI but from the 90's (not pulling rank, I just thought about these things maybe more than most people have) i.e. before the "Deep Learning" breakthrough.  I thought back then and still think today  that the hype this stuff generates with our already primed imaginations (primed by science fiction) is far, far greater than the practical reality.  If you could show me where these things are useful or likely to be?  You type in some words and it produces something with the same statistical properties as the 2 million tagged images you trained it on.  You can't actually ask it what those properties are (that would at least be interesting).

    There are some very focused cases I've seen where it seems pretty good and useful, like Cascadeur physics, protein folding and Flowframes video upscaling.  But in the general sense I don't think anyone's going to be out of a job in the creative industries any time soon.  The lack of control is the main problem here.  You give it a word soup and it returns an image soup.  That's about it!

    We're also at the very beginning of this. Use your imagination where this will go in the decades ahead.
  • generalgameplayinggeneralgameplaying Posts: 517
    edited February 2023

    bluejaunte said:

    Pickle Renderer said:

    So you guys had 100 tries at getting a decent image out of these things, and one of them looked pretty good.  I have a 1st Class Hons in AI but from the 90's (not pulling rank, I just thought about these things maybe more than most people have) i.e. before the "Deep Learning" breakthrough.  I thought back then and still think today  that the hype this stuff generates with our already primed imaginations (primed by science fiction) is far, far greater than the practical reality.  If you could show me where these things are useful or likely to be?  You type in some words and it produces something with the same statistical properties as the 2 million tagged images you trained it on.  You can't actually ask it what those properties are (that would at least be interesting).

    There are some very focused cases I've seen where it seems pretty_ good and useful, like Cascadeur physics, protein folding and Flowframes video upscaling.  But in the general sense I don't think anyone's going to be out of a job in the creative industries any time soon.  The lack of control is the main problem here.  You give it a word soup and it returns an image soup.  That's about it!

    We're also at the very beginning of this. Use your imagination where this will go in the decades ahead.

    I'd rather say "where we would go", which means both that humans design those systems, as well as humans decide if it's accepted at all. In addition, the purely generative systems won't get more magic and more precise at the same time, technically. It doesn't progress linearly, exponentially neither.  It's not certain that we will be that much better within 10 years. Not saying such can't happen, but it's clearly not certain.

    Concerning precision... a way more complex system, consisting of different types of components, including different types of ais, may make more precise posing possible for the generator. That's light years off the current implementations, i would say. Can't fully foretell the speed they're moving at either, though.

    Actual advancements i could imagine, would be using generative tools (or just algorithms, like daz3d scripting) to create images, that you can be pretty sure of, that they're tagged correctly, and an improved learning+detection system then is capable to run detection much better than what you get by "ordinary" supervision (so far). That's not yet the generator. Maybe it'll somehow be possible to use the inpaininting abilities and design systems, one of which poses something like the shapes or say wireframes rather, and the other doing texturing and effects, which when combined with a generative system with the actually artwork-wise precious training data, will then interpret stuff in a feedback loop (somewhat) to actually create stuff with precision and taste and sugar (and so on...). We're back at what i wrote above, then... "currently sci-fi", "maybe in range within decades".

     

    (Concerning accountability: With the current lawsuits no one who trained on images against consent of rights owners wants accountability. Competition might, though.)

    (With very good detection, there could be some shortcuts to some quality, e.g. if you can make different ais for "objects" as well as background and postwork/rather - that would allow for a relatively dumb composition. But those kind of systems already would be more complicated to build, and likely clearly more expensive to train. Plus, of course, the yet-unknown-to-me/what-i-forgot-to-mention...)

    Post edited by generalgameplaying on
  • SnowSultanSnowSultan Posts: 3,630

    So you guys had 100 tries at getting a decent image out of these things, and one of them looked pretty good. I have a 1st Class Hons in AI but from the 90's (not pulling rank, I just thought about these things maybe more than most people have) 

    Eh, I've got about a hundred very good tries saved in a folder, so I'm not complaining.

     

    The lack of control is the main problem here.  You give it a word soup and it returns an image soup.  That's about it!

    I don't think you've paid any attention to AI since the 90s either then. The addition of ControlNet to Stable Diffusion basically solved character posing and composition issues overnight. 

  • WonderlandWonderland Posts: 7,019

    Snow said:

    Wonderland said:

    So my computer was in for repair and I wanted to make a sexy Valentine's Day card for my boyfriend so I tried Midjourney on my iPad. ..


     

    Lucky boyfriend!

    Lol, I can't get jealous of my own art, can I? 

  • WonderlandWonderland Posts: 7,019

    SnowSultan said:

    I'm getting less concerned about its ability to steal art from other artists. No matter what I do, I can't get anything remotely similar to my art or likeness by putting in my name.

    We are not nearly famous enough for that to happen. Even when your prompt asks the AI to copy an artist's style, like Artgerm or Sakimichan, it rarely looks THAT similar. It can copy things like anime series and celebrity likenesses very well, but it wouldn't know you or me from Adam.

    If you are anywhere on the internet it should find you. If you put your name in the title of your art before you post it to major sites, it should find you. Somehow it found my watermark and put a mixed version of that on generated art and I could tell it was originally my watermark but thankfully, not my art. Somehow it can read print imbedded in images. Out of curiosity, I tried putting in Mousso, Daz3d, but I just got a weird mix of ad copy with illegible print on it with some interesting characters that didn't look anything like Mousso's.
     

    Another time, I tried to put in Shag the artist's style in there and it almost banned me lol, so I put in his actual name Josh Agle and the style was fairly similar but no where near as good and very convoluted. I think it may be trained to alter everything so much that it's only reminiscent of a certain famous artist or celebrity. I wasn't using the prompt photoreal, but the celebrities looked more like an expressive artistic depiction of them and some were barely recognizable. Elvis Presley's hair and leather jacket had more likeness to Elvis Presley than his face. But I guess if you kept trying you could get closer.  It might be programmed to avoid lawsuits lol. Out of curiosity, I'm going to try the photoreal prompt with celebrities and see how close I can get. If it's too real, that could definitely be a problem for them!

  • WonderlandWonderland Posts: 7,019
    edited February 2023

    Well, I don't think celebrities have to worry. The hair may be right but not much else. 
     

    682BA220-9DDE-4CE6-BEA3-69D4DFE9430D.png
    1024 x 1536 - 2M
    2613B534-4498-45F9-876E-252DEB857D3C.png
    1024 x 1536 - 2M
    Post edited by Wonderland on
  • generalgameplayinggeneralgameplaying Posts: 517
    edited February 2023

    Wonderland said:

    It might be programmed to avoid lawsuits lol. Out of curiosity, 

    Maybe. The most sensible thing would be to not train with images against the consent of the rights owners in the firs place.

    More complicated, slower, more expensive - yes, but societally more sane. Could be from the cookbook of modernist tech people: Simple, fast, for free.

    Other's call it the devil's path, but ok ~ i'm just passing through...

    Technically, it's encyclopedic by nature, as it trains to reproduce source images as close as possible. Theoretically resulting in "technically no style", because it learns all styles (in a way). Of course it's not magic, but it comes close to both the statistical part due to tagging and model size, and the encyclopedic part, due to being trained to reproduce as closely as possible (with the model size, source image resolution, and so on). So to me it's no wonder that you can actually get such and similar results. A sufficiently large model with perfect tagging might perfectly reproduce, if you train it on high resolution images. From there on it would be a fundamental decision about at which numbers it's safe to assume "fair use" - we've had distantly similar questions about thumbnails in some context.

    Post edited by generalgameplaying on
  • PixelSploitingPixelSploiting Posts: 898
    edited February 2023

    TBH face recognition is the hard part since the beginning of the AI mostly because even with humans modern science is not capable of giving a realy complete answer what makes us so good at recognizing faces, so it'd be unfair to expect better. Lots of human artists fail at making portraits too.

    Post edited by PixelSploiting on
  • marblemarble Posts: 7,500

    Pickle Renderer said:

    So you guys had 100 tries at getting a decent image out of these things, and one of them looked pretty good.  I have a 1st Class Hons in AI but from the 90's (not pulling rank, I just thought about these things maybe more than most people have) i.e. before the "Deep Learning" breakthrough.  I thought back then and still think today  that the hype this stuff generates with our already primed imaginations (primed by science fiction) is far, far greater than the practical reality.  If you could show me where these things are useful or likely to be?  You type in some words and it produces something with the same statistical properties as the 2 million tagged images you trained it on.  You can't actually ask it what those properties are (that would at least be interesting).

    There are some very focused cases I've seen where it seems pretty good and useful, like Cascadeur physics, protein folding and Flowframes video upscaling.  But in the general sense I don't think anyone's going to be out of a job in the creative industries any time soon.  The lack of control is the main problem here.  You give it a word soup and it returns an image soup.  That's about it!

     

    This might be of interest here:

  • bluejaunte said:

    We're also at the very beginning of this. Use your imagination where this will go in the decades ahead.

    I'm no spring chicken so I have to say I've heard this before with other technologies, more than once.  Nothing came of them.  Worse perhaps the more you increase the complexity, the less control you have over them.  They're pretty opaque already.  You wouldn't put a black box in charge of something important would you, or make your studio's £50m investment dependent upon it.

  • JabbaJabba Posts: 1,460
    edited February 2023

    I think the limitations will be centred more around legal restrictions than creative ones... If the Stable Diffusion database is determined to be transformitive instead of derivative, then that will open the door for all sorts of applications to use it - we could even see a future where you could create a 3D scene as normal and then select an art style for your renders - and the same thing is happening with music and writing as we speak.

    Post edited by Jabba on
  • I quite like the ControlNet OpenPose

    shame batch image won't work for me with it for making videos

    00031-3876060701.png
    1856 x 1024 - 3M
    00039-350641213.png
    1856 x 1024 - 2M
    00040-3823326259.png
    1856 x 1024 - 3M
    00043-3637433478.png
    1856 x 1024 - 2M
  • generalgameplayinggeneralgameplaying Posts: 517
    edited February 2023

    Jabba said:

    I think the limitations will be centred more around legal restrictions than creative ones... If the Stable Diffusion database is determined to be transformitive instead of derivative, then that will open the door for all sorts of applications to use it - we could even see a future where you could create a 3D scene as normal and then select an art style for your renders - and the same thing is happening with music and writing as we speak.

    There are creative restrictions too, also with transformer technology. The point is YOUR creativity with using such tools, though. So if the tool enhances that, it'll be another game.

    Moving "the database" to transformative... not sure what you're pointing at there. I guess creativity fed from statistics and past works is more a miss than a hit ~ as i said... up to YOUR creativity. Of course i can imagine a combination of different kinds of tools, streamlined for creating swift (mock-up) scenes, using precise and generative tools alongside, up to whatever quality. However i still assume, that you can't have arbitrary amounts of magic and precision at the same time - which doesn't contradict it possibly becoming a killer tool in the hands of a creative person.

    Concerning creativity... we must be very careful, i've seen few different kinds of takes on that (noteworthy for me):

    1. Human creativity isn't so huge either. It essentially does the same. (I call this romanticism, as humans are not actually static, even if their evolving isn't as great as in some of the fairy tales. On the other hand, there are some concepts for manipulation of people, like "framing" or "setting the stage", e.g. verbally, and humans will in average act slightly differently, similar to a prepped prompt. But that's still not touching the fundamental difference i intend to point at.)

    2. Imagine supervised learning, which is done for all those systems for a part, imagine you're the training part of the ai, and you're asking the human supervisor all the time "Is this creative?"

    3. The tools are not really creative by themselves. It all depends on whomever is using the tool. However they have the potential to help you break barriers and take repetitive routine work off your shoulders. That would be a game changer, already. (Note that different people may have different kinds of games in mind.)

    4. (Independently: risks with tools that minify the amount of work ~ floods of still irrelevant and not really creative content, which looks as if it could be something, spam, manipulation. Question is if there will be balancing effects, like people not wanting to pay for something they get tired of, or detection of generated content working very well, for instance.)

     

    We're closer (to some things) and very far (from some/other things) at the same time. Like with secret services, you never know what they have in the basement by now, but societally, we obviously can't rely on such untold secrets (neither scientifically).

    Post edited by generalgameplaying on
This discussion has been closed.