• úvod
  • témata
  • události
  • tržiště
  • diskuze
  • nástěnka
  • přihlásit
    registrace
    ztracené heslo?
    KERRAYAI Art - Tvorba za pomocí umělé inteligence: Midjourney, DALL·E 2, Stable Diffusion, OpenAI
    Vše ohledně umění tvořeného pomocí umělé inteligence - obrázky, 'fotky', galerie, hudba, video, text + články, novinky apod.

    NSFW obsah prosím obalit spoiler tagem - <div class="spoiler">obrázek</spoiler>, a nemá to tu být klub na roštěnky a nahotinky bez nějaké přidané hodnoty

    Příbuzné diskuze:
    - [DALL·E mini i Craiyon - having sex with AI since [date format unknown]]
    - [I Hope This Does Not Exist ​ ​ ​ ▌​ ​ Vedlejší efekty v AI visuálech]
    - AI obecně [Artificial Intelligence AI]
    - Vtipy [Umělá inteligence, chatboti - vtipné konverzace aneb "Hoří hovno?"]
    - [generativní modely] Jak konstruovat prompty, kde získat váhy i jak to vše interpretovat


    Prosím zkusme pro vkládanou tvorbu používat tagy
    #galerie (2-3 obrázky na ukázku, další po rozkliku) #obrázek #video #hudba #text #hry #původní (pro vlastní tvorbu) #roštěnky

    #článek #nástroj

    (návrhy na další tagy apod. vítány)

    Texty, programování: https://beta.openai.com/playground | https://chat.openai.com/
    Obrázky online: https://www.midjourney.com/ | https://beta.dreamstudio.ai/
    Lokálně: https://github.com/AUTOMATIC1111/stable-diffusion-webui | https://github.com/invoke-ai/InvokeAI | Civitai repository custom modelů pro SD
    AI na vytvoření textového zadání z existujícího obrázku: https://huggingface.co/spaces/pharma/CLIP-Interrogator
    rozbalit záhlaví
    GALADAR
    GALADAR --- ---
    HONZA09: Když narážíš na content restrictions, model tě upozorní. (a rád i sám poradí, jak je obejít) Tohle je jen chyba v promptu. Mmch "angel dust" už je direct drug reference, dívím se, že tě nestopl. Doporučuji flour, white powder, powdered caffeine atd.
    HONZA09
    HONZA09 --- ---
    Tak jsem našel omezení. nejsem za boha schopnej v Dall E vygenerovat obrázek boha dávajícího si lajnu. Ten generativní model záměrně igonruje přesný instrukce, že ten bůh má mít to brčko v nose.

    A depiction of a god inhaling the Milky Way through a straw into their nose, as if snorting a line of angel dust. The scene is high contrast, resembling the style of a tapestry. The god, adorned in majestic and detailed attire, is shown inhaling the glittering Milky Way through the straw, with cosmic elements swirling around them. The background features a vast, star-studded galaxy with vivid colors and intricate patterns, enhancing the grandeur of the universe. The format is wide to capture the expansiveness of the scene.





    KILLUA
    KILLUA --- ---
    AODHFIN: A funguje tak i člověk, pokud tě po narození zavřou do bílé krabice a nebude s tebou nikdo nijak 20 let interagovat max ti tam spadne kostka s jídlem tak jediná informační hodnota vzejde z naprogramovaných DNA vzorců typu dávivý reflex, strach a smyslových podnětů ze tvého vlastního těla tj např z prohlížení vlastních rukou mohou vzejít nějaké představy.

    Jinak nemá mozek ani ai model z čeho brát...
    FRK_R23
    FRK_R23 --- ---
    Zkoušel jsem jestli Fooocus umí taky třeba metahuman render :)

    prompt: Unreal engine metahuman, unreal engine render, lumen
    Base model: juggernautXL_v8

    INK_FLO
    INK_FLO --- ---
    The Age of Noise - by Eryk Salvaggio - Cybernetic Forests
    https://cyberneticforests.substack.com/p/the-age-of-noise

    Noise is a slippery word. It means both the presence and absence of information. Today it's in the urbanisation of our world, the hum of traffic and jet engines. Noise is also where we go to escape noise. In August of 2023, Spotify announced that users had listened to 3 million hours of white noise recordings. Noise to sleep to, noise to drown out noise. Noise is also the mental cacophony of data, on social media, of smartphones, and the algorithmic spectacle. The age of noise is a logical conclusion, a successful ending for the information age. And information, which was once scarce, is now spilling from the seams of our fibre optic cables and airwaves. The information age is over. Now we enter the age of noise. We can pin the information age to the invention of the transistor in 1947. The transistor was quaint by today's standards, a mechanism for handling on-off signals. Engineers built pathways through which voltage flowed, directing and controlling that voltage in response to certain inputs. We would punch holes into cards and feed them to a machine, running light through the holes into sensors. The cards became a medium, a set of instructions written in the language of yes and no.

    In other words, it all started with two decisions, yes or no, one or zero. The more we could feed the machine, the more the decisions the machine could make. Eventually it seemed the number of decisions began to encroach on our own. The machine said yes or no so that we didn't have to. By the start of the social media era, we were the ones responding to these holes. Like or don't like. Swipe left or swipe right. It all began with that maze of circuitry. The first neural networks, our adding machines, the earliest computers, were designed to reveal information. Noise meant anything that crept into the circuits, and the history of computing is in part a history of noise reduction. The noise in our telephone wires and circuit boards, even our analog TV broadcasts, was background radiation. Energy pulsing invisibly in the air, lingering for millennia after the Big Bang exploded our universe into being. Our task was to remove any traces of it from our phone calls. Today millions of on off calculations can take place in a single second. Put enough of these signals together, run them fast enough and you can do remarkably complex things with remarkable speed. Much of that has been harnessed into lighting up pixels. Put enough pixels together and you get a digital image. You get video games. You get live streams. You get maps, interfaces and you collect and process responses to live streams, maps and interfaces.

    With what we call generative AI today, we obviously aren't using punch cards. Now we inscribe our ones and zeros into digital images. The data mining corporations behind social media platforms take these digital images and they feed them to massive neural nets and data centres. In substance, the difference between punch cards and today's computation is only that our holes are smaller. Every image that we take becomes a computer program. Every caption and every label becomes a point of information. Today's generative AI models have learned from about 2.3 billion images with about 24 bits of information per pixel. All of them still at their core, a yes or no decision moving through a structure. I don't say this to give you a technical overview of image processing. I mention it because the entirety of human visual culture has a new name. We used to call these collections archives or museum holdings or libraries. Today we call them data sets. This collected culture has been harnessed to do the work of analog punch cards. And these cards, these physical objects, were once stamped with a warning. Do not fold, spindle or mutilate. Our collected visual heritage in its digital form carries no such warning.

    We don't feed our visual culture into a machine by hand anymore, and the number of decisions that we have automated are so large that even the words are ridiculous. Teraflops. We upload images to the internet, pictures of our birthday parties, our weddings, embarrassing nights at the club (not so much me anymore). Our drawings, our paintings, these personal images meant to communicate with others are clumped together with other archives. Cultural institutions share a wealth of knowledge online for the sake of human education and the arts history and beyond. And in training an AI model, all of these images are diffused, a word that is so neatly parallel to this diffusion of unfiltered information that we surround ourselves with. And for once, it's a technology named in a way that describes what it actually does. Diffusion models actually diffuse! This word means what it says. It dissolves the images, it strips information away from them until they resemble nothing but the fuzzy chaos of in between television channels. Images are diffused into noise. Billions of good and bad images all diffused into noise for the sake of training an artificial intelligence system that will produce a billion more images. From noise into noise, we move from the noise of billions of images taken from our noisy data-driven visual culture, isolate them and dissolve them into the literal noise of an empty JPEG, to be recreated again into the noise of one more meaningless image generated by AI among the noise of billions of other images, a count of images that already overwhelms any one person's desire to look at them.

    The information age has ended and we have entered the age of noise.

    We often think of noise as a presence. In America, we call it snow, the static. I've heard of other things as well. It's called ants in Thailand. Other places have other metaphors. But snow is a presence. We see snow. We see noise. We hear noise. Noise from a communication engineering perspective is the absence of information. Sometimes that absence is the result of too much information, a slippery paradox. Information which cannot be meaningfully discerned is still noise. Information has been rushing at us for about two decades now, pushing out information in the frame of content to such an extent that almost no signal remains that is worth engaging with. Here's a map of the internet visualised 20 years ago. Since then, it has only grown, today becoming a disorienting flood of good and bad information coming through the same channels. And what we are calling generative AI is the end result of a successful information age, which in just 24 years has rewritten all cultural norms about surveillance, public sharing, and our trust in corporatised collections of deeply personal data. Server farms mined this data through regimes of surveillance and financialisation. The guiding principle of social media has always been to lure us into sharing more so that more data could be collected, sold, and analysed. They've calibrated the speed of that sharing to meet the time scales of data centres rather than human comprehension or our desire to communicate. And all this data has become the food for today's generative AI.

    The words we shared built chat GPT, the images we shared built Stable Diffusion. Generative AI is just another word for surveillance capitalism. Taking our data with dubious consent and activating it through services it sells back to us. It is a visualisation of the way we organise things, a pretty picture version of the technologies that sorted and categorised us all along. Instead of social media feeds or bank loans or police lineups, these algorithms manifest as uncanny images, disorienting mirrors of the world rendered by a machine that has no experience of that world. If these images are unsettling because they resemble nothing like the lives they claim to represent, it's because that is precisely what automated surveillance was always doing to us. The internet was the Big Bang of the information era, and its noisy debris lingers within the Big Bang of generative AI. Famously, Open AI's chatbot stopped learning somewhere in April of 2021. That's when the bulk of its training was complete, and from there it was all just fine-tuning and calibration. Perhaps that marks the start of the age of noise, the age where streams of information blended into and overwhelmed one another in an indecipherable wall of static, so much information that truth and fiction dissolved into the same fuzz of background radiation.

    I worry that the age of noise will mark the era where we turn to machines to mediate this media sphere on our behalf. It follows a simple logic. To manage artificial information, we turn to artificial intelligence. But I have some questions. What are the strategies of artificial intelligence? The information management strategies that are responsible for the current regime of AI can be reduced to two, abstraction and prediction. We collect endless data about the past, abstract it into loose categories and labels, and then we draw from that data to make predictions. We ask the AI to tell us what the future will look like, what the next image might look like, what the next text might read like. It's all based on these abstractions of the data about the past. This used to be the role of archivists. Archivists used to be the custodians of the past, and archives and curators, facing limited resources of space and time, often pruned what would be preserved. And this shaped the archives. The subjects of these archives adapt themselves to the spaces we make for them. Just as mold grows in the lightest part of a certain film, history is what survives the contours we make for it. We can't save everything. But what history do we lose based on the size of our shelves? These are a series of subjective, institutionalised decisions made by individuals within the context of their positions and biases and privileges and ignorances. The funding mandates, the space, and the time. (No offence!)

    Humans never presided over a golden age of inclusivity, but at least the decisions were there on display. The archive provided its own evidence of its gaps. What was included was there, and what was excluded was absent. And those absences could be challenged. Humans could be confronted. Advocates could speak out. I'm reminded of my work with Wikipedia, simultaneously overwhelmed with biographies of men, but also host to a remarkable effort by volunteers to organise and produce biographies of women. When humans are in the loop, humans can intervene in the loop.

    I'm often asked if I fear that AI will replace human creativity, and I don't remotely understand the question. Creativity is where agency rises, and as our agency is questioned, it is more important than ever to reclaim it, through creativity, not adaptability. Not contorting ourselves to machines, but agency — contorting the machines to us. I fear that we will automate our decisions and leave out variations of past patterns based on the false belief that only repetition is possible. Of course, my work is also a remix. It has a lineage. To Nam June Paik, who famously quipped, “I use technology in order to hate it properly.” And this is part of the tension, the contradictions that we're all grappling with. I'm trying to explore the world between archive and training data, between the meaningful acknowledgement of the past and the meaningless reanimation of the past through quantification. Archives are far more than just data points. We're using people's personal stories and difficult experiences for this. There's a beauty of lives lived and the horrors, too. Training images are more than data. There is more to our archives than the clusters of light-coloured pixels. Our symbols and words have meaning because of their context in collective memory. When we remove that, they lose their connection to culture. If we strip meaning from the archive, we have a meaningless archive. We have five billion pieces of information that lack real-world connections. Five billion points of noise. Rather than drifting into the mindset of data brokers, it is critical that we as artists, as curators, as policymakers approach the role of AI in the humanities from a position of the archivist, historian, humanitarian, and storyteller. That is, to resist the demand that we all become engineers and that all history is data science.

    We need to see knowledge as a collective project, to push for more people to be involved, not less, to insist that meaning and context matters, and to preserve and contest those contexts in all their complexity. If artificial intelligence strips away context, human intelligence will find meaning. If AI plots patterns, humans must find stories. If AI reduces and isolates, humans must find ways to connect and to flourish. There is a trajectory for humanity that rests beyond technology. We are not asleep in the halls of the archive, dreaming of the past. Let's not place human agency into the dark, responsive corners. The challenge of this age of noise is to find and preserve meaning. The antidote to chaos is not enforcing more control. It's elevating context. Fill in the gaps and give the ghosts some peace.

    Looking at the Machine | FACT24 Symposium
    https://www.youtube.com/watch?v=Eqw7U8BA5aM
    ARAON
    ARAON --- ---
    More workflow tests. Here it's photoshop sketch to Krea to 3D. It takes roughly 15 seconds or so to generate a model from the sketch.
    https://twitter.com/MartinNebelong/status/1763900406759882944
    KERRAY
    KERRAY --- ---
    ad KERRAY, tady je k tomu paper a v něm spousta dalších zajímavostí a videí - například dokážou "renderovat" Minecraft a jsou tam k tomu #video - protože to není "jen" generátor obrázků, ale vytváří si vlastně model světa, včetně fyziky atd., ze kterého pak teprve vychází to video

    Simulating digital worlds. Sora is also able to simulate artificial processes–one example is video games. Sora can simultaneously control the player in Minecraft with a basic policy while also rendering the world and its dynamics in high fidelity. These capabilities can be elicited zero-shot by prompting Sora with captions mentioning “Minecraft.”

    Video generation models as world simulators
    Video generation models as world simulators
    https://openai.com/research/video-generation-models-as-world-simulators
    KILLUA
    KILLUA --- ---
    KERRAY: Jaký je to model nová MJ?
    KERRAY
    KERRAY --- ---
    #nástroj #midjourney
    Our first major update to V6 alpha is now live. All major qualities of the model are improved; aesthetics, coherence, prompt adherence, image quality, and text rendering. Higher values of --stylize also work much better and upscaling is now ~2x faster. Enjoy!
    VOZKA
    VOZKA --- ---
    Zaregistrovali jste vydání SDXL Turbo?

    Je to vydestilovaný model, který generuje použitelné obrázky už ve dvou krocích. A u mě, starší počítač a 6 GB VRAM s nutností použití pomalého --lowvram módu, se každý krok počítá cca o polovinu rychleji. Takže zatímco u SDXL se mi použitelný obrázek generuje přes tři minuty, tady to mám za dvacet vteřin. Na rychlejších současných GPU to je prý víceméně realtime.

    Funguje to v ComfyUI, s nutností použít trochu jiný workflow (u nich na webu mají někde ukázku), s A1111 předpokládám taky.

    Kvalita je někde subjektivně stejně dobrá, někde, když po tom chcete generovat méně standardní věci (typu screenshot ze staré pixel artové videohry), funguje hůř. Testoval jsem jen chvilku, ale přijde mi, že slabé stránky SDXL jsou ještě o něco slabší, silné stránky jsou srovnatelné.
    KAJJAK
    KAJJAK --- ---
    PEETIK: natrenovat vlastni model v SD?
    MORPHLER
    MORPHLER --- ---
    PECA: jo jeste je samozrejme cesta ze to syn vsechno instaluje z githubu a bezi mu nejaky model lokalne, ale to by asi bylo spise na pochvalu, ne?
    E2E4
    E2E4 --- ---
    Bing chat a Twitter mi omylem neco prozradily o bing image creatoru. Vnitrne pouziva jakysi graphic_art.

    1. daji se zadat parametry, zrejmena --r rozliseni.
    2. druhak se tam da zadat primo prompt pro dall-e 3, ne ze by ho AI vygenerovala. je to i o chlup rychlejsi..

    #graphic_art("sofa couch chia armchair ceramic orb acorn mushroom porcelain futuristic crochet")

    Da se to zadat jak do bing chatu, tak do https://www.bing.com/images/create

    Zdroje:

    1. https://twitter.com/zer0int1/status/1709809615582450085

    2. odpoved bingu kde se prokecl
    The image I created for you was a one-time output of my graphic_art tool, which uses an artificial intelligence model to generate images based on prompts. The tool is not deterministic, meaning that it can produce different results each time it is invoked with the same prompt. Therefore, I can’t guarantee that I can reproduce the exact same image that I showed you before. 😕

    However, if you want, I can try to create a new image for you with the same prompt or a different one. You can also adjust the parameters of the tool, such as the number of colors (—c), the aspect ratio (—ar), or the resolution (—r). Just let me know what you would like me to do. 😊

    Akorat jsem neprisel na to jak to i presne dat dohromady.

    Kazdopadne, heh, nemuze, ale pak to stejne udela.. :)

    VOZKA
    VOZKA --- ---
    KAJJAK: Tam jen vložíš ty dva safetensors soubory do klasický model složky ComfyUI/models/checkpoints, spustíš, copypastneš do okna co se ti otevře ten textový soubor na který linkuju v příspěvku níž a tím se ti vytvoří sestava nodů která funguje a po spuštění vygeneruje něco jako ten můj řízek.
    VOZKA
    VOZKA --- ---
    VOZKA: Ale co to tak čtu, tak tenhle patch zatím podporuje jen base model, ne refiner. SDXL jsou dva modely - jeden který vytvoří obrázek a druhý který ho pak doladí do lepší kvality, a Automatic zatím umí jen tu první polovinu. Takže "plná" verze je zatím jen v tom ComfyUI.
    KERRAY
    KERRAY --- ---
    a druhý #text #tts #nástroj je Suno Bark, který už jsem tu myslím dřív postoval, otevřeli Discord, kde jde generovat - to češtinu neumí, ale jinak to generovalo dost přirozené hlasy
    GitHub - suno-ai/bark: 🔊 Text-Prompted Generative Audio Model
    https://github.com/suno-ai/bark

    Suno
    https://www.suno.ai/
    FRK_R23
    FRK_R23 --- ---
    FRK_R23: Jo zkouším to na leonardo.ai, model Leonardo Diffusion.
    KERRAY
    KERRAY --- ---
    #nástroj tipy na trénování vlastních LoRA modelů
    Reddit - Dive into anything
    https://www.reddit.com/r/StableDiffusion/comments/13dh7ql/after_training_50_lora_models_here_is_what_i/

    Style Training :

    - use 30-100 images (avoid same subject, avoid big difference in style)

    - good captioning (better caption manually instead of BLIP) with alphanumeric trigger words (styl3name).

    - use pre-existing style keywords (i.e. comic, icon, sketch)

    - caption formula styl3name, comic, a woman in white dress

    - train with a model that can already produce a close looking style that you are trying to acheive.

    - avoid stablediffusion base model beacause it is too diverse and we want to remain specific


    Person/Character Training:

    - use 30-100 images (atleast 20 closeups and 10 body shots)

    - face from different angles, body in different clothing and in different lighting but not too much diffrence, avoid pics with eye makeup

    - good captioning (better caption manually instead of BLIP) with alphanumeric trigger words (ch9ractername)

    - avoid deep captioning like "a 25 year woman in pink printed tshirt and blue ripped denim striped jeans, gold earing, ruby necklace"

    - caption formula ch9ractername, a woman in pink tshirt and blue jeans

    - for real person, train on RealisticVision model, Lora trained on RealisticVision works with most of the models

    - for character training use train with a model that can already produce a close looking character (i.e. for anime i will prefer anythinv3)

    - avoid stablediffusion base model beacause it is too diverse and we want to remain specific
    DAVE2
    DAVE2 --- ---
    ARAON: Takové nástroje jsou tu už dlouho, například v podobě generovaných stromů, popínavek apod. Ale je to podobné jako stará GOFAI - člověk tomu musí vysvětlit co je kmen co podružné větve a že na tom mají být listy. Spousta lidské práce s omezeným využitím (z generátoru stromů padají jenom různé typy stromů). Představují si difuzní 3D model, který začne shlukem nějakých polygonů a postupně se odbourá šum natolik, že z toho vznikne ten strom. Nebo třípodlažní budova z roku 1905 nebo cokoliv si člověk vymyslí.
    Kliknutím sem můžete změnit nastavení reklam