• úvod
  • témata
  • události
  • tržiště
  • diskuze
  • nástěnka
  • přihlásit
    registrace
    ztracené heslo?
    KERRAYAI Art - Tvorba za pomocí umělé inteligence: Midjourney, DALL·E 2, Stable Diffusion, OpenAI
    Vše ohledně umění tvořeného pomocí umělé inteligence - obrázky, 'fotky', galerie, hudba, video, text + články, novinky apod.

    NSFW obsah prosím obalit spoiler tagem - <div class="spoiler">obrázek</spoiler>, a nemá to tu být klub na roštěnky a nahotinky bez nějaké přidané hodnoty

    Příbuzné diskuze:
    - [DALL·E mini je mrtvé, Dejve, ať žije Craiyon, mage.space, Stable Diffusion atd. - having sex with AI since [date format unknown]]
    - [I Hope This Does Not Exist ​ ​ ​ ▌​ ​ Vedlejší efekty v AI visuálech]
    - AI obecně [Artificial Intelligence AI]
    - Vtipy [Umělá inteligence, chatboti - vtipné konverzace aneb "Hoří hovno?"]
    - [generativní modely] Jak konstruovat prompty, kde získat váhy i jak to vše interpretovat


    Prosím zkusme pro vkládanou tvorbu používat tagy
    #galerie (2-3 obrázky na ukázku, další po rozkliku) #obrázek #video #hudba #text #hry #původní (pro vlastní tvorbu) #roštěnky

    #článek #nástroj

    (návrhy na další tagy apod. vítány)

    Texty, programování: https://beta.openai.com/playground | https://chat.openai.com/
    Obrázky online: https://www.midjourney.com/ | https://beta.dreamstudio.ai/
    Lokálně: Webui-Forge https://github.com/lllyasviel/stable-diffusion-webui-forge | ComfyUI https://github.com/comfyanonymous/ComfyUI | Civitai repository custom modelů pro SD
    AI na vytvoření textového zadání z existujícího obrázku: https://huggingface.co/spaces/pharma/CLIP-Interrogator
    rozbalit záhlaví
    DOUPLE
    DOUPLE --- ---
    par dni stary model Flux umi neuveritelny veci s textem
    FLUX: This new AI image generator is eerily good at creating human hands | Ars Technica
    https://arstechnica.com/information-technology/2024/08/flux-this-new-ai-image-generator-is-eerily-good-at-creating-human-hands/

    THERIDANE
    THERIDANE --- ---
    KILLUA: proto mají abliterované/ortogonalizované modely vyšší score než originální, odstraní se ten zmatek a model pak dělá méně chyb (ale řekne opravdu všechno, co ví, a neodmítá)
    KILLUA
    KILLUA --- ---
    HONZA09: Mistral má oproti velkým hráčům malé zdroje takže se snaží udělat svůj model konkurenceschopný všemi prostředky a jeden z nejsilnějších je co nejméně ho cenzurovat.

    Když se nad tím totiž zamyslíš model se něco naučí je to pro něj "pravda" no a pak přijde ten "tréning" kde mu říkáš že to pravda není... výsledek je že v tom je nějakej "šum" zmatek.
    KERRAY
    KERRAY --- ---
    #nástroj #midjourney
    Hey! We're releasing an early test version of model personalization today, here's how it works

    What is Model Personalization?
    Every time you write a prompt there's a lot that remains 'unspoken'. Our algorithms usually fill in the blank with their own 'preferences', which are really the combined biases and preferences of our community.
    But of course everyone is different! Model personalization learns what you like so that it's more likely to fill in the blanks with your tastes.

    Model Personalization Requirements
    Right now model personalization learns from votes in pair ranking and images that you like from the explore page.
    You need to have roughly 200 pair rankings / likes in order for the feature to work
    You can see how many ratings you have on the above page or by typing /info

    How do I use Model Personalization?
    Just type --p after your prompt, or turn on personalization for all prompts under /settings or the prompt settings button on the website
    When you enable personalization you get a 'code' added after your prompts, you can share this code to share the personalization effect being applied to that image
    You can control the strength of the personalization effect by using --s 100 (0 is off and 1000 is maximum and 100 is default)

    PLEASE NOTE: The personalization is not a stable feature right now, it will change as you do more pair ratings (subtly) and we may be pushing out algorithm updates over the next few weeks. So just assume that it's a 'fun feature constantly in flux'

    This is a totally new frontier of image synthesis.
    ARAON
    ARAON --- ---
    Sora by OpenAI is insane.

    But KWAI just dropped a Sora-like model called KLING, and people are going crazy over it.

    Here are 10 wild examples you don't want to miss:

    1. A Chinese man sits at a table and eats noodles with chopsticks
    x.com
    https://x.com/i/status/1798777783952527818
    GALADAR
    GALADAR --- ---
    HONZA09: Když narážíš na content restrictions, model tě upozorní. (a rád i sám poradí, jak je obejít) Tohle je jen chyba v promptu. Mmch "angel dust" už je direct drug reference, dívím se, že tě nestopl. Doporučuji flour, white powder, powdered caffeine atd.
    HONZA09
    HONZA09 --- ---
    Tak jsem našel omezení. nejsem za boha schopnej v Dall E vygenerovat obrázek boha dávajícího si lajnu. Ten generativní model záměrně igonruje přesný instrukce, že ten bůh má mít to brčko v nose.

    A depiction of a god inhaling the Milky Way through a straw into their nose, as if snorting a line of angel dust. The scene is high contrast, resembling the style of a tapestry. The god, adorned in majestic and detailed attire, is shown inhaling the glittering Milky Way through the straw, with cosmic elements swirling around them. The background features a vast, star-studded galaxy with vivid colors and intricate patterns, enhancing the grandeur of the universe. The format is wide to capture the expansiveness of the scene.





    KILLUA
    KILLUA --- ---
    AODHFIN: A funguje tak i člověk, pokud tě po narození zavřou do bílé krabice a nebude s tebou nikdo nijak 20 let interagovat max ti tam spadne kostka s jídlem tak jediná informační hodnota vzejde z naprogramovaných DNA vzorců typu dávivý reflex, strach a smyslových podnětů ze tvého vlastního těla tj např z prohlížení vlastních rukou mohou vzejít nějaké představy.

    Jinak nemá mozek ani ai model z čeho brát...
    FRK_R23
    FRK_R23 --- ---
    Zkoušel jsem jestli Fooocus umí taky třeba metahuman render :)

    prompt: Unreal engine metahuman, unreal engine render, lumen
    Base model: juggernautXL_v8

    INK_FLO
    INK_FLO --- ---
    The Age of Noise - by Eryk Salvaggio - Cybernetic Forests
    https://cyberneticforests.substack.com/p/the-age-of-noise

    Noise is a slippery word. It means both the presence and absence of information. Today it's in the urbanisation of our world, the hum of traffic and jet engines. Noise is also where we go to escape noise. In August of 2023, Spotify announced that users had listened to 3 million hours of white noise recordings. Noise to sleep to, noise to drown out noise. Noise is also the mental cacophony of data, on social media, of smartphones, and the algorithmic spectacle. The age of noise is a logical conclusion, a successful ending for the information age. And information, which was once scarce, is now spilling from the seams of our fibre optic cables and airwaves. The information age is over. Now we enter the age of noise. We can pin the information age to the invention of the transistor in 1947. The transistor was quaint by today's standards, a mechanism for handling on-off signals. Engineers built pathways through which voltage flowed, directing and controlling that voltage in response to certain inputs. We would punch holes into cards and feed them to a machine, running light through the holes into sensors. The cards became a medium, a set of instructions written in the language of yes and no.

    In other words, it all started with two decisions, yes or no, one or zero. The more we could feed the machine, the more the decisions the machine could make. Eventually it seemed the number of decisions began to encroach on our own. The machine said yes or no so that we didn't have to. By the start of the social media era, we were the ones responding to these holes. Like or don't like. Swipe left or swipe right. It all began with that maze of circuitry. The first neural networks, our adding machines, the earliest computers, were designed to reveal information. Noise meant anything that crept into the circuits, and the history of computing is in part a history of noise reduction. The noise in our telephone wires and circuit boards, even our analog TV broadcasts, was background radiation. Energy pulsing invisibly in the air, lingering for millennia after the Big Bang exploded our universe into being. Our task was to remove any traces of it from our phone calls. Today millions of on off calculations can take place in a single second. Put enough of these signals together, run them fast enough and you can do remarkably complex things with remarkable speed. Much of that has been harnessed into lighting up pixels. Put enough pixels together and you get a digital image. You get video games. You get live streams. You get maps, interfaces and you collect and process responses to live streams, maps and interfaces.

    With what we call generative AI today, we obviously aren't using punch cards. Now we inscribe our ones and zeros into digital images. The data mining corporations behind social media platforms take these digital images and they feed them to massive neural nets and data centres. In substance, the difference between punch cards and today's computation is only that our holes are smaller. Every image that we take becomes a computer program. Every caption and every label becomes a point of information. Today's generative AI models have learned from about 2.3 billion images with about 24 bits of information per pixel. All of them still at their core, a yes or no decision moving through a structure. I don't say this to give you a technical overview of image processing. I mention it because the entirety of human visual culture has a new name. We used to call these collections archives or museum holdings or libraries. Today we call them data sets. This collected culture has been harnessed to do the work of analog punch cards. And these cards, these physical objects, were once stamped with a warning. Do not fold, spindle or mutilate. Our collected visual heritage in its digital form carries no such warning.

    We don't feed our visual culture into a machine by hand anymore, and the number of decisions that we have automated are so large that even the words are ridiculous. Teraflops. We upload images to the internet, pictures of our birthday parties, our weddings, embarrassing nights at the club (not so much me anymore). Our drawings, our paintings, these personal images meant to communicate with others are clumped together with other archives. Cultural institutions share a wealth of knowledge online for the sake of human education and the arts history and beyond. And in training an AI model, all of these images are diffused, a word that is so neatly parallel to this diffusion of unfiltered information that we surround ourselves with. And for once, it's a technology named in a way that describes what it actually does. Diffusion models actually diffuse! This word means what it says. It dissolves the images, it strips information away from them until they resemble nothing but the fuzzy chaos of in between television channels. Images are diffused into noise. Billions of good and bad images all diffused into noise for the sake of training an artificial intelligence system that will produce a billion more images. From noise into noise, we move from the noise of billions of images taken from our noisy data-driven visual culture, isolate them and dissolve them into the literal noise of an empty JPEG, to be recreated again into the noise of one more meaningless image generated by AI among the noise of billions of other images, a count of images that already overwhelms any one person's desire to look at them.

    The information age has ended and we have entered the age of noise.

    We often think of noise as a presence. In America, we call it snow, the static. I've heard of other things as well. It's called ants in Thailand. Other places have other metaphors. But snow is a presence. We see snow. We see noise. We hear noise. Noise from a communication engineering perspective is the absence of information. Sometimes that absence is the result of too much information, a slippery paradox. Information which cannot be meaningfully discerned is still noise. Information has been rushing at us for about two decades now, pushing out information in the frame of content to such an extent that almost no signal remains that is worth engaging with. Here's a map of the internet visualised 20 years ago. Since then, it has only grown, today becoming a disorienting flood of good and bad information coming through the same channels. And what we are calling generative AI is the end result of a successful information age, which in just 24 years has rewritten all cultural norms about surveillance, public sharing, and our trust in corporatised collections of deeply personal data. Server farms mined this data through regimes of surveillance and financialisation. The guiding principle of social media has always been to lure us into sharing more so that more data could be collected, sold, and analysed. They've calibrated the speed of that sharing to meet the time scales of data centres rather than human comprehension or our desire to communicate. And all this data has become the food for today's generative AI.

    The words we shared built chat GPT, the images we shared built Stable Diffusion. Generative AI is just another word for surveillance capitalism. Taking our data with dubious consent and activating it through services it sells back to us. It is a visualisation of the way we organise things, a pretty picture version of the technologies that sorted and categorised us all along. Instead of social media feeds or bank loans or police lineups, these algorithms manifest as uncanny images, disorienting mirrors of the world rendered by a machine that has no experience of that world. If these images are unsettling because they resemble nothing like the lives they claim to represent, it's because that is precisely what automated surveillance was always doing to us. The internet was the Big Bang of the information era, and its noisy debris lingers within the Big Bang of generative AI. Famously, Open AI's chatbot stopped learning somewhere in April of 2021. That's when the bulk of its training was complete, and from there it was all just fine-tuning and calibration. Perhaps that marks the start of the age of noise, the age where streams of information blended into and overwhelmed one another in an indecipherable wall of static, so much information that truth and fiction dissolved into the same fuzz of background radiation.

    I worry that the age of noise will mark the era where we turn to machines to mediate this media sphere on our behalf. It follows a simple logic. To manage artificial information, we turn to artificial intelligence. But I have some questions. What are the strategies of artificial intelligence? The information management strategies that are responsible for the current regime of AI can be reduced to two, abstraction and prediction. We collect endless data about the past, abstract it into loose categories and labels, and then we draw from that data to make predictions. We ask the AI to tell us what the future will look like, what the next image might look like, what the next text might read like. It's all based on these abstractions of the data about the past. This used to be the role of archivists. Archivists used to be the custodians of the past, and archives and curators, facing limited resources of space and time, often pruned what would be preserved. And this shaped the archives. The subjects of these archives adapt themselves to the spaces we make for them. Just as mold grows in the lightest part of a certain film, history is what survives the contours we make for it. We can't save everything. But what history do we lose based on the size of our shelves? These are a series of subjective, institutionalised decisions made by individuals within the context of their positions and biases and privileges and ignorances. The funding mandates, the space, and the time. (No offence!)

    Humans never presided over a golden age of inclusivity, but at least the decisions were there on display. The archive provided its own evidence of its gaps. What was included was there, and what was excluded was absent. And those absences could be challenged. Humans could be confronted. Advocates could speak out. I'm reminded of my work with Wikipedia, simultaneously overwhelmed with biographies of men, but also host to a remarkable effort by volunteers to organise and produce biographies of women. When humans are in the loop, humans can intervene in the loop.

    I'm often asked if I fear that AI will replace human creativity, and I don't remotely understand the question. Creativity is where agency rises, and as our agency is questioned, it is more important than ever to reclaim it, through creativity, not adaptability. Not contorting ourselves to machines, but agency — contorting the machines to us. I fear that we will automate our decisions and leave out variations of past patterns based on the false belief that only repetition is possible. Of course, my work is also a remix. It has a lineage. To Nam June Paik, who famously quipped, “I use technology in order to hate it properly.” And this is part of the tension, the contradictions that we're all grappling with. I'm trying to explore the world between archive and training data, between the meaningful acknowledgement of the past and the meaningless reanimation of the past through quantification. Archives are far more than just data points. We're using people's personal stories and difficult experiences for this. There's a beauty of lives lived and the horrors, too. Training images are more than data. There is more to our archives than the clusters of light-coloured pixels. Our symbols and words have meaning because of their context in collective memory. When we remove that, they lose their connection to culture. If we strip meaning from the archive, we have a meaningless archive. We have five billion pieces of information that lack real-world connections. Five billion points of noise. Rather than drifting into the mindset of data brokers, it is critical that we as artists, as curators, as policymakers approach the role of AI in the humanities from a position of the archivist, historian, humanitarian, and storyteller. That is, to resist the demand that we all become engineers and that all history is data science.

    We need to see knowledge as a collective project, to push for more people to be involved, not less, to insist that meaning and context matters, and to preserve and contest those contexts in all their complexity. If artificial intelligence strips away context, human intelligence will find meaning. If AI plots patterns, humans must find stories. If AI reduces and isolates, humans must find ways to connect and to flourish. There is a trajectory for humanity that rests beyond technology. We are not asleep in the halls of the archive, dreaming of the past. Let's not place human agency into the dark, responsive corners. The challenge of this age of noise is to find and preserve meaning. The antidote to chaos is not enforcing more control. It's elevating context. Fill in the gaps and give the ghosts some peace.

    Looking at the Machine | FACT24 Symposium
    https://www.youtube.com/watch?v=Eqw7U8BA5aM
    ARAON
    ARAON --- ---
    More workflow tests. Here it's photoshop sketch to Krea to 3D. It takes roughly 15 seconds or so to generate a model from the sketch.
    https://twitter.com/MartinNebelong/status/1763900406759882944
    KERRAY
    KERRAY --- ---
    ad KERRAY, tady je k tomu paper a v něm spousta dalších zajímavostí a videí - například dokážou "renderovat" Minecraft a jsou tam k tomu #video - protože to není "jen" generátor obrázků, ale vytváří si vlastně model světa, včetně fyziky atd., ze kterého pak teprve vychází to video

    Simulating digital worlds. Sora is also able to simulate artificial processes–one example is video games. Sora can simultaneously control the player in Minecraft with a basic policy while also rendering the world and its dynamics in high fidelity. These capabilities can be elicited zero-shot by prompting Sora with captions mentioning “Minecraft.”

    Video generation models as world simulators
    Video generation models as world simulators
    https://openai.com/research/video-generation-models-as-world-simulators
    KILLUA
    KILLUA --- ---
    KERRAY: Jaký je to model nová MJ?
    KERRAY
    KERRAY --- ---
    #nástroj #midjourney
    Our first major update to V6 alpha is now live. All major qualities of the model are improved; aesthetics, coherence, prompt adherence, image quality, and text rendering. Higher values of --stylize also work much better and upscaling is now ~2x faster. Enjoy!
    VOZKA
    VOZKA --- ---
    Zaregistrovali jste vydání SDXL Turbo?

    Je to vydestilovaný model, který generuje použitelné obrázky už ve dvou krocích. A u mě, starší počítač a 6 GB VRAM s nutností použití pomalého --lowvram módu, se každý krok počítá cca o polovinu rychleji. Takže zatímco u SDXL se mi použitelný obrázek generuje přes tři minuty, tady to mám za dvacet vteřin. Na rychlejších současných GPU to je prý víceméně realtime.

    Funguje to v ComfyUI, s nutností použít trochu jiný workflow (u nich na webu mají někde ukázku), s A1111 předpokládám taky.

    Kvalita je někde subjektivně stejně dobrá, někde, když po tom chcete generovat méně standardní věci (typu screenshot ze staré pixel artové videohry), funguje hůř. Testoval jsem jen chvilku, ale přijde mi, že slabé stránky SDXL jsou ještě o něco slabší, silné stránky jsou srovnatelné.
    KAJJAK
    KAJJAK --- ---
    PEETIK: natrenovat vlastni model v SD?
    MORPHLER
    MORPHLER --- ---
    PECA: jo jeste je samozrejme cesta ze to syn vsechno instaluje z githubu a bezi mu nejaky model lokalne, ale to by asi bylo spise na pochvalu, ne?
    E2E4
    E2E4 --- ---
    Bing chat a Twitter mi omylem neco prozradily o bing image creatoru. Vnitrne pouziva jakysi graphic_art.

    1. daji se zadat parametry, zrejmena --r rozliseni.
    2. druhak se tam da zadat primo prompt pro dall-e 3, ne ze by ho AI vygenerovala. je to i o chlup rychlejsi..

    #graphic_art("sofa couch chia armchair ceramic orb acorn mushroom porcelain futuristic crochet")

    Da se to zadat jak do bing chatu, tak do https://www.bing.com/images/create

    Zdroje:

    1. https://twitter.com/zer0int1/status/1709809615582450085

    2. odpoved bingu kde se prokecl
    The image I created for you was a one-time output of my graphic_art tool, which uses an artificial intelligence model to generate images based on prompts. The tool is not deterministic, meaning that it can produce different results each time it is invoked with the same prompt. Therefore, I can’t guarantee that I can reproduce the exact same image that I showed you before. 😕

    However, if you want, I can try to create a new image for you with the same prompt or a different one. You can also adjust the parameters of the tool, such as the number of colors (—c), the aspect ratio (—ar), or the resolution (—r). Just let me know what you would like me to do. 😊

    Akorat jsem neprisel na to jak to i presne dat dohromady.

    Kazdopadne, heh, nemuze, ale pak to stejne udela.. :)

    VOZKA
    VOZKA --- ---
    KAJJAK: Tam jen vložíš ty dva safetensors soubory do klasický model složky ComfyUI/models/checkpoints, spustíš, copypastneš do okna co se ti otevře ten textový soubor na který linkuju v příspěvku níž a tím se ti vytvoří sestava nodů která funguje a po spuštění vygeneruje něco jako ten můj řízek.
    Kliknutím sem můžete změnit nastavení reklam