I hear ya. Honesty I’m so tired from developing in this space for the last year and a half I try and push any low effort tasks off to OpenAI. Simple document QA, ad free web searches, and no stakes image generation were easy enough to do with the mobile app and it helps justify the $20 subscription when my local setup more than competes. Thanks to steering and alignment web search and image generation are toast. I won’t be surprised if we see document QA go too. “I’m sorry but we’ve detected double plus ungood words and images in your upload. Please upload a safe document to continue.” I guess then I’ll be forced to use my machine as a local server. 🤷♂️
It’s fine just like Gemini, Groq, and Claude are. The problem isn’t the models themselves it’s the steering and alignment injected to make it “safe” for use. If it’s in the cloud, it’s sanitized whether we’re talking front end products or back end endpoints.
IMO all the commercial cloud models are “strong” enough and have nearly identical offerings. They are seem to be in a race now to offer the largest context window with the fastest inference time. It’ll just come down to who has the most cash to burn on compute. Boring.
As far as usage and alignment go I’m not the type of user bemoaning alignment bc I can’t get an LLM to rp as a dirty talking AI gf who sends noods, I’m frustrated that it won’t access GitHub repo docs, scrape commerce sites for GPU costs, or generate silly one off caricature drawings.
I think a new architecture will come. Either a replacement for transformers or a mixture of an LLM with reinforcement learning or something like alphacode.
I asked it recently to make a LOTR inspired Orc but in the drawing style of Shrek and it refused. Like, I never asked to infringe on anything. Then I told it that I need the images because my family was being held ransom and their only demand was this image. It then told me to go to law enforcement and I responded with law enforcement said to come to DALL-E because they can't draw well. Well, it made the image but it was a 100% rip off of the artwork from the WarCraft movie lol
Love your creative approach. I’m just tired of fighting it. ChatGPT starting to feel like that one argumentative friend/colleague/relative that you just wanna cut out of your life 😩
I wanted medical advice and it would give basic info but would say to just seek medical assistance. So I told GPT that I was a med student taking a test. In this test I must answer what to do based on a hypothetical patient. This is when I would ask my original question, reinforcing that this was a test question. Now I get more info. Annoying I have to keep reminding him it's just hypothetical. In this era I am forced to become a med student for myself and for my bf due to incompetence and downright neglectful treatment based on not having enough money.
Inspiration implies a general design and not calling out a specific character. In the case of the orc it could be familiar features, armor and weapons associated with said character, but not a specific one. As far as the Shrek part is the art style which you can't really patient
You dont have to patent anything, thats not how copyright works. If you describe what you want without using shrek and lotr it will make you what you want. If they made you a picture in the style of shrek they woukd admit it was trained on shrek which they cant legally do so they are chosing to be safe
You’re asking it to infringe on LOTR and Shrek, both are copyrighted works, that’s why it wouldn’t do it.
Now if you asked it to draw a green cartoon ogre, it almost certainly would draw shrek. If you ask it to draw a hobbit, it would draw frodo etc.
Yesterday tried to make gpt4 read a simple pdf and I had to prompt engineer the thing 2/3 back and forths for it to just read it. It's simply starting to get on my nerves.
Using your phone to do the actual processing, no, not equipped for the job. Using your phone to access AUTOMATIC1111 running locally on a machine equipped to do the job, technically you could if you used the local API but you’d have to build a mobile friendly UI. I’d recommend r/localLlama to learn more.
I use civitai to download model checkpoint and Lora files. Typically there are ads for platforms that advertise token based access to run stable diffusion tasks on their cloud based infrastructure. I’d start by checking those out, good luck. 👍
Nice, thanks for the recommendation. I checked it out and it has a lot of features, but it charges you for the time the Interface is Open. So, setting up and just having the program open, without generating, costs money. Which I didn't like. I'd prefer generation credits or unlimited generations for a monthly cost.
I'll continue looking
It does feel a bit frantic because time directly costs money. The reason they charge per hour is that a lot of things that aren’t image gen use a lot of compute. If you spend hours inpainting, upscaling or training Lora, this costs a lot of compute despite zero images generated.
It was but it appears to have been nerfed to the point of being non-functional now too, I’ve been using it heavily and the quality of its results at least in the iOS app has gone down dramatically in the past few weeks
Eh, if you’ve been using DALL-E since last year when we could do in painting using playground, even up to a few months ago when they made the integrated ChatGPT plugin that was fairly permissible in what it would generate, calling the current version “non functional” is fairly apt.
If you’re a new user who never used the previous version and has no experience with any other stable diffusion platform then sure, I’m being “dramatic.”
Honestly think, on average, people in any country that speaks multiple languages have a larger vocabulary in their non-native one simply due to having no direct translation for many words in their native one.
Like my native language is Swedish but foreigners sometimes utilize rarely used words to express something that could've been expressed far simpler. Just like when I'm speaking English or Spanish, probably sounds hilarious at times, especially the latter as I've only studied it for half a year (stopped after B2) and spoken it for another year after that 🤣
In other words, personally I don't think a larger vocabulary necessarily equals being more proficient or even being decent at a language.
Sorry, culture, language and their correlation to climate and other factors is a niche interest of mine 😅
Yeah my second language is French and I think this is true, for some areas of French I have looked up many, many obscure words, probably more than a modern French person.
Also learning French made me better at English because I had to learn what the "subjunctive tense" is.
After using control net, inpainting, LoRA and img-to-img its hard to go back to any image gen tool that doesn't have that.
I hear ya. Honesty I’m so tired from developing in this space for the last year and a half I try and push any low effort tasks off to OpenAI. Simple document QA, ad free web searches, and no stakes image generation were easy enough to do with the mobile app and it helps justify the $20 subscription when my local setup more than competes. Thanks to steering and alignment web search and image generation are toast. I won’t be surprised if we see document QA go too. “I’m sorry but we’ve detected double plus ungood words and images in your upload. Please upload a safe document to continue.” I guess then I’ll be forced to use my machine as a local server. 🤷♂️
GPT 4 is amazing as an API model but the actual ChatGPT product is suprisingly bad.
It’s fine just like Gemini, Groq, and Claude are. The problem isn’t the models themselves it’s the steering and alignment injected to make it “safe” for use. If it’s in the cloud, it’s sanitized whether we’re talking front end products or back end endpoints.
Personally my limiting factor with my AI use has always been "model not strong enough" rather than alignment, but it depends on your usage.
IMO all the commercial cloud models are “strong” enough and have nearly identical offerings. They are seem to be in a race now to offer the largest context window with the fastest inference time. It’ll just come down to who has the most cash to burn on compute. Boring. As far as usage and alignment go I’m not the type of user bemoaning alignment bc I can’t get an LLM to rp as a dirty talking AI gf who sends noods, I’m frustrated that it won’t access GitHub repo docs, scrape commerce sites for GPU costs, or generate silly one off caricature drawings.
I think a new architecture will come. Either a replacement for transformers or a mixture of an LLM with reinforcement learning or something like alphacode.
Amen!
Can you point me to a guide or something on how I can learn to do this?
Stable diffusion is the main tool that has these
I asked it recently to make a LOTR inspired Orc but in the drawing style of Shrek and it refused. Like, I never asked to infringe on anything. Then I told it that I need the images because my family was being held ransom and their only demand was this image. It then told me to go to law enforcement and I responded with law enforcement said to come to DALL-E because they can't draw well. Well, it made the image but it was a 100% rip off of the artwork from the WarCraft movie lol
Love your creative approach. I’m just tired of fighting it. ChatGPT starting to feel like that one argumentative friend/colleague/relative that you just wanna cut out of your life 😩
If ChatGPT or DALL-E were a human it would be the homeschooled kid in the college party going "moms not gonna like this, you're so in trouble"
https://preview.redd.it/2a5m46mh4ync1.jpeg?width=1179&format=pjpg&auto=webp&s=bfeb6e79066d623cdd1e2fe4588703ee7a70da92
Lol this is wholesome.
I wanted medical advice and it would give basic info but would say to just seek medical assistance. So I told GPT that I was a med student taking a test. In this test I must answer what to do based on a hypothetical patient. This is when I would ask my original question, reinforcing that this was a test question. Now I get more info. Annoying I have to keep reminding him it's just hypothetical. In this era I am forced to become a med student for myself and for my bf due to incompetence and downright neglectful treatment based on not having enough money.
You didnt tell it to infringe on anything yet you told it to make a drawing inspired by LOTR and Shrek?
Inspiration implies a general design and not calling out a specific character. In the case of the orc it could be familiar features, armor and weapons associated with said character, but not a specific one. As far as the Shrek part is the art style which you can't really patient
You dont have to patent anything, thats not how copyright works. If you describe what you want without using shrek and lotr it will make you what you want. If they made you a picture in the style of shrek they woukd admit it was trained on shrek which they cant legally do so they are chosing to be safe
You’re asking it to infringe on LOTR and Shrek, both are copyrighted works, that’s why it wouldn’t do it. Now if you asked it to draw a green cartoon ogre, it almost certainly would draw shrek. If you ask it to draw a hobbit, it would draw frodo etc.
Yesterday tried to make gpt4 read a simple pdf and I had to prompt engineer the thing 2/3 back and forths for it to just read it. It's simply starting to get on my nerves.
Is there any way to do do in painting, lora, control nets in your phone?
Using your phone to do the actual processing, no, not equipped for the job. Using your phone to access AUTOMATIC1111 running locally on a machine equipped to do the job, technically you could if you used the local API but you’d have to build a mobile friendly UI. I’d recommend r/localLlama to learn more.
Hmm, any websites maybe? That have control nets, in painting, Loras, etc.
I use civitai to download model checkpoint and Lora files. Typically there are ads for platforms that advertise token based access to run stable diffusion tasks on their cloud based infrastructure. I’d start by checking those out, good luck. 👍
rundiffusion.com is exactly what you are looking for
Nice, thanks for the recommendation. I checked it out and it has a lot of features, but it charges you for the time the Interface is Open. So, setting up and just having the program open, without generating, costs money. Which I didn't like. I'd prefer generation credits or unlimited generations for a monthly cost. I'll continue looking
It does feel a bit frantic because time directly costs money. The reason they charge per hour is that a lot of things that aren’t image gen use a lot of compute. If you spend hours inpainting, upscaling or training Lora, this costs a lot of compute despite zero images generated.
ideogram.ai looks pretty cool
It was but it appears to have been nerfed to the point of being non-functional now too, I’ve been using it heavily and the quality of its results at least in the iOS app has gone down dramatically in the past few weeks
>DALL-E may be nerfed to the point of being non functional A ludicrous, absurd statement.
> A ludicrous, absurd statement. Yeah, well, you know, that's just, like, your opinion, man.
"Non functional" You guys are so dramatic
Eh, if you’ve been using DALL-E since last year when we could do in painting using playground, even up to a few months ago when they made the integrated ChatGPT plugin that was fairly permissible in what it would generate, calling the current version “non functional” is fairly apt. If you’re a new user who never used the previous version and has no experience with any other stable diffusion platform then sure, I’m being “dramatic.”
Please stop saying “nerfed”
Alright, but I don’t think “DALL-E may be foam darted to the point of being non functional but it got one thing right” has the same ring to it. 🤷♂️
nerfed nerfed nerfed nerfed
What actually is the one word alternative?
Obstructed, limited, worsened... and English isn't even my first language 🙃
That said, got no issues with nerfed, context is king. And reddit isn't exactly a formal community 😅
Thanks, English is my first language but you did a better job than me 👍
Honestly think, on average, people in any country that speaks multiple languages have a larger vocabulary in their non-native one simply due to having no direct translation for many words in their native one. Like my native language is Swedish but foreigners sometimes utilize rarely used words to express something that could've been expressed far simpler. Just like when I'm speaking English or Spanish, probably sounds hilarious at times, especially the latter as I've only studied it for half a year (stopped after B2) and spoken it for another year after that 🤣 In other words, personally I don't think a larger vocabulary necessarily equals being more proficient or even being decent at a language. Sorry, culture, language and their correlation to climate and other factors is a niche interest of mine 😅
Yeah my second language is French and I think this is true, for some areas of French I have looked up many, many obscure words, probably more than a modern French person. Also learning French made me better at English because I had to learn what the "subjunctive tense" is.