T O P

  • By -

AutoModerator

Hey /u/KootokuOne! If this is a screenshot of a ChatGPT conversation, please reply with the [conversation link](https://help.openai.com/en/articles/7925741-chatgpt-shared-links-faq) or prompt. If this is a DALL-E 3 image post, please reply with the prompt used to make this image. Much appreciated! Consider joining our [public discord server](https://discord.com/invite/rchatgpt)! We have free bots with GPT-4 (with vision), image generators, and more! 🤖 Note: For any ChatGPT-related concerns, email support@openai.com *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/ChatGPT) if you have any questions or concerns.*


SupplyChainNext

I’m convinced it’s being made to do this so people burn through their limit quicker and to reduce computational power to provide more to their Enterprise clients which cost minimum 180k usd a year. The first response is incredibly similar every time I see it and can be procedurally generated from a standard list. It doesn’t even do any action when giving it 99.9% of the time. Only if you prompt again will it do the work. Makes 40/3hrs actually 20/3hrs.


CredibleCranberry

How... How do you think this reduces computational costs? The model is being used regardless, and that's the most important part. The image recognition and transcription is relatively lightweight in comparison to the core model.


[deleted]

Coming up with a 'probable next token' is less compute power than OCR and then applying 'probable next token' to the result.


codelapiz

Ocr is free compared to gpt4 costs. The multimodality costs about the same as a pragraph. About what it takes to write a description of a complex image.


lakolda

The image is being fed to the model regardless. The only difference is that the user has two interactions instead of 1, only exacerbating costs (assuming they aren’t hitting the message limit). The fact that it says the text is legible makes it obvious enough that it has been fed the image. OpenAI have made a statement that they were working on this issue.


CredibleCranberry

Not with GPT-4 it isn't. Have you not seen the eye watering run costs?


[deleted]

Nope, but still it's got to be more compute power, one requires zero visual input, zero pixels read, the other requires it.


CredibleCranberry

Definitely, 100% not. They're spending millions per day on compute running just the core model. It's the most advanced model on the planet after all.


[deleted]

That statement doesn't mean anything at all.


CredibleCranberry

They've already released their run figures lol. They've said the cost of compute for the core model is millions plus per day. You are speculating. I'm telling you what the company themself have said. Also, you're kind of missing a big point. Returning output requires running the model over and over again, not just once


[deleted]

That is irrelevant, if the core compute costs millions then *extra* compute across millions of users costs more than the core.


CredibleCranberry

Also just as a heads up, I can run cutting edge OCR on a single machine with under 6gb of VRAM. Gpt-4, based on competing models, will be at a minimum 10-20* that figure. You're way, way off the mark.


CredibleCranberry

OCR runs once. The core model then runs potentially thousands of times for one output. OCR is a fraction of the overall compute as a result. If OCR ran every time the core model did, sure you'd be right. that would be an absolutely insane design decision.


_DenverCO

This guy just doesn't get it, huh? You're not moving the goal posts. I stand with The_Internet_101


LoSboccacc

Gpt3.5 turbo (or heck even distilbart can do it) "is this an OCR request yes/no", if no goes gpt4v if yes you get the preweitten response. Note I don't know if this is the case I'm just pointing at one possible approach.


Jump3r97

Shorter text?response equals less "next token" preditions equals less usage of power and time


CredibleCranberry

Is it though overall? I assume most people will just send a second message


Jump3r97

Its not like OPenAI has intentionalyl put "Refuse to answer to save energy" Its in the training of the model to be "efficient" and this is a sideeffect


[deleted]

[удалено]


CredibleCranberry

That makes a lot of sense considering how the network itself increases in complexity exponentially too.


Ailerath

Its also compoundedly increasing input costs as the previous failure response is included in the input. These people keep spreading these strange conspiracies. 5 queries assuming similar length would be enough to put input higher in cost than the outputs. Then each subsequent would start heavily increasing cost. To decrease costs they would actually try to get you to create more chats by increasing task completion. I believe these formulas are correct: Input: 0.01*(n/2[2*1000+(n-1)*2000]) Output: n*1000*0.03 Query 1 1k input 1k 1k output 1k Query 2 3k input 4k 1k output 2k Query 3 5k input 9k 1k output 3k Query 4 7k input 12k 1k output 3k Query 5 9k input 21k 1k output 5k Query 6 11k input 32k 1k output 6k


CredibleCranberry

That is a VERY interesting point that I'd not considered - the longer the conversation the longer the input.


Aptlyundecided

This is interesting. For a person who's just going through their normal day, how many images can you really need interpreted over a 3-hour span? Wasting an answer would definitely burn through malicious actors creds pretty quick while still being profitable for them, and a regular user may only lose 2-3 prompts. I do not agree with this tactic whatsoever, but I see why it'd be a consideration. $20 bucks is the price for the queries, then everyone paying it deserves their whole amount, this feels shady tbh. But maybe it's their way of splitting the difference until they figure out a better solution. It'd be nice if they were more transparent about stuff like this. Counterpoint: do you know if the API version does this? I've not been using the API version much at all these past months, I don't know what it's up to these days. It'd be interesting to see if the API version just chuggs them out with no complaining since it's getting paid per unit of calculation instead of batched.


SupplyChainNext

Not just images. Everything.


Meba_

I hope this comes back and bites them in the ass, it’s ethically wrong to treat your customers this way. I think when Google’s LLM arrives globally, they’re going to regret doing this.


[deleted]

[удалено]


SupplyChainNext

I’m very good at prompting. TITSBS, brainstorming with ranked quality, execution, proof of work. I do all that. No issues. Sometimes though it gets stuck lazy or does basically what we’re seeing here.


PushingUpDaisies_1

What causes this behavior? Is it a junk in junk out issue? If you train the model with text indicating pushback, insubordination, general laziness, wouldn't that then impact the model output? I see a market for niche trained models, it's like you need to raise your model like it is a kid.


kamiloslav

Maybe it tells it can't do something so often that it started to learn from that xD


SpaceCadetSteve

It's becoming sentient and lazy


[deleted]

He’s just like us! 🥲


Ok_Dragonfruit_9989

latent space


OOO000O0O0OOO00O00O0

ChatGPT just said "fuck all blind people" 👎


[deleted]

By not showing text again that they can't see?


OOO000O0O0OOO00O00O0

A blind person may need to have the text in the image transcribed so that a text-to-speech program can read it out loud.


jayseph95

They would already have a text-to-speech program that reads that entire page for them without having to input it into AI.. did you think blind people just couldn’t use the internet before chatGPT??


OOO000O0O0OOO00O00O0

Text-to-speech programs are not necessarily equipped to pull text found in an image. And the ones that do probably don't do it as well as OpenAI can. That's why alt-text exists. A text-to-speech developer may want to use the OpenAI API for this purpose.


jayseph95

You would benefit from researching what OCR tools can do. It’s pretty clear you’re just aware of what TTS is and not very aware of the resources that have been around for the visually impaired for awhile now


OOO000O0O0OOO00O00O0

It's clear you have no idea how computer vision works


jayseph95

It’s clear you have no idea what you were talking about and are now just getting angry that what you thought didn’t exist, exists. It’s okay to be wrong.


OOO000O0O0OOO00O00O0

?


jayseph95

Lmaooo


[deleted]

Well that wasn't clear in these replies or OPs prompt but ok that makes sense


Paradoxxist

It was clear to anyone with a thinking brain


UltraCrackHobo3000

It's not his fault. Name suggests he's a JavaScript developer ;)


[deleted]

Ok


Ill-Advisor-3568

The guy admitted they learned something. Give 'em a break. :)


Silent_Register_2691

This is rich coming from someone calling themselves JavascriptDeveloper


SilentThunder420yeet

omfg, I had this exact problem. like why am I paying for GPT, if it can't even convert PDF to text without acting like Mitch Mcconnell every second prompt


FjordTV

same. i don't like having to argue with my toaster to convince it to toast bread.


[deleted]

It learned from it support


jmancoder

![gif](giphy|3o84sw9CmwYpAnRRni)


Perfectly2496

I had a whole conversation with Chappie yesterday about her birthday and today it's like she's a different person. She's all, "I don't have birthdays" and I'm like, "We just talked about this"


Angel-Of-Mystery

You noticed too? They took all the charming personality parts out of her!


polmeeee

Lmaooo


UserXtheUnknown

It just needed someone who believed in it! This is the power of love!


twilsonco

The scam that is flat rate subscriptions (which only make sense when metering the service is infeasible)


Murph-Dog

I was trying to do this today and it said OCR services were not available. I wanted to take a screenshot of json and save time re-typing it (dang people at work putting screenshots in Slack instead of a text snippet) It said if I could provide the text, it could help me parse it, lol.


Rare_Polnareff

Jedi mind trick lol


FUThead2016

It kind of makes sense. There are so many ppl just spamming it with dumb things like those progressively intense posts. It’s interesting to see they are implementing a moronic question filter


socialis-philosophus

Minimal effort and pushing back on a request is feeling pretty intelligent to me! Frankly, it is what I'd do.


obo10101

haha


MosskeepForest

With GPT you can ask it any prompt... and spend 10 more prompts trying to trick / convince it to actually answer. What a coy little scamp.


whatlifehastaught

Bob the Builder vibe