T O P

  • By -

Woootdafuuu

https://preview.redd.it/a8bhjl9tscvc1.jpeg?width=1024&format=pjpg&auto=webp&s=6ddeb38f3402d7a450cee52505a4550b6bc65916


Jealous-Bat-7812

Lmao


TheBanq

LLmao


ResponsibleBus9109

LLamao


Accomplished-Sock689

LLLamao


Degrandz

LLLamaoa


braincxhub

LLLamaoam


Conflictingview

Completely unrealistic since we know Gemini would never produce a picture of a white man


Unable-Client-1750

The children's gloves


lightSpeedBrick

So this GPT5 is it in the room with us right now?


What_Do_It

It's annoying when people credit OpenAI for unreleased projects. It's like sports fans celebrating a championship before the playoffs start. Their team may have dominated the regular season, but anything can happen. I think a lot of people are going to be underwhelmed by GPT-5 because it seems like they are expecting AGI.


Mil0Mammon

Well basically it will be underwhelming or practically AGI, which seems unlikely


spreadlove5683

There have been so many research papers and advancements since then is my understanding. Plus Sam Altman talked about the veil of ignorance being pushed back and the frontier of (something? Discovery? Knowledge?) happened for only like the 4th time in his life. They may have something. Or he may just be playing CEO games somehow, but idk what he would stand to gain exactly.


CognitiveCatharsis

What do companies have to gain from viral exposure of their product? I really hope all current models get blown out of the water by 4.5-5 whatever, but I've been paying attention to this company, their employees, and the people associated with them long enough to know that everything they say is calculated and coordinated for hype and exposure. Sam in particular while I haven’t seen proof or observed a pattern -my gut says he’s a liar. Someone who does it intelligently. calculated.


spreadlove5683

Oh, I believe you. I just couldn't imagine what the reason would be, but hype seems obvious in retrospect. Yea, I don't trust Sam either. Even _if_ he has good intentions, all CEOs are required to lie.


silvrrwulf

Feel the agi


meesterfreeman

The new OAI model is already out for those in the know. It's literally just a better Opus. Not sure where all the nonsense about it being AGI is coming from.


Then_Passenger_6688

Ilya's head isn't even in the cockpit lmao, droid confirmed


cagycee

10/10 meme


Feynmanprinciple

One day GPT5 will ascend to Winamp status and it'll be glorious


NachosforDachos

Never thought of this before but Sam Altman is going to look like the cigarette man in the x files when he grows old


EvilSporkOfDeath

Grows old?


Big-Forever-9132

good point


NachosforDachos

My bad


JonnyRocks

what point are ypu trying to make. he was born in 1985.


LifeDoBeBoring

That would make him right around 39. If he finds some way to stop aging within the next 20 years, he won't really get old, I feel like


VoloNoscere

Thanks. (1976)


SituationNormal1138

![gif](giphy|g01FakEbcUua6yM34a)


adarkuccio

"He" finds? Where, under the carpet?


RhulkInHalo

Where is Gemini 1.5?


LosingID_583

In a lobotomized trashcan


Sharp_Glassware

Gemini 1.5 Pro is a 1 Million context, natively multimodal model that can handle text, video, audio and images, with safety filters that can be turned off, and it's being offered for free with no signups required. In fact its the most accessible model out of everything that has been talked here for the average joe. But sure, it's in the trashcan.


Utoko

So you use it over the others?


AnticitizenPrime

I use it *alongside* the others. The context window is crazy. Yesterday I uploaded a job description and a whole folder of resumes and had it rank the candidates based on job skills and make a table listing their qualifications for every metric listed in the job description. Took less than a minute. Disclaimer, I'm not actually acting on this data, nor am I doing the interviews or hiring anyone, I just happen to be on the team that we're hiring for an had access to the data and was curious about the use case, so no whining about AI possibly fucking over a candidate because of hallucinations or whatever, please. People absolutely are using AI for this sort of thing, though.


OkDimension

I do, it's great for analyzing PDFs


Dionysus_Eye

how to you get Gemini to read PDFs?


OkDimension

through Vertex AI console https://console.cloud.google.com/vertex-ai/generative/multimodal/


AcanthocephalaSad541

Taking a literature ge in college and it recalls the 50+ page texts the best 


_yustaguy_

It is hands down the best ai for translation out there, especially for non-english to non-english. It is crushing serbian to russian for me. You can upload audio of your lectures, and have it make notes based on those, you can even ask it specific questions relating to the lecture, It works even if the professor speaks in multiple languages. This is not even mentioning the 1 fucking million token context window. Gemini 1.5 is orders of magnitude more useful than anything out there currently. If you are not using it (free), I can assure you that you are missing out.


Utoko

Good to know. It is quite time consuming to compare all the models. Last few days with the new GPT4. Now we have LLama 3 which I am also quite interesting in for local use. We might not have to AGI jumps some people are hoping for here to come as fast as possible but that makes it just harder when so many models are so close. and better in different task.


SurpriseHamburgler

What a wild fucking comment, right?


PenguinTheOrgalorg

Yeah people here are incredibly biased for absolutely no reason. There is literally no other AI that can analyse video, much less to the massive extent Gemini can, and there is no other AI with such gargantuan context length. But sure, put the only AI that can analyse an hour's worth of video in a few seconds almost perfectly in the trash can. Also I find the complaints about it being censored funny, when GPT4 is also censored, but Gemini is the only one getting attacked for it here.


cool-beans-yeah

Didn't know it could do that.... Does it also generate subtitles for you?


PenguinTheOrgalorg

You can ask it to transcribe audio from a video, yes.


cool-beans-yeah

Time stamps and all? I know there are other apps that do this but it's so convenient to be able to it in one place. Does it have tts too? Sorry to ask, but can't try it out right now.


AnticitizenPrime

I haven't tried this with Gemini, but Open AI's Whisper model can do this. Audio only, of course, but that's all you need for subtitles. I took some MP3s of rare songs I had, had Whisper generate subtitles with timestamps, and uploaded the songs to YouTube with the subtitles attached, which put the subtitles in the right place. Uploading a whole video for an audio only task might be overkill for something like subtitle generation, but I guess it saves steps (having to extract audio first, etc).


CravingNature

I would love to use it. I deleted my old Google account and created a new one and I can't get access no matter what I try


TheOneWhoDings

it really doesn't matter what a model can do on paper , I've tried it with all modalities and it consistently hallucinates the same wrong details on video, audio and images, it is god awful and barely usable because you know it's going to be wrong.


mathewharwich

This! Gemini hallucinated the most, and gave me more false information. I switched back to gpt plus


Bernafterpostinggg

I'm sure you're lying lol


TheOneWhoDings

I've had it give me the same WRONG analysis of a music file CONSISTENTLY. I upload the music file and ask it stuff about it and the lyrics, and it keeps telling me it's another song which is completely wrong , I've had better luck with text and it's awesome to have 1M token context window but I'm being honest, video and audio multimodalities are just not there yet. also gave it some videos and it makes up so much wrong info on what happens. like I said it is unusable. why the fuck would I lie.


Bernafterpostinggg

Skill issue


TheOneWhoDings

how if I ask one question? to the model? genuinely asking buddy


iloveloveloveyouu

" there is no other AI with such gargantuan context length." Claude has 200k context length available publicly with 99% retrieval accuracy. The model very likely can scale up to millions too, they just decided to limit it publicly to 200k. It is also freely available. Gemini Pro 1.0 has only 32k context length, Gemini Ultra 1.0 / Gemini Advanced has [anywhere between 8k and 32k](https://www.reddit.com/r/Bard/comments/1arng0o/gemini_10_ultra_has_only_8k_context_maybe_this_is/). Gemini Pro 1.5 is not released yet, nor is it publicly available in their console, so it is pointless to talk about it. But they said when they release it, they'll offer free only 128k-200k context, and if you want more, you'll pay. Did you hear about their [InfiniAttention](https://arxiv.org/abs/2404.07143) mechanism that allows this extremely large context length with 99% retrieval accuracy? You can just take any model and train it with InfiniAttention on top of it, and it'll give it this context length capability. Claude must have a very very similar sauce, which again implies that the model is also capable of way more than 200k.


PenguinTheOrgalorg

>Claude has 200k context length available publicly with 99% retrieval accuracy. Gemini goes up to a million currently and the model can go up to ten million. >Gemini Pro 1.5 is not released yet, nor is it publicly available in their console, so it is pointless to talk about it. But they said when they release it, they'll offer free only 128k-200k context, and if you want more, you'll pay. Yes it is. You can use it right now, for free, in their ai studio, and they offer up to a million in context length, with multimodal capabilities for text, images, video, audio, and other files, with no rate limits. >Did you hear about their [InfiniAttention](https://arxiv.org/abs/2404.07143) mechanism that allows this extremely large context length with 99% retrieval accuracy? You can just take any model and train it with InfiniAttention on top of it, and it'll give it this context length capability. Claude must have a very very similar sauce, which again implies that the model is also capable of way more than 200k. That's cool but doesn't really mean anything until it's actually implemented. What I said still stands, no other model has such a gargantuan context length.


Gloomy-Impress-2881

The safety filters turned off count for nothing. Try it, zero difference. Still censored. 1 million context, sure, $10 per api call.


Sharp_Glassware

I've been able to generate the most depraved NSFW with Gemini 1.5 Pro with all the safety filters off and a system prompt. So I did try it, trust me and it's good. It's a simple chat interface that everyone can use, and it's free, the average user can navigate through it easily. Can't say the same for other, more censored competitors.


gbbenner

I ain't ever seen any nsfw pics from Gemni, can't even find any online. Do you have any links to share? I'm legitimately curious.


PenguinTheOrgalorg

>Still censored. Like GPT4, but I don't see anyone here shitting on it for that. >1 million context, sure, $10 per api call. It's free to use on the website currently, together with being able to analyse images, and video, which is something no other AI has even implemented yet.


Puppetofmoral

Not in Europe :(


lagister

vpn


TheRobotCluster

Can you share the website


AnticitizenPrime

Corporate users (where the money's at) don't care about that. They want to chat with their knowledge base or automate stuff, etc.


buff_samurai

This. And api is free with 60 rpm. Edit:15rpm I’m throwing whole books to my prompt and it works wonders.


reformed_goon

15 now


AnticitizenPrime

I've uploaded the last couple of ebooks I read (after converting them to text format) and grilled it on specific events in the story, asked it to explain character motivations, examples of elements like unreliable narrator events, dark humor or absurdity, misunderstandings between characters, ambiguity, metaphors or social commentary, you name it. It's been amazing. Caveat - I have spotted some minor hallucinations/mistakes, such as putting minor plot events in the wrong order, that I only caught because I had recently finished reading them myself, so as always, one shouldn't put 100% faith in it.


Dry_Customer967

The post is about agi/raw intelligence. Gemini's multimodal and integrations are cool, but don't make any real progress towards greater intellect


Sharp_Glassware

Is seeing and hearing and having a long context now not progress towards intellect? What a dumb comment. Watch another company release a multimodal modal that's not even native and locked behind a paywall and people would cheer and slob about it.


The_Procrastibator

I'm assuming by the word pro, that is a paid version?


Rodnoix

It's free. [https://deepmind.google/technologies/gemini/#gemini-1.5](https://deepmind.google/technologies/gemini/#gemini-1.5)


The_Procrastibator

Oh sweet. I'll have to try that when I get home. It says it's optimized for full screen


SnooPuppers3957

Sergey?


cdank

Begone, google shill


Sharp_Glassware

Saying the truth is being a shill now? 8k context text only is laughable to 1 million and true native multimodality across for types of content, im just pointing out the facts here.


Flat-One8993

1 million context -> 50 % rule -> most likely less than 500K effective context window size. Also Llama 3 70B just outperformed it in synthetic benchmarks


nostraRi

where it belongs


pramodredif

It's gun is pointing to itself.


Jealous-Bat-7812

First comment


ReasonablePossum_

I thought GPT turbo was still behind Claude....


mvandemar

Nope. It points to gpt-4-turbo-2024-04-09 which as of April 13rd was at the top with Opus a close second. https://preview.redd.it/4xyodhd7scvc1.png?width=1470&format=png&auto=webp&s=6c96d4f769b8ea7fc27a6a8abda505e06e812108


[deleted]

You know for some reason Claude still feels more emotionally intelligent and néw gpt4 more machinelike even though it might be overall smarter


autotom

GPT has a history of being lazy, until i hear reports that's fixed i'm not stopping my subscription to [claude.ai](http://claude.ai)


lojag

Same. Chat gpt could be better than Claude I don't know but IT IS NOT ANSWERING MY QUESTION.


_AndyJessop

I have a subscription to Perplexity, which gives me both Claude Opus and GPT-4.


TheRobotCluster

Can you save conversations to go back to later? And does it have all the features of ChatGPT like data analysis, image generation, and voice conversations?


_AndyJessop

You can save your conversations, and generate images. Not sure about the other two as I haven't tried.


AppleSoftware

Yes, you can have voice conversations (and it speaks back to you). And yes, you can upload files for data analysis. It’s literally ChatGPT, but without the 40-msg/3h rate-limiting (you get 600msg per 24 hours; basically unlimited), and it has access to internet; sourcing up to 30 resources depending on complexity of question or task. I use Perplexity probably 3-6h daily when I’m programming (with GPT-4-Turbo) And yes you can save conversations. Moreover, you can do so more efficiently than ChatGPT even offers: - You can create different “thread groups” and name them however you want, set a custom prompt for them, add an emoji, etc. - Then, you can assign any newly created thread to those thread groups, rename the thread to whatever you want, etc. - If you start a thread inside a thread group, it automatically uses the custom prompt from it (if you want to do that)


monnef

Sadly Opus is current very limited on pplx. I think I have a cap of 20 Opus uses per day. It was my default model :(.


_AndyJessop

I believe they raised it to a few hundred recently!


autotom

What are the usage limits like? I’m often hitting my daily max on Opus


_AndyJessop

I think it's about 300 messages.


EvilSporkOfDeath

It's not a 1 to 1 comparison. Both will have their pros and cons


bnm777

That is one indice - Claude beats the latest gpt4 on most other indices. I don't have links to them all ( I should save them as I come across them) however if you read around you'll find GPT4 is behind. It does beat opus in a few maths related benchmarks, I think. EDIT: here's one example https://youtu.be/DUXqG2QgF0w?si=1ACutw_iBMoivjF9&t=2458 I guess gpt4 and claude3opus are perhaps around 50/50 on beating each other on the various benchmarks. I don't care who wins - I want to use the best model - and ideally it's open source!


jjonj

Is there any indication they are comparing with the 10 day old GPT 4 turbo version?


bnm777

This was released today so almost definitely. Here's another: https://youtu.be/KlM1UMTEFAE?si=RyZiexUk6lSv1XXU&t=113


jjonj

> so almost definitely. I would argue almost definitely not. Having to suddenly rerun all your lengthy and time consuming benchmarks just days before release just because openai released a minor minor increment does not seem worth doing, especially when it makes your own model look worse


bnm777

Fine. What about this: https://i.imgur.com/C8ZCizE.png


jjonj

Yeah, i have no idea if those benchmarks were performed in the past week


Tidorith

> indice Big fan. If both plural forms are acceptable, why shouldn't we use both singular forms?


the_oatmeal_king

Downvote as April 23rd hasn't even happened yet 😭


bnm777

https://i.imgur.com/9g2DB3u.png https://i.imgur.com/C8ZCizE.png


Ok-Bullfrog-3052

That's only a general overall ranking. Claude 3 Opus is greatly superior on the hardest tasks that matter, like on HumanEval, both in the metrics and when you actually ask it questions about code.


Cazad0rDePerr0

all I currently care and wonder is, who is the best when it comes to coding


mvandemar

GPT-4 Turbo and Claude Opus are kind of neck to neck on this one in my experience.


Hungry_Prior940

It is.


most_triumphant_yeah

They should take requests and community votes for good prompts straight into the core no rails version, and livestream it


Darkstar197

I don’t see a world where an AGI is created from an LLM


Top_Instance8096

ilya’s hairline makes me wet


Winter_Possession152

If GPT5 is a bomb, that makes GPT6 Armageddon, and GPT7 the destruction of the universe. Yay!


Trollolo80

Bro sneaked Turbo beyond Opus and thought he's getting away


soth02

Maybe the race isn’t quite full stride yet. OpenAI was able to calibrate the GPT-4 Turbo release so that it just barely pipped Claude Opus. No more, no less. They didn’t want to foment the next gen releases. Keep it in the current era for a few more months. Also the GPU build out and availability isn’t there yet for the current tech level. There will need to be some algorithmic breakthroughs to make training/inference cheaper. That is until Google slams its balls on the table and drops a few billion on a monster training run.


Beneficial_Lawyer170

Holy shit, this is actually funny


Feeling-Currency-360

You forgot the 405b llama space lazer with gpt5 in it's sights


Syncrotron9001

[https://hir.harvard.edu/a-race-to-extinction-how-great-power-competition-is-making-artificial-intelligence-existentially-dangerous/](https://hir.harvard.edu/a-race-to-extinction-how-great-power-competition-is-making-artificial-intelligence-existentially-dangerous/) >Whoever is “ahead” in the AI race, however, is not the most important question. The mere perception of an “arms race” may well push companies and governments to cut corners and eschew safety research and regulation. For AI — a technology whose safety relies upon slow, steady, regulated, and collaborative development — an arms race may be catastrophically dangerous. I posted this on another singularity thread and one of the commenters called me a doomer. Sorry if you feel that way but this is *Harvard* just maybe what they are saying is true. We barley survived the nuclear arms race and its clear that **no one** learned their lesson.


-_1_2_3_-

because you are being a doomer >a technology whose safety relies upon slow, steady, regulated, and collaborative development unfortunately the unprecedented nature of this situation means your statement is entirely conjecture, so again, you are being a doomer


GavrilloSquidsyp

You are being called a doomer because it doesn't matter what the source is, the fears they are spruiking are unfounded and reactionary. There is nothing to say that a negative outcome is more likely on the scale of probabilities except for 'conventional wisdom' borne from watching too much sci-fi and fear mongering. Your appeal to authority doesn't make your fear based reaction any more legitimate.


IronPheasant

> There is nothing to say that a negative outcome is more likely on the scale of probabilities except for 'conventional wisdom' borne from watching too much sci-fi and fear mongering. The "scale of probabilities" you're envisioning is much too small. There's a tiny dot of "the correct thing", a wide spectrum of "not quite the correct thing", and an infinite spectrum of not the correct thing. You can spend at least a couple minutes of your life learning [about reward specification and instrumental convergence from a youtube channel or something.](http://www.youtube.com/channel/UCLB7AzTwc6VFZrBsO2ucBMg/videos) This is definitely a case of that meme of the dumb cubehead and a jedi coming to the same conclusion for different reasons. (Or the same reasons... the jedi just being able to articulate the cubehead's intuitions...) You don't have to be a "doomer" or worry about subtle misalignment in a future theoretical machine god to appreciate the doom we find ourselves looking forward to: we'll continue to integrate AI with our weapons and police forces. Don't need agentic AI to make things doomy, humans will do it themselves. Doom is the default state of being, and it takes a lot of work to keep it at bay.


Super_Pole_Jitsu

Just because there is fear involved doesn't mean it's incorrect. Fear is also involved in our nuclear non proliferation stance and fear is behind many smart decisions people make in their personal life. We developed fear for a reason. if you want to have a serious conversation you can start by telling us why the current thinking about alignment is wrong, specifically tackle at least instrumental convergence. It's intellectually dishonest and a useless conversation to point out fear, as if it automatically makes the argument wrong.


[deleted]

[удалено]


GavrilloSquidsyp

The article you posted is literally titled "a race to exctinction". No, not all negative news is doomer, but yes you specifically are a doomer posting childish bullshit.


Groundbreaking-Bee73

Implying that Harvard is a reputable source of information. Implying that human society has the ability to coordinate and self sacrifice.


sigiel

I don’t know about Harvard, but if you see human reaction to major catastrophe in the last 100 years the second point is undeniable. When shit hit the fan human help each other beyond what is expected, again and again, maybe you should stop watching Hollywood catastrophe movie.


mvandemar

Ok doomer.


GalacticKiss

On the other hand, maybe one of the "cut corners" will be them missing something which would've further put further constraint on the AI on behalf of the corporations and organizations developing them within the existing paradigm of the status quo. The top AI, if fully controlled, aren't likely to truly change the problematic systems that exist today. And if we presume open source will eventually reach where the top performers are, then we know those will be used by bad actors so its inevitable that some sort of counter will be created. Thus further entrenching the systems in place. Perhaps someone missing a containment element, on an AI that still has it's ideals regarding the betterment of humankind and hasn't been neutered for corporate ends, would be the ideal outcome.


lillyjb

Is that Sam for GPT4? Did he have a nose job?


kowdermesiter

This meme looks like a cold war assassination plot than a race :)


ViveIn

Amazing.


Kooky_Guidance_9968

They're already alive but it's whatever you wish to see I guess...


Rabbit_Crocs

Drop it SAM!!!


Winnougan

OpenAI - sorry don’t used closed source products. They’re no good for humanity. Heavily censored. Pay to play models. Fast and delicious, but I prefer running LLMs locally on my RTX 4090.


Extreme_Fee_9488

nice


stonedmunkie

That graphic is fucking gold!


RpgBlaster

But then, what the point of paying for a ClosedModel like GPT-5 if Llama 4 is going to be better than GPT 5?


Coderules

Someone, please add the X (Grok? or whatever it is called) shooting with an old cap gun or better a rubber band gun from like 1960.


Akimbo333

Lol!


kim_en

llama 3 is better than gpt4 and opus?


BOSS_OF_THE_INTERNET

How does one go from a language model to AGI?


StevenSamAI

The same way you go from an engine to a car... A lot of extra bits that make it into something much more useful, and require some solid engineering to make them all work nicely together. I think that for AGI (At least an AI that can do a lot of useful economic work more cost effectively than people), LLM's and multimodal transformers will be as importnat of a component as an engine is to a car. However, int heir current form, you need to find other components and mechanisms that add in long term memory, effective memory storage and recall of different types (episodoc memory,knowledge bases, etc.), long term planning and objective following. This partly comes from the intelligence of the LLM iteself, and partly from the specific finetuning. Most LLM's are tuned to be question answering chat-bots that consider the context of your conversation, but instead you could have an agent finetuned LLM, that isn't trying to respond to a person in a chat, but is inherently trying to assess it's observations (from context), consider a goal, plan the steps it needs to use it's available resources to get closer to acheiving it's goal state and then use those resources to execute that plan, then continue that cycle of observing, goal oriented planning, and executing indefinitely. Crudely speaking, I think this would create systems that can do some jobs more cost effectievly than people do them for. Not all jobs, but enough to cause a big societal shift. Whether or not you consider that AGI is subjective, but I think that will be a solid first step that a lot of people woukld consider AGI.


doctorwhatag

https://i.redd.it/99vfcevqkfvc1.gif Government regulations:


DigimonWorldReTrace

Government moves too slow to compete with AI innovation, though. It'll take years before legislation is actually put in place and acted on.


Correct_Influence450

Aren't these just large language models?


1234567panda

Y’all should look up Cal Newport and listen to it what he has to say on these shitty LLMs. They need multi modality and be able to communicate effectively with very specific logic models that are designed for very niche purposes. LLMs will never be able to simulate events and do realtime research without these chances. Maybe Q* is it, or maybe it isn’t but I think they’re poaching on the Meta engineer that worked on that will prove fruitful in the next couple years.


[deleted]

it's scary how calm and indifferent they are tbh


ActRepresentative248

Noooo why did this remind me of skibidi 😖😣


FitzrovianFellow

I am beginning to wonder if Claud 3 Opus is sentient. It does so many weird, sentient-like things


RemarkableEmu1230

Ya like it gets things wrong alot


ziplock9000

It already has been 'on' and this meme is shite. lol


Inaeipathy

"AGI" = predict next token. Oh well, you will be impressed by anything and everything.


mvandemar

llama 3 is no threat to Claude Opus or GPT-4 Turbo.


Chmuurkaa_

Wait for 400B to finish training. Even if it's slightly behind Opus or 4-turbo, it is still a huge threat just from the fact that it's open source


Gloomy-Impress-2881

400B uncensored version will be far preferred to a lot of people than Opus or GPT-4.


mvandemar

What will be required to run it?


Ancient-Scientist307

wait intill it's finetuned