Тёмный

LLM hallucinations discover new math solutions!? | FunSearch explained 

AI Coffee Break with Letitia
Подписаться 48 тыс.
Просмотров 13 тыс.
50% 1

Опубликовано:

 

7 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 85   
@tahir2443
@tahir2443 8 месяцев назад
thanks for this excellent explanation. exciting times!
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@ceilingfun2182
@ceilingfun2182 8 месяцев назад
I'm not a mathematician, but I love mathematics. Wolfram Alpha did not make mathematics obsolete, and in my opinion, nothing will.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Thanks, this is a wise take! I agree, that we now can make progress on some old problems will not mean we are running out of problems, but that we are pushing the boundary of science to discover new problems.
@sanjose6010
@sanjose6010 8 месяцев назад
I really like your information delivery style ✨
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Oh thank you!
@terjeoseberg990
@terjeoseberg990 8 месяцев назад
This is how humans solve problems. We come up with a million dumb ideas, think a bit, and discard the 99.999% that are obviously wrong, and then verify the ones that aren’t. Every once in a while, we find a gem.
@CaptTerrific
@CaptTerrific 8 месяцев назад
Whoa... so they essentially ran a genetic learning algorithm with LLMs? That's rather clever (if not very very expensive :D)
@homewardboundphotos
@homewardboundphotos 8 месяцев назад
So... if we are back to brute force, what could be the implications of IBM's quantum roadmap they just released? I said right from the start, the hallucinations aren't a bug they are a feature. It's not a coincidence people who think very abstractly are more likely to get their wires crossed about things. You're never going to come up with a completely novel good idea without trying out 1000 ideas that don't work in the process. That's pretty much a law of nature in human intelligence, no reason why it wouldn't be a law of AI. You'll be able to make an AI that can always give you the right answer but has zero imagination and can do nothing original. And you can make an AI that will be able to come up with new ideas but 99.999 % of them will be bad.
@divineigbinoba4506
@divineigbinoba4506 8 месяцев назад
It'll be interesting if Quantum computers can instantly find the answer after the million generation. That'll be interesting.
@CodexPermutatio
@CodexPermutatio 8 месяцев назад
Hallucinations are not bugs... They are features! :]
@Mtaalas
@Mtaalas 8 месяцев назад
So infinite monkeys on typewriters? :D Is this how far we've gotten as a species? :D
@MaJetiGizzle
@MaJetiGizzle 8 месяцев назад
Ayyyy!!! One AI’s trash can be a human’s treasure! 😆 Thanks for the breakdown of the FunSearch paper! It pairs well with my morning coffee break. 😉
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@ricowallaby
@ricowallaby 4 месяца назад
Love mathematics and science. . . I remember my maths teacher when we worried so much about how solving problems, and he said "remember every problem contains its own solutions". How true. Cheers from Sydney '
@nunoalexandre6408
@nunoalexandre6408 8 месяцев назад
Love it!!!!!!!!!!!!!!!!!!
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Thanks!
@dahahaka
@dahahaka 8 месяцев назад
Creativity is a myth, there is nothing new you can create, only discover patterns in things which haven't been discovered yet.
@judgeomega
@judgeomega 8 месяцев назад
given the correct level of abstraction its simple and seeing it alone in isolation one would probably say 'thats not creativity', but then dropped back down into reality you would be amazed at how creative is!
@opusdei1151
@opusdei1151 6 месяцев назад
Wow Letitia, your videos are amazing :) thank you
@AICoffeeBreak
@AICoffeeBreak 6 месяцев назад
Thank You! Hope to see you here often. :)
@opusdei1151
@opusdei1151 6 месяцев назад
@@AICoffeeBreak dont worry I watch your videos all the time, also under a different account :) you one of my favorite ML youtuber
@vitalyl1327
@vitalyl1327 8 месяцев назад
I keep saying that hallucinations are the best feature of the LLMs. They are exactly the last missing piece of puzzle I needed to automate discovery and invention.
@hannesstark5024
@hannesstark5024 8 месяцев назад
Nice.
@MachineLearningStreetTalk
@MachineLearningStreetTalk 8 месяцев назад
Awesome video Letitia! 🤩👌
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Thanks, Tim!
@micbab-vg2mu
@micbab-vg2mu 8 месяцев назад
Thank you for the video:)
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@micbab-vg2mu
@micbab-vg2mu 8 месяцев назад
I asked GPT4 for creating a prompt for FunSearch this is what I got: "You are tasked with a challenging problem: [describe the specific problem or task succinctly, including any key details or constraints]. Your objective as a pre-trained large language model is to generate innovative and diverse solutions for this issue. Your outputs will be evaluated based on [define the evaluation criteria briefly, linking it directly to the problem’s objectives]. It is essential that your suggestions are innovative and feasible within the specified constraints. This process will be iterative. Based on the evaluation of your initial suggestions, the problem statement will be refined, and you will be asked to generate further ideas. This iterative process aims to enhance the quality of the solutions progressively. You are expected to utilize your extensive training and capability to generate a wide range of potential solutions. Approach this challenge with creativity, considering various perspectives. Begin with your first set of ideas and solutions for the stated problem" Usually during brainstorming of ideas additionaly I ask for randomization of presented solutions. LMMs have a bias toward the first answer - and assess it the highest. @@AICoffeeBreak
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@@micbab-vg2mu it's a start, but it needs code samples and specific problems.
@amigalemming
@amigalemming 8 месяцев назад
AI will not make mathematicians unemployed but Wolfram Research will have to implement some new techniques in Mathematica to stay up-to-date.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
They do have this ChatGPT integration, don't they? But sure, it's not nearly enough.
@ciragoettig1229
@ciragoettig1229 8 месяцев назад
This sounds broadly like brainstorming; you intentionally suspend criticism at first and just try to come up with many possibilities, no matter how off-the-wall they may be, valuing instead their volume and diversity, and only then in a second step you evaluate their quality. Wonder if other approaches to systematic thinking one can find in popular literature might have some lessions for ML research as well. I think I've heard of this kind of overgenerate + rerank approaches in NN papers before, though not specifically the added step of then prompting with the best ones and iterating the whole process again in that manner.
@theosalmon
@theosalmon 8 месяцев назад
Thank you for covering this. This is a great overview about something I was curious about. I hope Ms. Coffee Bean didn't expand her mind too much with hallucinations.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Thanks! Your last sentence made me laugh so much.
@DaNa-px7ol
@DaNa-px7ol 8 месяцев назад
Interesting times are coming in most fields, I think if we learn to use the advantages and understand our new roles, this cooperation could help us mitigate many shortcomings. You wouldn't be afraid of a powerful ally now, would you?
@DrJaneLuciferian
@DrJaneLuciferian 8 месяцев назад
I think an excellent point was made, that from the brute force of learning we experience emergent knowledge that is not always a logical continuation of that learning; lateral thinking is what we're particularly good at (for now). I think offloading the voluminous learning processes to the AI, and from that experiencing our lateral aha moments, a synthesis of mathematician and AI will speed up progress. Now, I live under no delusions that this will continue as-is. AI's will eventually begin to not only make lateral inferences from unaligned learning, but they will eventually self-realize the value of lateral thinking. I hope to be retired by then, lol
@amigalemming
@amigalemming 8 месяцев назад
Mathematicians searched for centuries for solutions of algebraic equations using radicals. They ended up with Galois Theory which explains everything but is practically of little use for solving actual algebraic equations. Maybe AI can help finding new solution formulas to algebraic equations?
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
I don't know the answer to this question. Maybe what is missing to answer it, is motivated people with lots of GPUs to try out?
@Laszer271
@Laszer271 8 месяцев назад
I like it. It basically takes AI which is known to provide false answers and which we cannot fully understand and makes it create something that is deterministic, something that we can verify and even understand. It basically addresses the biggest drawbacks of LLMs. Future won't be asking LLM to do stuff for you, it will be asking LLM to write a program that does stuff for you. I wonder how it would do on some problems that were always thought to be unsolvable by programs like classifying Imagenet images. It was always an example of why we need AI - because otherwise we would need to code unimaginable number of rules into our algorithms. Maybe AI could also come up with that unimaginable number of rules though. Otherwise, maybe it could write a program that not necessarily solves the problem but complements the AI in solving the problem? Like an AI that creates new ways to leverage another AI or even itself to solve a specified problem. Interesting times if you have access to such compute.
@ApPillon
@ApPillon 8 месяцев назад
Fit our correct frame of physics into a verifier and lets see where it goes
@automatescellulaires8543
@automatescellulaires8543 8 месяцев назад
What about packing coffee bins then ?
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@harumambaru
@harumambaru 8 месяцев назад
@@AICoffeeBreak The emojis are brilliant! Do you have them on Discord? How come I didn't find Discord link in description on RU-vid, Patreon and your website. Is this true that you have discord and patreon benefits?
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Yes, I have them on my Patreon- and YT-member-only Discord Server as well.
@DerPylz
@DerPylz 8 месяцев назад
@divineigbinoba4506
@divineigbinoba4506 8 месяцев назад
Damn. A million of a million is so much compute... Fun search is definitely not search.
@TemporalOnline
@TemporalOnline 8 месяцев назад
Almost like letting a million monkeys bashing a million keyboards for a million years and then checking if any of them wrote Shakespeare ¯\_(ツ)_/¯
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@BradleyZS
@BradleyZS 8 месяцев назад
What is imagination if not guided hallucination?
@earleyelisha
@earleyelisha 8 месяцев назад
How is this any different than generic algorithms?
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
It is a genetic algorithm.
@bipl8989
@bipl8989 8 месяцев назад
It's a selection of the fittest, but the candidate solutions do not appear to be the result of genetic evolution. Are they, or are they developed by other means? It isn't clear to me.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
They are generated by using the "previous generation" algorithms as prompts.
@earleyelisha
@earleyelisha 8 месяцев назад
@@AICoffeeBreak Yeah this is typical of evolutionary algorithms from decades ago. They don’t suffer from local min/max the way that gradient methods do. They don’t even need the LLMs to “solve math”. They are just taking already existing solutions and recapitulating within LLMs paradigm to benefit from the hype. “LLM hallucinations discover new math solutions” seems a bit misleading when the LLM wasn’t what made it work, but instead the evolutionary algorithm.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
I accept your point of a clickbaity title, though I argue (1) it is both the evolutionary algorithm and the hallucinations: Take the LLM outputs out of it, and you're back at a search without good solution candidates. (2) one needs catchy titles to convince the audience to watch this video instead of cute cat videos. 🐈 Once caught, the audience can be educated by the video that it is a smart search through the hallucinations and not just hallucinations. 🪤
@Neomadra
@Neomadra 8 месяцев назад
Thanks for this great breakdown! Are you also on Threads? I'm trying to get more AI contacts there as Twitter has become unbearable for me.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Sorry, Threads is not available in Germany. :(
@Neomadra
@Neomadra 8 месяцев назад
@@AICoffeeBreakIt is since a few days ago :)
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Wow, cool! Now I am on Threads, thanks to you! ai.coffee.break Let's connect.
@DanielSeacrest
@DanielSeacrest 8 месяцев назад
I would like to say i bleieve hallucinations are a feature and a bug. Neither do i believe the answer to making models more accurate (this is the bug part - hallucinations can make models innacurate) is removing hallucinations, but rather, aligning them with reality / what we percieve as truthful.
@minecraftermad
@minecraftermad 8 месяцев назад
Mathematics will be obsolete once we become omniscient, id say we have some ways till that.
@tatsujin1404
@tatsujin1404 8 месяцев назад
Would like to compare to evolutionary methods of code generation.
@BR-hi6yt
@BR-hi6yt 8 месяцев назад
Once AIs have been properly trained in mathematics they will go through it quickly solving all the problems in math remaining. Its only about following rules carefully.
@NeoKailthas
@NeoKailthas 8 месяцев назад
So they are using a process like RL to see if they can make these models come up with something humans didn't come up with.
@444haluk
@444haluk 8 месяцев назад
That's not hallucinations, that is randomness, you try to give meaning to it
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
Yes, there are people advocating for the term "confabulation" instead. Since LLMs are not grounded, everything they say is confabulated, it just sometimes happens it corresponds to reality.
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
But it is nonetheless an established technical term nowadays. Just like "foundation models", which of I do not like much.
@Neomadra
@Neomadra 8 месяцев назад
@@AICoffeeBreak What exactly do you mean by saying that LLMs are not "grounded"? What would one need to do to "ground" an LLM?
@AICoffeeBreak
@AICoffeeBreak 8 месяцев назад
@@Neomadra To ground would mean to have them see the same world context as you do at the moment you are asking something. A simple, short question in the form we are usually formulating to ChatGPT, such as "What should I wear this evening?" is underspecified in all aspects: where do you live and what are the customs geographically? What is culturally accepted when you mean by evening (the spanish 10PM or the German 7PM)? How do you look like and what would fit you? What I mean by not grounded is that they do not share context with you.
@Neomadra
@Neomadra 8 месяцев назад
@@AICoffeeBreakAh that makes sense. Thanks!
@naromsky
@naromsky 8 месяцев назад
This is research?
@poopytowncat
@poopytowncat 8 месяцев назад
You are beautiful but your lipstick is excessive and distracting.
Далее
MAMBA and State Space Models explained | SSM explained
22:27
кого отпустят гулять чееек
00:53
Bike vs Super Bike Fast Challenge
00:30
Просмотров 4,4 млн
Has Generative AI Already Peaked? - Computerphile
12:48
Solving Math's Map Coloring Problem Using Graph Theory
9:04
Transformers explained | The architecture behind LLMs
19:48
DeepMind’s AlphaGeometry AI: 100,000,000 Examples!
7:49
The moment we stopped understanding AI [AlexNet]
17:38
Why Does Diffusion Work Better than Auto-Regression?
20:18
кого отпустят гулять чееек
00:53