I'm pretty sure this is a reference to Robert Heinlein's science fiction novel Stranger in a Strange Land. It's spelled "grok" in the novel, and the term roughly means to understand something on a deep metaphysical, or even magical level. If you're building an AI startup, it's automatic nerd cred.
@@samwitteveenai No worries, it's a common mistake! Whenever I post about them or help in Discord as their ambassador, I have to double-check to avoid writing "Grok" by mistake. They've corrected me a few times 😄😄
I wonder if infinite context windows might be how we end up doing continual learning. There's this idea of a model that learns at inference so it can adapt to new problems dynamically, but if you're able to do something like Textgrad where you can backpropogate through text (essentially it's the same as self reflection, but packaged like pytorch), you could have an LLM dynamically build its own in-context learning notes at inference time.
Thanks Sam!! Good to be reminded that we haven’t arrived at our AI destination… we are just starting the journey. Yes Sam Altman could have phrased the 5% / 95% thing better. I’d hate to have his job…
@@ri3m4nn Not to mention that in my opinion there is ample evidence that Sam is not to be trusted under any condition, and I'm referring to what others that have left the company have said, not to mention some of the key decisions being made such as requiring signatures approved by Open AI for GPU cards, like WTF kind of big brother dystopia mission is this guy on. I think he's the second comming of Bill Gates who was notorious for stealing IP from start-ups, only much, much worse.
you know what would've been great to have from Altman? An example of a startup in this "giant space" that will not get steamrolled by them. He didn't provide one because there isn't one.
Great video! I am amazed by the recent Claude 3.5. I hope Google and OpenAI updates will be as good as Anthropic's. I use LLMs from these three companies for different tasks :) - I only use the best one. The price is not a problem for me, as at the moment I use LLMs just for experimentation, not for scaling.
there is a pattern if you havent notice: claud release killer model, then gemini kills it then openAI kills both, and so on... the span is 1-2 months between each release
Altman is absolutely right on his point. If you plan your product around a flaw or shortcoming, and the shortcoming gets fixed, you're screwed. It's not good long term business planning.
C´mon Sam Altman! A startup (or any startup) needs to use what is available with teh tools that are available. You can´t sit and wait for the next and better version. It would mean waiting forever. What should be said. I suppose, is that one should expect the models becoming better, of course. As Sam Witteveen says.
People seem to be sticking to the terms LLM and VLM. I haven't seen LMM even though it is technically correct. most the VLMs actually bolt on a image encoder and are not Multi modal in the way some other models are that are training just end to end.
Hey Sam, I love the content you produce! I have a question for you. Been learning LLMs and AI for close to 5 years now, and I'm finding it really interesting. Your channels has been really instrumental too. I haven't been this excited about anything in software or computer science in a long time. I have a software engineering background, been at it since 2010. But I'll be honest, I haven't loved any new revolution in the field for a while - until recently. So here's my question: With all this knowledge I'm gaining about LLMs, I don't know what role, position, title or name to give someone with such expertise. I'm finding it hard because I don't want to just call it "software engineering" - that gets lost in the sea of other things. It's not really computer science. Not really a prompt engineer either, because that's not yet an established role. At the same time, it's knowledge that's distinct from the mainstream roles that exist in the software/CS world, like backend dev or full stack or whatever. So what is this field? What are the roles and titles? How does someone with this knowledge present themselves or describe what their role? To put it plainly, What will ones "linkedin" page say?😄 . If anyone in the comments has any ideas, please guide me as well. Thanks!
2 years ago it would have been ML Engineer vs ML Ops etc. Now I am seeing LLM Engineer for people who actually run models and work on models and AI Engineer for people who use APIs.
@@samwitteveenai Oh, so I guess I'm an AI engineer then. Hehe. Thank you. I could be an LLM engineer as well, but I rarely run models. As you know, most models are resource-heavy, and my laptop can't handle some of them, but I have run the lighter models. So, I'm treading between LLM engineer and AI engineer. I also have a feeling "AI engineer" is going to be overused and abused very soon by anyone who even knows how to use ChatGPT, calling themselves an AI engineer. Hehe. But thanks again for this. I love the content you create; keep making it. You are helping a lot.
I love the idea of using ASIC powered MoA frameworks. However, the state of using ASICs is that one still has to stay within the API call rate limits which basically is not usable for group discussion or iterative query applications. Rate limit's aren't just by minute, they are also by day. In many projects that I've used chatGPT to grind with I would say that in a day I've done over 1500 queries. easily. Using an agent library so that it's better than GPT seems to be between 15 and 20 messages per query. So I've just run into rate limit problems enough to find myself just returning to the GPT webui because it just cost less money and don't run out of queries in a day. When 4 first came out it was a really big pain in the neck because I'd run out of queries in like a couple of hours, but that hasn't happened since 4o, for sure.
What???? Modern ASICs, especially those designed for high-performance tasks like running large language models (LLMs) or handling API calls for chatbots, are capable of parallel processing. They can handle multiple operations simultaneously, similar to how GPUs operate.
To paraphrase Sam "There are two types of start ups. Start ups that use with the OpenAI right mindset toward us... and those that use OpenAI with the wrong mindset toward us... No one else huh?
Event and Focus: Coverage of Google I/O held in May in San Francisco. Focus on the evolution of Large Language Models (LLMs) and their impact on startup strategies. Key Topics Discussed: Instruction finetuning. Use of synthetic data. Integration of multimodality to enhance model performance. Significant reduction in token costs. Improvements in LLMs. Expansion of context windows. Evolution of RetrievalAugmented Generation (RAG) systems. Detailed Look at RAG System Design: Embedding models. Context learning. Caching. Dynamic example selection. Designing Applications Utilizing LLMs: Incontext learning. Chunking. Embedding. Promptability. Actionable Strategies: Preparing for smarter AI models. Adapting products to leverage current and future capabilities. Integrating these technologies into products and services.
Well i found it to be a strange claim that models are getting better because they just aren't. I mean benchmarks get +1% and you get crazy stuff like Phi-3 3.8b outperforming bigger models, but let's be honest (GPT-4o is not better than Turbo, which isn't better than the base model). Even Claude 3.5 does not feel better than the previous model. What makes a big difference are the things like artifacts. But let's be honest if we just use the raw model GPT-4 how it came out and Claude 3.5 how it is now. I don't see a noticeable difference. It still has the same issues that are bound to how a LLM works (it's still just text prediction and without proper RAG skills you're not going to get good results). So, who really claims that models are getting better? What can Claude 3.5 do what baser GPT-4 can't? (aside from the 1million token bullshit, let's be honest that's just API provider dreams of ppl doing 1 million token calls xD)
1. Lots of people will happily pay ( I am often surprised when I see some peoples OAI bills, many companies are spending 6 figures a month on GPT models (very annoying when they could achieve the same elsewhere for a fraction of the price)) 2. With things like Context Caching (I made a video on this) you can preload most of the tokens and you just pay a flat fee based on time use for the prefixed tokens.
Q. Why are computer programs becoming smarter than the typical human? A. Computer programs are becoming smarter than the typical human because programmers are focused on the thinking process. If we were at apply this same focus on teaching humans to think, we can greatly improve the average intelligence of all humans.
Not surprisingly, Sam Altman just proved that he's completely detached from how businesses actually function... there's a reason why 95% use the "former" sam.
Well, you seem to be the one bringing in totally unrelated dimensions into a conversation about whether this video has relevant content or not. Like a teenager would do. Besides, you seem to be looking for underage people on the internet, which is concerning.
Cheers to @samw over @sama. Sama is a charlatan grifter. Please can we move on from him. @samw is a much better representative and explainer of what is going on.