How to ride an exponential wave
It’s been a year since ChatGPT was launched. As with all new, disruptive things, it’s had unexpected impacts as well as disappointments where it wasn’t as magical as we wanted it to be. We are, as always, becoming accustomed to it, so even though a lot is still changing in the world of AI, there is a temptation to settle in to the “new normal”.
But that’s a mistake. We are still very much on an exponential curve in terms of the capabilities of LLMs (and likely other kinds of models). Exponential curves are hard to understand - we tend to think linearly by default, so it’s hard to really visualize something that is moving faster than that. There’s an additional challenge, too - at low values (that is, in the early stages), linear curves can actually be higher than exponential curves. This is the reason for the famous “technology is overestimated in the near term, but underestimated in the long term”. We have a linear model of impact, which is higher than exponential at first, but much lower eventually.
So given that a lot has changed and even more will change, what do we do? How do you ride this wave? The best thing to do, other than just trying your best to stay familiar and educated with what is happening, is to look for invariants. What isn’t likely to change, or will change in degree but not nature, as this space evolves?
I can think of some things. Models will continue to get smarter and fill in gaps they have now. Things like planning, hallucinations, token windows, performance, and cost will all continue to get better. It’s hard to find a new thing to do - the “0 to 1” problem - but much easier to apply a lot of effort in parallel to optimize it once you’ve found it. So anything that can be incrementally improved (even if that increment is very expensive in capital) is likely to be.
The landscape of models will also continue to get more complex. There will be lots of choices for developers - some at the frontier stage, but more and more looking at particular niches, places where some particular mix of data, cost, latency, and quality is better served by a different approach. This will continue to drive the need for tools to manage all of this
We are likely to continue to spend more and more time in front of these models, probably in some kind of “assistant” interface. It’s likely that ChatGPT isn’t the end state of that, and that we will continue to get richer multi-modal interactions
Finally, what do we do, as people, to stay relevant and valuable
This seems to be a constant: that there will always be some advantage or value in being able to work with a model. In that world, the premium is not on learning facts (which can be looked up or explained by the model) but on learning reasoning. How does the world work? What is occam’s razor, or an falsifiable hypothesis, or a fermi question? How do you explore a space you don’t know anything about, without being fooled or lost? What are the basic behaviors of physics, societies, laws, politics?
We are moving from a world where the value isn’t as much in being able to answer the question, but in being able to ask the right one
Give me work to enable you with AI
1yIn some of our non technical roles, we're changing our interview process from "Do this assignment" to "Do this assignment using ChatGPT and send us your prompt history". To your point, that shifts the demonstration of skills in interviews from answering questions to asking the right ones. It is incredibly telling.
Principal SW Engineering Manager at Microsoft
1yThanks for sharing this Sam. This is great. Philosophy i.e. epistemology seems becoming more relevant in leveraging these growing models.
Security Leader
1yThis is great. Thanks Sam!
Founder 🧠 Psychologist 🤖 NLP + GenAI
1y“Be a centaur!” I’m going to use this!
CEO @ Oliver NYC | BA English, Psychology
1yAgreed. It’s always about asking the “right” question!