AGI's already here, it's just not evenly organized
Published on July 30, 2023
Knowledge work can be explained – fairly reductively – as labour that involves the usage of structured (trends, demographics, financials etc) and unstructured data (expertise, corroborated evidence, and new knowledge, all of which may or may not have been stored in plain text) to make decisions – or to help some other people make decisions. Virtually every job depicted as ‘knowledge work’ or ‘intellectual labour’ can be described this way, from the heads of enterprises to their lowliest analysts. Wikipedia puts it succinctly: “Knowledge workers are workers whose main capital is knowledge.“
Language plays a crucial role in our species, not least because it is one of the primary modalities – perhaps the dominant modality – through which unstructured data is transmitted. The reasons 2022-and-beyond-class LLMs have garnered a lot of attention (hah!) boil down to the fact that they seem to, at the very least, be perfectly capable of acting like they perceive language in the same way as us. Are we next-token predictors too? I tend to err on the side of the negative but nonetheless, machines seem to 'understand’ language (well)! That’s huge.
So, now what? A tempting observation to make – and it is one that a lot of people are making – would be to claim that we’re all going to die knowledge work will become increasingly automated. Perhaps the primary reason it wasn’t automated already is the machines’ inability to wrangle with the tools of language and, as a consequence, unstructured data. Those seem like valid points – and AI companies and laboratories are working towards making the automation of knowledge work, and more, a reality. The stated goal of OpenAI is to build Artificial General Intelligence (AGI), defined in the OpenAI charter as “highly autonomous systems that outperform humans at most economically valuable work.” Ilya Sutskever, Chief Scientist at OpenAI, recently offered a somewhat less circuitous description, “AGI is a computer system which can automate the great majority of intellectual labour.”
Casting my aspirations on the moral nature of these efforts aside (broadly speaking, I believe that pursuing AI progress is very worth it), my wish for this essay is to expand upon how I think large-scale automation of intellectual labour will play out and why we might be closer than we think.
In 2018, Dan Wang, Technology/China-US Industrial Policy Analyst at Gavekal Dragonomics, wrote an oft-quoted essay on the importance of process knowledge, which he describes as “what we can also refer to as tacit knowledge, know-how, technical experience. Process knowledge is the kind of knowledge that’s hard to write down as an instruction.” He then brings the point home with a clear example:
"You can give someone a well-equipped kitchen and an extraordinarily detailed recipe, but unless he already has some cooking experience, we shouldn’t expect him to prepare a great dish."
Though written with industrial policy in mind, I find the construct “process knowledge” useful when visualising and extrapolating current and future states of AGI[1] research. Consider the following two premises:
-
The lack of some form of weak or strong process knowledge (high-quality data that renders its owner expertise) will hinder the most common way the big AI Labs are approaching building AGI systems: by brute-forcing ever larger foundational language models with low- (internet corpora) to somewhat high- (SFT and RLHF) signal generalised data. Basically, there are some intuitive processes that humans currently perform that you can’t expect an AI to ‘know’ and, therefore, replicate/outperform without better information, especially that which is not precisely available in neat, well-defined packages. This could be anything from how the best traders set stop-losses and manage downside risk and the subtleties in balance sheets and annual reports that the best equity investors and stock pickers look for to how the best drivers drive and a physician’s intuitive understanding of therapeutic best practices for his patients.
Depending on the validity of this premise and the observed strength of the notion of ‘process knowledge’, a centralised model achieving ultra-human general performance using current methods would be, at the limit, perhaps impossible and, at the very least, operationally inefficient. GPT will achieve driving capabilities (if it does at all) long after we have self-driving cars.
- Current foundational LLMs and/or LLMs of the very near future, incentivised and fine-tuned to purpose, already possess – or can be fine-tuned and gated to possess – the requisite linguistic, syllogistic, and, in general, reasoning capabilities for human-level performance (and, in some circumstances, beyond) across a decent variety of intellectual tasks. The most promising early use cases of this class of LLMs – across Customer Service and client-facing tools like Shopify’s Sidekick – more or less offer satisfactory evidence for the validity of this premise.
With the premises in mind, I think the better way to approach the AGI problem – and deliver on its immense promise – is to look at it as a set of discrete tractable tasks, each with its process knowledge (high-signal data and fine-tuning) requirements. I also think developing superhuman agentic systems in narrow domains – eventually all working together to automate the majority of intellectual labour – is possible without any breakthrough improvements to today’s (July 2023) technology. As described above, this has some anthropomorphic hints: humans, after all, only command expertise over very narrow domains of complex systems.
Viewing our approach in the context of Dan’s example (mentioned in the first paragraph above): sufficient data quality, granularity and specificity should yield performance indistinguishable from that of the control group. In this case, this would mean giving the amateur (an AI) a recipe that contained everything he could need (high-signal information/methodology) – it would tell him exactly how to cut the vegetables, visual cues for spotting doneness, and every other infinitesimally small written and unwritten detail that separates the best of chefs from the dilettantes. This should, in theory, allow him to come up with a dish that rivals that of an expert. Maybe training general foundational models gets to this point too but one would assume – if they do at all – that they will arrive much later than the narrow and focused model.
Note that I don’t mean we should attempt to hard-code human intuitions and processes into these machines. Focus allows for unique processes, better data, and devoted computing specific to solving domain-specific problems. AlphaFold, GPT-4, and AlphaZero all have unique architectural characteristics that enable superhuman performance in verticals.
There’s an added benefit to approaching individual intellectual domains in isolation: solving related subsequent domains are easier. An AI that’s an expert equities investor will also possess characteristics that can help it be a stellar bond analyst or a management consultant. An AI that can write beautiful software will also have abilities that can help in chip design.
This, in my opinion, is the path to delivering AGI’s promise.
Surprisingly, there are few efforts to build practical systems that can automate the most highly valued intellectual labour. The major AI labs are focused on developing some form of generalized super-intelligence, while startups focus on highly linguistic but far less socially valuable tasks like copywriting and summarization. Large incumbents who could benefit from increased automation are reluctant to commit resources to the task. They are hindered by, amongst other things, a lack of research and technical talent, innovator’s dilemma and institutional drudgery, and a general apathy towards automation which, admittedly, has its risks but is worth pursuing nonetheless. They are not early adopters anyway.
I fear that (admitting that the possibility is minuscule) AI’s promise may not be realized, simply because we’ve overlooked practical solutions in a search for some mythical super-intelligent being – and worrying about the safety concerns that finding that mythical super-intelligent mechanical being will entail. This class of LLMs are tools capable of ‘understanding’ and, to some extent, reasoning in natural language fairly accurately. That’s their essence and that’s what makes them great.
If we start working on these practical narrow systems, I suspect we’ll quickly discover that 'AGI' has been with us all this time.
Footnotes:
- I only realised as I was writing this that the term “AGI” has some weird connotations signalling some god-being for what is essentially a set of tools that can augment human productivity. (On this, I’m very much in alignment with Marc Andreessen.). If you notice any reticence on my part to state the term AGI explicitly, this is probably the reason. ↩