7 Comments
Mar 5Liked by Jurgen Gravestein

To me, prompt engineering has a lot of classical control and regulation theory, like you find it in all engineering disciplines when dealing with more complex and non-deterministic systems and processes.

You know basically what you want, but cannot set your parameters exactly to get the desired result. Feedback is fuzzy. So instead of computing everything fully deterministically, you'll have to tinker around a little in the hopes of coming closer to what you originally intended.

It's also no wonder AI behaves so similar to that. Neural networks are just statistical estimators, which are also used in classical engineering to model complex systems.

And if you look at software engineering, building a piece of software with agile methodologies quite closely resembles the way you interact with an LLM to get the answer you want.

Expand full comment

I’ve found the greatest amount of creativity shall we say occurs in prompts with even a scintillaty of a wisp of moral or ethical contest. Prompt isn’t the best word. It resonates with motivated, evocative, invitation, directions, assignments, etc. squishy human attitudes toward speculation, uncertainty, weakly situated in epistemic clarity. I prefer command mode. Nothing human about it. Essentially, the bot has learned to read our mind. If we hold a clouded or surface sense of our project and treat the bot like a teenager we get teenage output. Might be an interesting experiment for a discourse analysis—an adult talking like a teenager vs a teen talking like an adult. Epistemic experts get better output from general LMs but even better from niche bots. How? Not by taking to them like adolescents. Command posture.

Expand full comment
Apr 11Liked by Jurgen Gravestein

I'd say it's engineering in about the same sense as social engineering - except maybe 10X harder.

Expand full comment
Mar 5Liked by Jurgen Gravestein

I agree that, in the words of Jeff Bezos, LLMs have been discovered rather than invented, and that empirical investigation is necessary in order to fully understand their capabilities. However, I disagree that *behavioural* psychology is the appropriate methodology; we can understand many aspects of their capabilities, or lack of, by taking into account not just their outward behaviour but also their internal architecture and the way they have been trained. Humans have many biases that result from the fact that we were "trained" to maximise inclusive fitness in a highly-changeable social environment. Similarly, LLMs have many biases that result from the fact that they were trained for next-token prediction. Analogous to evolutionary psychology, there will be branches of the new field of machine psychology that take into account the underlying objective-function, e.g. "auto-regressive machine psychology". This methodological approach is advocated in the seminal paper "Embers of autoregression: Understanding large language models through the problem they are trained to solve" (McCoy et al. 2023). I have written a short essay on this paper here: https://sphelps.substack.com/p/a-teleological-approach-to-understanding.

McCoy, R. Thomas, et al. "Embers of autoregression: Understanding large language models through the problem they are trained to solve." arXiv preprint arXiv:2309.13638 (2023).

Expand full comment