The fine art of human prompt engineering: How to talk to a person like ChatGPT
People are more like AI language models than you might think. Here are some prompting tips. In a break from our normal practice, Ars is publishing this helpful guide to knowing how to prompt the human brain, should you encounter one during your daily routine.
While AI assistants like ChatGPT have taken the world by storm, a growing body of research shows that it’s also possible to generate useful outputs from what might be called human language models, or people. Much like large language models (LLMs) in AI, HLMs have the ability to take information you provide and transform it into meaningful responses — if you know how to craft effective instructions, called prompts.
Human prompt engineering is an ancient art form dating at least back to Aristotle’s time, and it also became widely popular through books published in the modern era before the advent of computers.
Understanding human language models
LLMs like those that power ChatGPT, Microsoft Copilot, Google Gemini, and Anthropic Claude all rely on an input called a prompt, which can be a text string or an image encoded into a series of tokens (fragments of data). The goal of each AI model is to take those tokens and predict the next most-likely tokens that follow, based on data trained into their neural networks. That prediction becomes the output of the model.
Similarly, prompts allow human language models to draw upon their training data to recall information in a more contextually accurate way. For example, if you prompt a person with Mary had a, you might expect an HLM to complete the sentence with little lamb based on frequent instances of the famous nursery rhyme encountered in educational or upbringing datasets. But if you add more context to your prompt, such as In the hospital, Mary had a, the person instead might draw on training data related to hospitals and childbirth and complete the sentence with baby.
Despite the black-box nature of their brains, most experts believe that humans build a world model (an internal representation of the exterior world around them) to help complete prompts and that they possess advanced mathematical capabilities, though that varies dramatically by model, and most still need access to external tools to complete accurate calculations. Still, a human’s most useful strength might lie in the verbal-visual user interface, which uses vision and language processing to encode multimodal inputs (speech, text, sound, or images) and then produce coherent outputs based on a prompt.
Humans also showcase impressive few-shot learning capabilities, being able to quickly adapt to new tasks in context (within the prompt) using a few provided examples. Their zero-shot learning abilities are equally remarkable, and many HLMs can tackle novel problems without any prior task-specific training data (or at least attempt to tackle them, to varying degrees of success).
Interestingly, some HLMs (but not all) demonstrate strong performance on common sense reasoning benchmarks, showcasing their ability to draw upon real-world knowledge to answer questions and make inferences. They also tend to excel at open-ended text generation tasks, such as story writing and essay composition, producing coherent and creative outputs.