Skip to content Skip to footer

How I Won Singapore’s GPT-4 Prompt Engineering Competition | by Sheila Teo | Dec, 2023


A deep dive into the strategies I learned for harnessing the power of Large Language Models

Celebrating a milestone — The real win was the priceless learning experience!

Last month, I had the incredible honor of winning Singapore’s first ever GPT-4 Prompt Engineering competition, which brought together over 400 prompt-ly brilliant participants, organised by the Government Technology Agency of Singapore (GovTech).

Prompt engineering is a discipline that blends both art and science — it is as much technical understanding as it is of creativity and strategic thinking. This article is a compilation of the prompt engineering strategies and insights that I learned along the way, that push any LLM to do exactly what you need and more!

This article covers the following, with 🟢 referring to beginner-friendly prompting techniques while 🟠 refers to advanced strategies:

1. [🟢] Structuring prompts using the CO-STAR framework

2. [🟢] Sectioning prompts using delimiters

3. [🟠] Creating system prompts with LLM guardrails

4. [🟠] Analyzing datasets using only LLMs, without plugins or code
With a hands-on example of analyzing a real-world Kaggle dataset

Effective prompt structuring is crucial for eliciting optimal responses from an LLM. The CO-STAR framework, a brainchild of GovTech Singapore’s Data Science & AI team, is a handy template for structuring prompts. It considers all the key aspects that influence the effectiveness and relevance of an LLM’s response, leading to more optimal responses.

CO-STAR framework — Image by author

Here’s how it works:

(C) Context: Provide background information on the task

This helps the LLM understand the specific scenario being discussed, ensuring its response is relevant.

(O) Objective: Define what the task is that



Source link