Iterate Continuously
Small wording changes can make or break behavior. Constantly iterate and test different phrasings to optimize performance. Example: In one experiment, changing “inaudible” to “unintelligible” in instructions for handling noisy input significantly improved model performance. After your first attempt at a system prompt, have an LLM check it for ambiguity or conflicts.Use Bullet Points Over Paragraphs
Clear, short bullet points are better than long paragraphs. Real-time models follow concise bullet points more effectively. Long paragraphs are harder for models to parse and follow. Breaking them into bullet points makes instructions clearer:- Paragraph (Harder to Follow)
- Bullet Points (Easier to Follow)
Provide Example Phrases
The model strongly follows example phrases. Provide short, varied examples for common conversation moments. Include sample phrases with variations so the model doesn’t repeat the same responses. For example, if providing greeting examples:Be Precise
Ambiguity and conflicting instructions degrade performance. Similar to GPT models, precision matters.- Check for vague terms and replace with specific language
- Eliminate conflicting instructions
- Test different wordings to find what works best
- Use LLMs to review prompts for ambiguity and conflicts
Control Language
If you see language drift, pin the output to the target language. Add a dedicated “Language” section to your prompt. Simple language matching:Reduce Repetition
Add variety rules to reduce robotic phrasing. If responses sound repetitive or mechanical, include explicit variety instructions.Use Capitalization for Emphasis
Use CAPITALIZED text for important rules to make them stand out to the model. It’s also helpful to convert non-text rules (like numeric conditions) to text before capitalizing. Convert numeric conditions to text before capitalizing. For example, instead of writing “if x > 3 escalate”, write “IF MORE THAN THREE FAILURES OCCUR, ESCALATE.”- Numeric Condition
- Text with Capitalization