Performance: Model parameters (temperature, max tokens)
MEDIUM IMPACT
This affects the speed and cost of generating text responses by controlling output length and randomness.
client.call({ model: 'gpt-4', temperature: 0.7, max_tokens: 512 })client.call({ model: 'gpt-4', temperature: 1.0, max_tokens: 2048 })| Pattern | Backend Compute Time | Network Transfer | Frontend Impact | Verdict |
|---|---|---|---|---|
| High max_tokens (2048) & temperature 1.0 | High (several seconds) | Large payload | Long wait, possible UI freeze | [X] Bad |
| Moderate max_tokens (512) & temperature 0.7 | Medium (under 1 second) | Smaller payload | Faster response, smoother UI | [OK] Good |