If I'm using multiple prompts per data point during inference, should I aim for a lower latency per prompt to keep overall processing time in check?