Batch processing sends multiple LLM requests together to be processed asynchronously, typically at reduced cost and without strict latency requirements. Results are retrieved later when processing completes.
This is ideal for large-scale data processing tasks where immediate response isn't needed.