Hi Bidur Nepali,
The token_limit_exceeded error in Azure OpenAI Batch API doesn’t mean that an individual .jsonl file is too large, but rather that the total number of tokens across all enqueued and in‑progress batch jobs has exceeded the configured limit, which is 30,000 tokens for the gpt‑4.1‑batch deployment. Azure applies this limit at the queue level, so even if your files are well under 30K tokens, you’ll still get this error if the combined tokens across jobs exceed that threshold. Currently, there is no dedicated endpoint to check the available batch token quota before submitting jobs. The best practice is to track your batch jobs by listing their status, summing their estimated token usage, and submitting new jobs only when enough capacity is available. Alternatively, you can request a quota increase if your workload needs higher throughput.
For more information: Azure OpenAI in Azure AI Foundry Models quotas and limits
Getting started with Azure OpenAI batch deployments
I hope this information helps. Thank you!