llm_rerank always chunks input tuples, even when the entire set would fit within the model’s context window. Additionally, the implementation uses a hard limit of 2048 tuples per batch, while in practice the LLM can handle more tuples if their combined token length fits in the context window.