VLLM runtime has a wealth of token metrics, example prompt_tokens_total and generation_tokens_total. Why does mlserver have none?