{"name":"Eco-LLM Inference Routing Environment","description":"Hierarchical RL environment for carbon-aware LLM query routing across accuracy, latency, energy, and caching objectives.","readme_content":null,"version":"1.0.0","author":null,"documentation_url":null}