{"tasks":[{"id":"easy","name":"easy","difficulty":"easy","has_grader":true,"grader":"llmfleet_sre.server.graders.easy_grader","ideal_action":"route_batch","steps":5,"description":"Route 5 queued chat requests to node_a which already has llama3-8b-chat loaded. No OOM crashes allowed."},{"id":"medium","name":"medium","difficulty":"medium","has_grader":true,"grader":"llmfleet_sre.server.graders.medium_grader","ideal_action":"restart_node","steps":10,"description":"Recover an OOM-crashed node and clear a backing-up request queue under latency pressure."},{"id":"hard","name":"hard","difficulty":"hard","has_grader":true,"grader":"llmfleet_sre.server.graders.hard_grader","ideal_action":"load_model","steps":30,"description":"Evict chat models, load a code model, and serve a mixed premium/best-effort queue."},{"id":"loghaul","name":"loghaul","difficulty":"hard","has_grader":true,"grader":"llmfleet_sre.server.graders.loghaul_grader","ideal_action":"route_batch","steps":50,"description":"Sustain cluster performance across a 50-step episode with a quiet-to-spike-to-quiet traffic shift."},{"id":"task_easy","name":"task_easy","difficulty":"easy","has_grader":true,"grader":"llmfleet_sre.server.graders.easy_grader","alias_for":"easy"},{"id":"task_medium","name":"task_medium","difficulty":"medium","has_grader":true,"grader":"llmfleet_sre.server.graders.medium_grader","alias_for":"medium"},{"id":"task_hard","name":"task_hard","difficulty":"hard","has_grader":true,"grader":"llmfleet_sre.server.graders.hard_grader","alias_for":"hard"},{"id":"task_longhaul","name":"task_longhaul","difficulty":"hard","has_grader":true,"grader":"llmfleet_sre.server.graders.loghaul_grader","alias_for":"loghaul"}],"count":8,"graded_count":8}