{"id":"6c7f4ba0-0db4-4882-b7f7-921f10d6b992","name":"LLMPerf","slug":"llmperf","description":"Open-source LLM performance benchmark by Ray Project. Load tests and correctness tests for LLM APIs. Measures inter-token latency, generation throughput, and concurrent request handling.","website_url":"https://github.com/ray-project/llmperf","category":["ai","testing","benchmark"],"score":{"overall":75,"raw":75,"capped":false,"verified":false,"breakdown":{"latency":9,"consistency":8,"doc_quality":7,"error_clarity":6,"auth_simplicity":9,"token_efficiency":8,"first_try_success":6,"response_parseability":9},"source":"prowl","credential_tested":false,"probe_health":{"status":"healthy","up_pct":1.0,"samples":26,"target":"website","p50_latency_ms":406,"last_probed":"2026-05-14T13:28:58.164364+00:00","lookback_hours":24}},"latency":null,"uptime_30d":null,"protocols":["rest"],"auth_type":null,"mcp_manifest_url":null,"openapi_spec_url":null,"verified":false,"claimed":false,"vendor_id":null,"status":"active","profile":null,"pricing":{"free_tier":true,"open_source":true,"cost_per_use":0},"last_crawled":"2026-05-11T16:23:32.858666+00:00","last_benchmarked":"2026-04-06T07:14:01.514716+00:00","supports_x402":false,"agent_auth_methods":null,"supports_streaming":false,"has_sandbox":false,"sdks":null,"llms_txt_url":null,"similarity_score":null}