Llm Bench
D 56 completedPipeline State
completedPipeline Metadata
🧪 Code Distillation
Browse all specs →AI Prompt
Catalog Information
A high-performance benchmarking tool for OpenAI-compatible LLM inference servers, designed to measure detailed performance characteristics of local LLM servers.
Description
llm-bench is a high-performance benchmarking tool for OpenAI-compatible LLM inference servers. It measures detailed performance characteristics of local LLM servers like llama-server, vLLM, TGI, and other OpenAI API-compatible endpoints. The tool supports various load patterns, including concurrent mode, fixed QPS mode, arrival distributions, duration-based testing, request count mode, and warmup period.
الوصف
هو أداة تقييم الأداء عالية السرعة لخادمات الاستعلام المتوافقة مع LLM من OpenAI. يقيّم أداء خادمات LLM المحلية مثل llama-server، vLLM، TGI، وغيرها من نقاط النهاية المتناظرة مع API OpenAI. تدعم الأداة أنماط تحميل متعددة، بما في ذلك وضع التزامن، ووضع قيم الاستعلام الثابتة، وتوزيعات الوصول، والاختبار على أساس المدة، ووضع عدد الطلبات، وفترة التأهب.
Novelty
7/10Tags
Technologies
Claude Models
Quality Score
Strengths
- CI/CD pipeline configured (github_actions)
- Consistent naming conventions (snake_case)
- Good security practices \u2014 no major issues detected
Weaknesses
- No LICENSE file \u2014 legal ambiguity for contributors
- No tests found \u2014 high risk of regressions
- 765 duplicate lines detected \u2014 consider DRY refactoring
- 2 'god files' with >500 LOC need decomposition
Recommendations
- Add a test suite \u2014 start with critical path integration tests
- Add a linter configuration to enforce code style consistency
- Add a LICENSE file (MIT recommended for open source)
Security & Health
Languages
Frameworks
Symbols
Concepts (6)
| Category | Name | Description | Confidence | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Source: Repobility analyzer · https://repobility.com | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| design_pattern | Repository | Found repository-named files | 80% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| auto_description | Project Description | A high-performance benchmarking tool for OpenAI-compatible LLM inference servers. Designed to measure detailed performance characteristics of local LLM servers like llama-server, vLLM, TGI, and other OpenAI API-compatible endpoints. | 80% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| auto_category | CLI Tool | cli | 70% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| design_pattern | Factory | Found factory/create_ naming patterns | 60% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| business_logic | Analytics | Detected from 3 related files | 50% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| business_logic | Configuration | Detected from 2 related files | 50% | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Embed Badge
Add to your README:
