https://github.com/bytedance/bytemlperf
AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.
Science Score: 26.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Committers with academic emails
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (11.2%) to scientific vocabulary
Keywords
Repository
AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.
Basic Info
- Host: GitHub
- Owner: bytedance
- License: apache-2.0
- Language: Python
- Default Branch: main
- Homepage: https://bytemlperf.ai/
- Size: 21.4 MB
Statistics
- Stars: 256
- Watchers: 13
- Forks: 81
- Open Issues: 22
- Releases: 1
Topics
Metadata Files
README.md
ByteMLPerf Benchmark Tool
ByteMLPerf is an AI Accelerator Benchmark that focuses on evaluating AI Accelerators from practical production perspective, including the ease of use and versatility of software and hardware. Byte MLPerf has the following characteristics: - Models and runtime environments are more closely aligned with practical business use cases. - For ASIC hardware evaluation, besides evaluate performance and accuracy, it also measure metrics like compiler usability and coverage. - Performance and accuracy results obtained from testing on the open Model Zoo serve as reference metrics for evaluating ASIC hardware integration.
Category
The ByteMLPerf benchmark is structured into three main categories: Inference, Training, and Micro, each targeting different aspects of AI accelerator performance:
Inference: This category is subdivided into two distinct sections to cater to different types of models:
- General Performance: This section is dedicated to evaluating the inference capabilities of accelerators using common models such as ResNet-50 and BERT. It aims to provide a broad understanding of the accelerator's performance across a range of typical tasks. Vendors can refer to this document for guidance on building general perf backend: ByteMLPerf General Perf Guide [中文版]
- Large Language Model (LLM) Performance: Specifically designed to assess the capabilities of accelerators in handling large language models, this section addresses the unique challenges posed by the size and complexity of these models. Vendors can refer to this document for guidance on building llm perf backend: ByteMLPerf LLM Perf Guide [中文版]
Micro: The Micro category focuses on the performance of specific operations or "ops" that are fundamental to AI computations, such as Gemm, Softmax, and various communication operations. This granular level of testing is crucial for understanding the capabilities and limitations of accelerators at a more detailed operational level. Vendors can refer to this document for guidance on building micro perf backend: ByteMLPerf Micro Perf Guide[中文版]
Training: Currently under development, this category aims to evaluate the performance of AI accelerators in training scenarios. It will provide insights into how well accelerators can handle the computationally intensive process of training AI models, which is vital for the development of new and more advanced AI systems.
Vendors looking to evaluate and improve their AI accelerators can utilize the ByteMLPerf benchmark as a comprehensive guide. The benchmark not only offers a detailed framework for performance and accuracy evaluation but also includes considerations for compiler usability and coverage for ASIC hardware, ensuring a holistic assessment approach.
For more details, you can visit our offical website here: bytemlperf.ai
Vendor List
ByteMLPerf Vendor Backend List will be shown below
| Vendor | SKU | Key Parameters | Inference(General Perf) | Inference(LLM Perf) | | :---- | :----| :---- | :---- | :---- | | Intel | Xeon | - | - | - | | Stream Computing | STC P920 |
Statement
ASF Statement on Compliance with US Export Regulations and Entity List
Owner
- Name: Bytedance Inc.
- Login: bytedance
- Kind: organization
- Location: Singapore
- Website: https://opensource.bytedance.com
- Twitter: ByteDanceOSS
- Repositories: 255
- Profile: https://github.com/bytedance
GitHub Events
Total
- Create event: 8
- Issues event: 17
- Watch event: 57
- Delete event: 8
- Member event: 1
- Issue comment event: 36
- Push event: 55
- Pull request review event: 1
- Pull request review comment event: 1
- Pull request event: 23
- Fork event: 24
Last Year
- Create event: 8
- Issues event: 17
- Watch event: 57
- Delete event: 8
- Member event: 1
- Issue comment event: 36
- Push event: 55
- Pull request review event: 1
- Pull request review comment event: 1
- Pull request event: 23
- Fork event: 24
Committers
Last synced: 8 months ago
Top Committers
| Name | Commits | |
|---|---|---|
| jianzhe.xiao | j****o@b****m | 58 |
| jiangzishan | j****n@b****m | 56 |
| Jack Zhang | j****z@g****i | 21 |
| HanTengfei99 | 1****9 | 10 |
| Ke Gong | k****g@g****i | 10 |
| Minghui-BD | 9****D | 9 |
| gaoyujia.01 | g****1@b****m | 8 |
| stc-qiupeng | p****u@s****m | 5 |
| kevinsouthByteDance | x****n@B****M | 4 |
| N | m****l@g****m | 3 |
| AngryPowman | p****2@g****m | 1 |
| Moffett AI | 1****i | 1 |
| huijuanz | h****u@i****m | 1 |
| root | r****t@i****g | 1 |
| root | r****t@n****g | 1 |
Committer Domains (Top 20 + Academic)
Issues and Pull Requests
Last synced: 6 months ago
All Time
- Total issues: 30
- Total pull requests: 108
- Average time to close issues: 29 days
- Average time to close pull requests: 5 days
- Total issue authors: 22
- Total pull request authors: 33
- Average comments per issue: 1.9
- Average comments per pull request: 0.54
- Merged pull requests: 74
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 14
- Pull requests: 25
- Average time to close issues: 11 days
- Average time to close pull requests: 1 day
- Issue authors: 13
- Pull request authors: 10
- Average comments per issue: 2.14
- Average comments per pull request: 0.6
- Merged pull requests: 13
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
- incomingflyingbrick (3)
- jackzipu (3)
- DeepTecher (3)
- Yi-sir (2)
- TianTengya (1)
- At1a8 (1)
- yizhou-xu (1)
- ZhaiFeiyue (1)
- wangrt1 (1)
- abelleeye (1)
- longerzone (1)
- danielhua23 (1)
- zhengy001 (1)
- yaoyixuejava (1)
- trinashih (1)
Pull Request Authors
- suisiyuan (59)
- HanTengfei99 (16)
- YJessicaGao (12)
- jackzipu (11)
- MikuGhoul (7)
- keg0704 (7)
- JianzheXiao (6)
- terU3760 (5)
- valarLip (4)
- kevinsouthByteDance (3)
- qinhj (2)
- gliangMT (2)
- hliuca (2)
- caizhi-mt (2)
- stc-qiupeng (2)