https://github.com/bytedance/bytemlperf

AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.

https://github.com/bytedance/bytemlperf

Science Score: 26.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
  • Committers with academic emails
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (11.2%) to scientific vocabulary

Keywords

python
Last synced: 6 months ago · JSON representation

Repository

AI Accelerator Benchmark focuses on evaluating AI Accelerators from a practical production perspective, including the ease of use and versatility of software and hardware.

Basic Info
  • Host: GitHub
  • Owner: bytedance
  • License: apache-2.0
  • Language: Python
  • Default Branch: main
  • Homepage: https://bytemlperf.ai/
  • Size: 21.4 MB
Statistics
  • Stars: 256
  • Watchers: 13
  • Forks: 81
  • Open Issues: 22
  • Releases: 1
Topics
python
Created almost 3 years ago · Last pushed 7 months ago
Metadata Files
Readme Contributing License

README.md

ByteMLPerf Benchmark Tool

ByteMLPerf is an AI Accelerator Benchmark that focuses on evaluating AI Accelerators from practical production perspective, including the ease of use and versatility of software and hardware. Byte MLPerf has the following characteristics: - Models and runtime environments are more closely aligned with practical business use cases. - For ASIC hardware evaluation, besides evaluate performance and accuracy, it also measure metrics like compiler usability and coverage. - Performance and accuracy results obtained from testing on the open Model Zoo serve as reference metrics for evaluating ASIC hardware integration.

Category

The ByteMLPerf benchmark is structured into three main categories: Inference, Training, and Micro, each targeting different aspects of AI accelerator performance:

  • Inference: This category is subdivided into two distinct sections to cater to different types of models:

    • General Performance: This section is dedicated to evaluating the inference capabilities of accelerators using common models such as ResNet-50 and BERT. It aims to provide a broad understanding of the accelerator's performance across a range of typical tasks. Vendors can refer to this document for guidance on building general perf backend: ByteMLPerf General Perf Guide [中文版]
    • Large Language Model (LLM) Performance: Specifically designed to assess the capabilities of accelerators in handling large language models, this section addresses the unique challenges posed by the size and complexity of these models. Vendors can refer to this document for guidance on building llm perf backend: ByteMLPerf LLM Perf Guide [中文版]
  • Micro: The Micro category focuses on the performance of specific operations or "ops" that are fundamental to AI computations, such as Gemm, Softmax, and various communication operations. This granular level of testing is crucial for understanding the capabilities and limitations of accelerators at a more detailed operational level. Vendors can refer to this document for guidance on building micro perf backend: ByteMLPerf Micro Perf Guide[中文版]

  • Training: Currently under development, this category aims to evaluate the performance of AI accelerators in training scenarios. It will provide insights into how well accelerators can handle the computationally intensive process of training AI models, which is vital for the development of new and more advanced AI systems.

Vendors looking to evaluate and improve their AI accelerators can utilize the ByteMLPerf benchmark as a comprehensive guide. The benchmark not only offers a detailed framework for performance and accuracy evaluation but also includes considerations for compiler usability and coverage for ASIC hardware, ensuring a holistic assessment approach.

For more details, you can visit our offical website here: bytemlperf.ai

Vendor List

ByteMLPerf Vendor Backend List will be shown below

| Vendor | SKU | Key Parameters | Inference(General Perf) | Inference(LLM Perf) | | :---- | :----| :---- | :---- | :---- | | Intel | Xeon | - | - | - | | Stream Computing | STC P920 |

  • Computation Power:128 TFLOPS@FP16
  • Last Level Buffer: 8MB, 256GB/s
  • Level 1 Buffer: 1.25MB, 512GB/s
  • Memory: 16GB, 119.4GB/S
  • Host Interface:PCIe 4, 16x, 32GB/s
  • TDP: 160W | STC Introduction | - | | Graphcore | Graphcore® C600 |
  • Compute: 280 TFLOPS@FP16, 560 TFLOPS@FP8
  • In Processor Memory: 900 MB, 52 TB/s
  • Host Interface: Dual PCIe Gen4 8-lane interfaces, 32GB/s
  • TDP: 185W | IPU Introduction | - | | Moffett-AI | Moffett-AI S30 |
  • Compute: 1440 (32x-Sparse) TFLOPS@BF16, 2880 (32x-Sparse) TOPS@INT8,
  • Memory: 60 GB,
  • Host Interface: Dual PCIe Gen4 8-lane interfaces, 32GB/s
  • TDP: 250W | SPU Introduction | - | | Habana | Gaudi2 |
  • 24 Tensor Processor Cores, Dual matrix multiplication engines
  • Memory: 96 GB HBM2E, 48MB SRAM | HPU Introduction | - |

    Statement

    ASF Statement on Compliance with US Export Regulations and Entity List

  • Owner

    • Name: Bytedance Inc.
    • Login: bytedance
    • Kind: organization
    • Location: Singapore

    GitHub Events

    Total
    • Create event: 8
    • Issues event: 17
    • Watch event: 57
    • Delete event: 8
    • Member event: 1
    • Issue comment event: 36
    • Push event: 55
    • Pull request review event: 1
    • Pull request review comment event: 1
    • Pull request event: 23
    • Fork event: 24
    Last Year
    • Create event: 8
    • Issues event: 17
    • Watch event: 57
    • Delete event: 8
    • Member event: 1
    • Issue comment event: 36
    • Push event: 55
    • Pull request review event: 1
    • Pull request review comment event: 1
    • Pull request event: 23
    • Fork event: 24

    Committers

    Last synced: 8 months ago

    All Time
    • Total Commits: 189
    • Total Committers: 15
    • Avg Commits per committer: 12.6
    • Development Distribution Score (DDS): 0.693
    Past Year
    • Commits: 48
    • Committers: 4
    • Avg Commits per committer: 12.0
    • Development Distribution Score (DDS): 0.229
    Top Committers
    Name Email Commits
    jianzhe.xiao j****o@b****m 58
    jiangzishan j****n@b****m 56
    Jack Zhang j****z@g****i 21
    HanTengfei99 1****9 10
    Ke Gong k****g@g****i 10
    Minghui-BD 9****D 9
    gaoyujia.01 g****1@b****m 8
    stc-qiupeng p****u@s****m 5
    kevinsouthByteDance x****n@B****M 4
    N m****l@g****m 3
    AngryPowman p****2@g****m 1
    Moffett AI 1****i 1
    huijuanz h****u@i****m 1
    root r****t@i****g 1
    root r****t@n****g 1

    Issues and Pull Requests

    Last synced: 6 months ago

    All Time
    • Total issues: 30
    • Total pull requests: 108
    • Average time to close issues: 29 days
    • Average time to close pull requests: 5 days
    • Total issue authors: 22
    • Total pull request authors: 33
    • Average comments per issue: 1.9
    • Average comments per pull request: 0.54
    • Merged pull requests: 74
    • Bot issues: 0
    • Bot pull requests: 0
    Past Year
    • Issues: 14
    • Pull requests: 25
    • Average time to close issues: 11 days
    • Average time to close pull requests: 1 day
    • Issue authors: 13
    • Pull request authors: 10
    • Average comments per issue: 2.14
    • Average comments per pull request: 0.6
    • Merged pull requests: 13
    • Bot issues: 0
    • Bot pull requests: 0
    Top Authors
    Issue Authors
    • incomingflyingbrick (3)
    • jackzipu (3)
    • DeepTecher (3)
    • Yi-sir (2)
    • TianTengya (1)
    • At1a8 (1)
    • yizhou-xu (1)
    • ZhaiFeiyue (1)
    • wangrt1 (1)
    • abelleeye (1)
    • longerzone (1)
    • danielhua23 (1)
    • zhengy001 (1)
    • yaoyixuejava (1)
    • trinashih (1)
    Pull Request Authors
    • suisiyuan (59)
    • HanTengfei99 (16)
    • YJessicaGao (12)
    • jackzipu (11)
    • MikuGhoul (7)
    • keg0704 (7)
    • JianzheXiao (6)
    • terU3760 (5)
    • valarLip (4)
    • kevinsouthByteDance (3)
    • qinhj (2)
    • gliangMT (2)
    • hliuca (2)
    • caizhi-mt (2)
    • stc-qiupeng (2)
    Top Labels
    Issue Labels
    Pull Request Labels
    add_op (1)