Welcome to the AI-Infra Benchmarks repository. This repository serves as a centralized location for benchmarks related to Large Language Model (LLM) inference and other performance evaluations. As the repository evolves, we will incorporate additional benchmarks for fine-tuning and inference to gauge performance across different hardware configurations.
/nvidia/
: Contains benchmarks specific to Nvidia hardware./common/
: Includes benchmarks that are not specific to any hardware, such as serving benchmarks./amd/
: Will contain benchmarks for running on AMD hardware.
-
Clone this repository:
git clone https://github.com/runpod/ai-infra-benchmarks.git cd ai-infra-benchmarks
-
Follow instructions in the respective hardware directory to run the benchmarks.
- For Nvidia benchmarks, navigate to the
/nvidia/
directory and follow the provided instructions. - For common benchmarks, check the
/common/
directory. - For AMD benchmarks, follow the instructions in the
/amd/
directory.
- For Nvidia benchmarks, navigate to the
- This repository currently contains inference benchmarks only. As our requirements evolve, we will actively update the repository to include additional benchmarks and tools.
Please follow the standard pull request process and include relevant details about the benchmarks or improvements you are adding.