Unverified Commit dfbd70b1 authored by Yifan Xiong's avatar Yifan Xiong Committed by GitHub
Browse files

Release - SuperBench v0.3.0 (#212)



**Description**

Cherry-pick  bug fixes from v0.3.0 to main.

**Major Revisions**
* Docs - Upgrade version and release note (#209)
* Benchmarks: Build Pipeline - Update rccl-test git submodule to dc1ad48 (#210)
* Benchmarks: Update - Update benchmarks in configuration file (#208)
* CI/CD - Update GitHub Action VM (#211)
* Benchmarks: Fix Bug - Fix wrong parameters for gpu-sm-copy-bw in configuration examples (#203)
* CI/CD - Fix bug in build image for push event (#205)
* Benchmark: Fix Bug - fix error message of communication-computation-overlap (#204)
* Tool: Fix bug - Fix function naming issue in system info  (#200)
* CI/CD - Push images in GitHub Action (#202)
* Bug - Fix torch.distributed command for single node (#201)
* CLI - Integrate system info for node (#199)
* Benchmarks: Code Revision - Revise CMake files for microbenchmarks. (#196)
* CI/CD - Add ROCm image build in GitHub Actions (#194)
* Bug: Fix bug - fix bug of hipBusBandwidth build (#193)
* Benchmarks: Build Pipeline - Restore rocblas build logic (#197)
* Bug: Fix Bug - Add barrier before 'destroy_process_group' in model benchmarks (#198)
* Bug - Revise 'docker run' in sb deploy (#195)
* Bug - Fix Bug : fix bug of error param operations to operation in rccl-bw of hpe config (#190)
Co-authored-by: default avatarYuting Jiang <v-yujiang@microsoft.com>
Co-authored-by: default avatarGuoshuai Zhao <guzhao@microsoft.com>
Co-authored-by: default avatarZiyue Yang <ziyyang@microsoft.com>
parent 37b15db9
......@@ -3,7 +3,7 @@
# Server:
# - Product: HPE Apollo 6500
version: v0.2
version: v0.3
superbench:
enable: null
var:
......@@ -40,17 +40,13 @@ superbench:
rccl-bw:
enable: true
modes:
- name: mpi
proc_num: 8
env:
NCCL_SOCKET_IFNAME: ens17f0
NCCL_IB_GDR_LEVEL: 1
- name: local
proc_num: 1
parallel: no
parameters:
maxbytes: 128M
minbytes: 32M
iters: 50
ngpus: 1
operations: allreduce
maxbytes: 8G
ngpus: 8
operation: allreduce
mem-bw:
<<: *default_local_mode
gemm-flops:
......@@ -75,15 +71,16 @@ superbench:
parameters:
block_devices: []
gpu-sm-copy-bw:
enable: false
enable: true
modes:
- name: local
proc_num: 32
prefix: CUDA_VISIBLE_DEVICES=$(({proc_rank}%8)) numactl -N $(({proc_rank}%4)) -m $(({proc_rank}%4))
prefix: HIP_VISIBLE_DEVICES=$(({proc_rank}%8)) numactl -N $(({proc_rank}%4)) -m $(({proc_rank}%4))
parallel: no
parameters:
dtoh: true
htod: true
mem_type:
- dtoh
- htod
gpt_models:
<<: *default_pytorch_mode
models:
......
......@@ -4,7 +4,7 @@
# - Product: G482-Z53
# - Link: https://www.gigabyte.cn/FileUpload/Global/MicroSite/553/G482-Z53.html
version: v0.2
version: v0.3
superbench:
enable: null
var:
......@@ -13,7 +13,7 @@ superbench:
modes:
- name: local
proc_num: 8
prefix: CUDA_VISIBLE_DEVICES={proc_rank}
prefix: HIP_VISIBLE_DEVICES={proc_rank}
parallel: yes
default_pytorch_mode: &default_pytorch_mode
enable: true
......@@ -36,6 +36,52 @@ superbench:
- train
pin_memory: yes
benchmarks:
kernel-launch:
<<: *default_local_mode
rccl-bw:
enable: true
modes:
- name: local
proc_num: 1
parallel: no
parameters:
maxbytes: 8G
ngpus: 8
operation: allreduce
mem-bw:
<<: *default_local_mode
gemm-flops:
<<: *default_local_mode
parameters:
m: 7680
n: 8192
k: 8192
ib-loopback:
enable: true
modes:
- name: local
proc_num: 2
prefix: PROC_RANK={proc_rank} IB_DEVICES=0,1
parallel: no
disk-benchmark:
enable: false
modes:
- name: local
proc_num: 1
parallel: no
parameters:
block_devices: []
gpu-sm-copy-bw:
enable: true
modes:
- name: local
proc_num: 32
prefix: HIP_VISIBLE_DEVICES=$(({proc_rank}%8)) numactl -N $(({proc_rank}%4)) -m $(({proc_rank}%4))
parallel: no
parameters:
mem_type:
- dtoh
- htod
gpt_models:
<<: *default_pytorch_mode
models:
......
# SuperBench Config
version: v0.2
version: v0.3
superbench:
enable: null
var:
......@@ -35,6 +35,51 @@ superbench:
<<: *default_local_mode
gemm-flops:
<<: *default_local_mode
nccl-bw:
enable: true
modes:
- name: local
proc_num: 1
parallel: no
parameters:
ngpus: 8
ib-loopback:
enable: true
modes:
- name: local
proc_num: 4
prefix: PROC_RANK={proc_rank} IB_DEVICES=0,2,4,6 NUMA_NODES=1,0,3,2
parallel: yes
- name: local
proc_num: 4
prefix: PROC_RANK={proc_rank} IB_DEVICES=1,3,5,7 NUMA_NODES=1,0,3,2
parallel: yes
mem-bw:
enable: true
modes:
- name: local
proc_num: 8
prefix: CUDA_VISIBLE_DEVICES={proc_rank} numactl -c $(({proc_rank}/2))
parallel: yes
disk-benchmark:
enable: false
modes:
- name: local
proc_num: 1
parallel: no
parameters:
block_devices: []
gpu-sm-copy-bw:
enable: true
modes:
- name: local
proc_num: 32
prefix: CUDA_VISIBLE_DEVICES=$(({proc_rank}%8)) numactl -N $(({proc_rank}%4)) -m $(({proc_rank}%4))
parallel: no
parameters:
mem_type:
- dtoh
- htod
cudnn-function:
<<: *default_local_mode
cublas-function:
......
# SuperBench Config
version: v0.2
version: v0.3
superbench:
enable: null
var:
......@@ -32,7 +32,10 @@ superbench:
enable: true
modes:
- name: local
prefix: NCCL_DEBUG=INFO NCCL_IB_DISABLE=1
proc_num: 1
parallel: no
parameters:
ngpus: 8
ib-loopback:
enable: true
modes:
......@@ -61,15 +64,16 @@ superbench:
prefix: CUDA_VISIBLE_DEVICES={proc_rank} numactl -c $(({proc_rank}/2))
parallel: yes
gpu-sm-copy-bw:
enable: false
enable: true
modes:
- name: local
proc_num: 32
prefix: CUDA_VISIBLE_DEVICES=$(({proc_rank}%8)) numactl -N $(({proc_rank}%4)) -m $(({proc_rank}%4))
parallel: no
parameters:
dtoh: true
htod: true
mem_type:
- dtoh
- htod
kernel-launch:
<<: *default_local_mode
gemm-flops:
......
......@@ -101,7 +101,7 @@
{{ '--security-opt seccomp=unconfined --group-add video' if amd_gpu_exist else '' }} \
-w /root -v {{ workspace }}:/root -v /mnt:/mnt \
-v /var/run/docker.sock:/var/run/docker.sock \
{{ docker_image }} bash && \
--entrypoint /bin/bash {{ docker_image }} && \
docker exec {{ container }} bash -c \
"chown -R root:root ~ && \
sed -i 's/[# ]*Port.*/Port {{ ssh_port }}/g' /etc/ssh/sshd_config && \
......
......@@ -123,20 +123,13 @@ def __get_mode_command(self, benchmark_name, mode):
elif mode.name == 'torch.distributed':
# TODO: replace with torch.distributed.run in v1.9
# TODO: only supports node_num=1 and node_num=all currently
torch_dist_params = '' if mode.node_num == 1 else \
'--nnodes=$NNODES --node_rank=$NODE_RANK --master_addr=$MASTER_ADDR --master_port=$MASTER_PORT '
mode_command = (
'python3 -m torch.distributed.launch '
'--use_env --no_python --nproc_per_node={proc_num} '
'--nnodes={node_num} --node_rank=$NODE_RANK '
'--master_addr=$MASTER_ADDR --master_port=$MASTER_PORT '
'{command} {torch_distributed_suffix}'
).format(
proc_num=mode.proc_num,
node_num=1 if mode.node_num == 1 else '$NNODES',
command=exec_command,
torch_distributed_suffix=(
'superbench.benchmarks.{name}.parameters.distributed_impl=ddp '
'superbench.benchmarks.{name}.parameters.distributed_backend=nccl'
).format(name=benchmark_name),
f'python3 -m torch.distributed.launch'
f' --use_env --no_python --nproc_per_node={mode.proc_num} {torch_dist_params}{exec_command}'
f' superbench.benchmarks.{benchmark_name}.parameters.distributed_impl=ddp'
f' superbench.benchmarks.{benchmark_name}.parameters.distributed_backend=nccl'
)
elif mode.name == 'mpi':
mode_command = (
......
......@@ -4,15 +4,19 @@
"""Generate system config."""
import json
import os
import subprocess
import xmltodict
from pathlib import Path
import xmltodict
from superbench.common.utils import logger
class SystemInfo(): # pragma: no cover
"""Systsem info class."""
def run_cmd(self, command):
"""Run the command as root or non-root user and return the stdout string..
def _run_cmd(self, command):
"""Run the command and return the stdout string.
Args:
command (string): the command to run in terminal.
......@@ -21,11 +25,17 @@ def run_cmd(self, command):
string: the stdout string of the command.
"""
output = subprocess.run(
command, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, shell=True, check=False, universal_newlines=True
command,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
shell=True,
check=False,
universal_newlines=True,
timeout=300
)
return output.stdout
def count_prefix_indent(self, content, symbol='\t'):
def __count_prefix_indent(self, content, symbol='\t'):
r"""Count the number of a specific symbol in the content.
Args:
......@@ -43,7 +53,7 @@ def count_prefix_indent(self, content, symbol='\t'):
break
return count
def parse_key_value_lines(self, lines, required_keywords=None, omitted_values=None, symbol=':'): # noqa: C901
def _parse_key_value_lines(self, lines, required_keywords=None, omitted_values=None, symbol=':'): # noqa: C901
"""Parse the lines like "key:value" and convert them to dict.
if required_keywords is None, include all line. Otherwise,
......@@ -83,7 +93,7 @@ def parse_key_value_lines(self, lines, required_keywords=None, omitted_values=No
while next_indent_index < length and self.__count_prefix_indent(lines[next_indent_index]) > indent:
next_indent_index += 1
value = self.__parse_key_value_lines(lines[i + 1:next_indent_index])
value = self._parse_key_value_lines(lines[i + 1:next_indent_index])
i = next_indent_index - 1
# split line by symbol
elif symbol in line:
......@@ -112,7 +122,7 @@ def parse_key_value_lines(self, lines, required_keywords=None, omitted_values=No
i += 1
return dict
def parse_table_lines(self, lines, key):
def _parse_table_lines(self, lines, key):
"""Parse lines like a table and extract the colomns whose table index are the same as key to list of dict.
Args:
......@@ -125,7 +135,6 @@ def parse_table_lines(self, lines, key):
index = []
list = []
valid = False
try:
for line in lines:
line = line.split()
if key[0] in line:
......@@ -139,8 +148,6 @@ def parse_table_lines(self, lines, key):
if index[i] < len(line):
dict[key[i]] = line[index[i]]
list.append(dict)
except Exception:
print('Error: key error in __parse_table_lines')
return list
def get_cpu(self):
......@@ -152,13 +159,13 @@ def get_cpu(self):
lscpu_dict = {}
try:
# get general cpu information from lscpu
lscpu = self.__run_cmd('lscpu').splitlines()
lscpu = self._run_cmd('lscpu').splitlines()
# get distinct max_speed and current_speed of cpus from dmidecode
speed = self.__run_cmd(r'dmidecode -t processor | grep "Speed"').splitlines()
lscpu_dict = self.__parse_key_value_lines(lscpu)
lscpu_dict.update(self.__parse_key_value_lines(speed))
speed = self._run_cmd(r'dmidecode -t processor | grep "Speed"').splitlines()
lscpu_dict = self._parse_key_value_lines(lscpu)
lscpu_dict.update(self._parse_key_value_lines(speed))
except Exception:
print('Error: get CPU info failed')
logger.exception('Error: get CPU info failed')
return lscpu_dict
def get_system(self):
......@@ -169,27 +176,27 @@ def get_system(self):
"""
system_dict = {}
try:
lsmod = self.__run_cmd('lsmod').splitlines()
lsmod = self.__parse_table_lines(lsmod, key=['Module', 'Size', 'Used', 'by'])
sysctl = self.__run_cmd('sysctl -a').splitlines()
sysctl = self.__parse_key_value_lines(sysctl, None, None, '=')
system_dict['system_manufacturer'] = self.__run_cmd('dmidecode -s system-manufacturer').strip()
system_dict['system_product'] = self.__run_cmd('dmidecode -s system-product-name').strip()
system_dict['os'] = self.__run_cmd('cat /proc/version').strip()
system_dict['uname'] = self.__run_cmd('uname -a').strip()
system_dict['docker'] = self.__get_docker_version()
lsmod = self._run_cmd('lsmod').splitlines()
lsmod = self._parse_table_lines(lsmod, key=['Module', 'Size', 'Used', 'by'])
sysctl = self._run_cmd('sysctl -a').splitlines()
sysctl = self._parse_key_value_lines(sysctl, None, None, '=')
system_dict['system_manufacturer'] = self._run_cmd('dmidecode -s system-manufacturer').strip()
system_dict['system_product'] = self._run_cmd('dmidecode -s system-product-name').strip()
system_dict['os'] = self._run_cmd('cat /proc/version').strip()
system_dict['uname'] = self._run_cmd('uname -a').strip()
system_dict['docker'] = self.get_docker_version()
system_dict['kernel_parameters'] = sysctl
system_dict['kernel_modules'] = lsmod
system_dict['dmidecode'] = self.__run_cmd('dmidecode').strip()
system_dict['dmidecode'] = self._run_cmd('dmidecode').strip()
if system_dict['system_product'] == 'Virtual Machine':
lsvmbus = self.__run_cmd('lsvmbus').splitlines()
lsvmbus = self.__parse_key_value_lines(lsvmbus)
lsvmbus = self._run_cmd('lsvmbus').splitlines()
lsvmbus = self._parse_key_value_lines(lsvmbus)
system_dict['vmbus'] = lsvmbus
except Exception:
print('Error: get system info failed')
logger.exception('Error: get system info failed')
return system_dict
def __get_docker_version(self):
def get_docker_version(self):
"""Get docker version info.
Returns:
......@@ -197,7 +204,7 @@ def __get_docker_version(self):
"""
docker_version_dict = {}
try:
docker_version = self.__run_cmd('docker version')
docker_version = self._run_cmd('docker version')
lines = docker_version.splitlines()
key = ''
......@@ -209,7 +216,7 @@ def __get_docker_version(self):
elif 'Version' in line and key not in docker_version_dict:
docker_version_dict[key] = line.split(':')[1].strip().strip('\t')
except Exception:
print('Error: get docker info failed')
logger.exception('Error: get docker info failed')
return docker_version_dict
def get_memory(self):
......@@ -220,14 +227,14 @@ def get_memory(self):
"""
memory_dict = {}
try:
lsmem = self.__run_cmd('lsmem')
lsmem = self._run_cmd('lsmem')
lsmem = lsmem.splitlines()
lsmem = self.__parse_key_value_lines(lsmem)
lsmem = self._parse_key_value_lines(lsmem)
memory_dict['block_size'] = lsmem.get('Memory block size', '')
memory_dict['total_capacity'] = lsmem.get('Total online memory', '')
dmidecode_memory = self.__run_cmd('dmidecode --type memory')
dmidecode_memory = self._run_cmd('dmidecode --type memory')
dmidecode_memory = dmidecode_memory.splitlines()
model = self.__parse_key_value_lines(
model = self._parse_key_value_lines(
dmidecode_memory, ['Manufacturer', 'Part Number', 'Type', 'Speed', 'Number Of Devices'],
omitted_values=['other', 'unknown']
)
......@@ -236,52 +243,47 @@ def get_memory(self):
memory_dict['clock_frequency'] = model.get('Speed', '')
memory_dict['model'] = model.get('Manufacturer', [''])[0] + ' ' + model.get('Part Number', [''])[0]
except Exception:
print('Error: get memory info failed')
logger.exception('Error: get memory info failed')
return memory_dict
def __get_gpu_nvidia(self):
def get_gpu_nvidia(self):
"""Get nvidia gpu info.
Returns:
dict: nvidia gpu info dict.
"""
gpu_dict = {}
try:
gpu_query = self.__run_cmd('nvidia-smi -q -x')
gpu_query = self._run_cmd('nvidia-smi -q -x')
gpu_query = xmltodict.parse(gpu_query).get('nvidia_smi_log', '')
gpu_dict['gpu_count'] = gpu_query.get('attached_gpus', '')
gpu_dict['nvidia_info'] = gpu_query
gpu_dict['topo'] = self.__run_cmd('nvidia-smi topo -m')
gpu_dict['nvidia-container-runtime_version'] = self.__run_cmd('nvidia-container-runtime -v').strip()
gpu_dict['nvidia-fabricmanager_version'] = self.__run_cmd('nv-fabricmanager --version').strip()
gpu_dict['nv_peer_mem_version'] = self.__run_cmd(
gpu_dict['topo'] = self._run_cmd('nvidia-smi topo -m')
gpu_dict['nvidia-container-runtime_version'] = self._run_cmd('nvidia-container-runtime -v').strip()
gpu_dict['nvidia-fabricmanager_version'] = self._run_cmd('nv-fabricmanager --version').strip()
gpu_dict['nv_peer_mem_version'] = self._run_cmd(
'dpkg -l | grep \'nvidia-peer-memory \' | awk \'$2=="nvidia-peer-memory" {print $3}\''
).strip()
except Exception:
print('Error: get nvidia gpu info failed')
return gpu_dict
def __get_gpu_amd(self):
def get_gpu_amd(self):
"""Get amd gpu info.
Returns:
dict: amd gpu info dict.
"""
gpu_dict = {}
try:
gpu_query = self.__run_cmd('rocm-smi -a --json')
gpu_query = self._run_cmd('rocm-smi -a --json')
gpu_query = json.loads(gpu_query)
gpu_per_node = list(filter(lambda x: 'card' in x, gpu_query.keys()))
gpu_dict['gpu_count'] = len(gpu_per_node)
gpu_mem_info = self.__run_cmd('rocm-smi --showmeminfo vram --json')
gpu_mem_info = self._run_cmd('rocm-smi --showmeminfo vram --json')
gpu_mem_info = json.loads(gpu_mem_info)
for card in gpu_per_node:
gpu_query[card].update(gpu_mem_info.get(card))
gpu_dict['rocm_info'] = gpu_query
gpu_dict['topo'] = self.__run_cmd('rocm-smi --showtopo')
except Exception:
print('Error: get amd gpu info failed')
gpu_dict['topo'] = self._run_cmd('rocm-smi --showtopo')
return gpu_dict
def get_gpu(self):
......@@ -290,10 +292,13 @@ def get_gpu(self):
Returns:
dict: gpu info dict.
"""
try:
if Path('/dev/nvidiactl').is_char_device() and Path('/dev/nvidia-uvm').is_char_device():
return self.__get_gpu_nvidia()
return self.get_gpu_nvidia()
if Path('/dev/kfd').is_char_device() and Path('/dev/dri').is_dir():
return self.__get_gpu_amd()
return self.get_gpu_amd()
except Exception:
logger.exception('Error: get gpu info failed')
print('Warning: no gpu detected')
return {}
......@@ -305,10 +310,10 @@ def get_pcie(self):
"""
pcie_dict = {}
try:
pcie_dict['pcie_topo'] = self.__run_cmd('lspci -t -vvv')
pcie_dict['pcie_info'] = self.__run_cmd('lspci -vvv')
pcie_dict['pcie_topo'] = self._run_cmd('lspci -t -vvv')
pcie_dict['pcie_info'] = self._run_cmd('lspci -vvv')
except Exception:
print('Error: get pcie gpu info failed')
logger.exception('Error: get pcie info failed')
return pcie_dict
def get_storage(self): # noqa: C901
......@@ -319,44 +324,45 @@ def get_storage(self): # noqa: C901
"""
storage_dict = {}
try:
fs_info = self.__run_cmd("df -Th | grep -v \'^/dev/loop\'").splitlines()
fs_list = self.__parse_table_lines(fs_info, key=['Filesystem', 'Type', 'Size', 'Avail', 'Mounted'])
fs_info = self._run_cmd("df -Th | grep -v \'^/dev/loop\'").splitlines()
fs_list = self._parse_table_lines(fs_info, key=['Filesystem', 'Type', 'Size', 'Avail', 'Mounted'])
for fs in fs_list:
fs_device = fs.get('Filesystem', 'UNKNOWN')
if fs_device.startswith('/dev'):
fs['Block_size'] = self.__run_cmd('blockdev --getbsz {}'.format(fs_device)).strip()
fs['Block_size'] = self._run_cmd('blockdev --getbsz {}'.format(fs_device)).strip()
fs['4k_alignment'] = ''
partition_ids = self.__run_cmd(
'parted {} print | grep -oE "^[[:blank:]]*[0-9]+"'.format(fs_device)
partition_ids = self._run_cmd(
'yes Cancel | parted {} print | grep -oE "^[[:blank:]]*[0-9]+"'.format(fs_device)
).splitlines()
for id in partition_ids:
fs['4k_alignment'] += self.__run_cmd('parted {} align-check opt {}'.format(fs_device,
id)).strip()
fs['4k_alignment'] += self._run_cmd(
'yes Cancel | parted {} align-check opt {}'.format(fs_device, id)
).strip()
storage_dict['file_system'] = fs_list
except Exception:
print('Error: get file system info failed')
logger.exception('Error: get file system info failed')
try:
disk_info = self.__run_cmd("lsblk -e 7 -o NAME,ROTA,SIZE,MODEL | grep -v \'^/dev/loop\'").splitlines()
disk_list = self.__parse_table_lines(disk_info, key=['NAME', 'ROTA', 'SIZE', 'MODEL'])
disk_info = self._run_cmd("lsblk -e 7 -o NAME,ROTA,SIZE,MODEL | grep -v \'^/dev/loop\'").splitlines()
disk_list = self._parse_table_lines(disk_info, key=['NAME', 'ROTA', 'SIZE', 'MODEL'])
for disk in disk_list:
block_device = disk.get('NAME', 'UNKNOWN').strip('\u251c\u2500').strip('\u2514\u2500')
disk['NAME'] = block_device
disk['Rotational'] = disk.pop('ROTA')
disk['Block_size'] = self.__run_cmd('fdisk -l -u /dev/{} | grep "Sector size"'.format(block_device)
disk['Block_size'] = self._run_cmd('fdisk -l -u /dev/{} | grep "Sector size"'.format(block_device)
).strip()
if 'nvme' in block_device:
nvme_info = self.__run_cmd('nvme list | grep {}'.format(block_device)).strip().split()
nvme_info = self._run_cmd('nvme list | grep {}'.format(block_device)).strip().split()
if len(nvme_info) >= 15:
disk['Nvme_usage'] = nvme_info[-11] + nvme_info[-10]
storage_dict['block_device'] = disk_list
storage_dict['mapping_bwtween_filesystem_and_blockdevice'] = self.__run_cmd('mount')
storage_dict['mapping_bwtween_filesystem_and_blockdevice'] = self._run_cmd('mount')
except Exception:
print('Error: get block device info failed')
logger.exception('Error: get block device info failed')
return storage_dict
def __get_ib(self):
def get_ib(self):
"""Get available IB devices info.
Return:
......@@ -364,19 +370,19 @@ def __get_ib(self):
"""
ib_dict = {}
try:
ibstat = self.__run_cmd('ibstat').splitlines()
ib_dict['ib_device_status'] = self.__parse_key_value_lines(ibstat)
ibv_devinfo = self.__run_cmd('ibv_devinfo -v').splitlines()
ibstat = self._run_cmd('ibstat').splitlines()
ib_dict['ib_device_status'] = self._parse_key_value_lines(ibstat)
ibv_devinfo = self._run_cmd('ibv_devinfo -v').splitlines()
for i in range(len(ibv_devinfo) - 1, -1, -1):
if ':' not in ibv_devinfo[i]:
ibv_devinfo[i - 1] = ibv_devinfo[i - 1] + ',' + ibv_devinfo[i].strip('\t')
ibv_devinfo.remove(ibv_devinfo[i])
ib_dict['ib_device_info'] = self.__parse_key_value_lines(ibv_devinfo)
except Exception as e:
print('Error: get ib info failed. message: {}.'.format(str(e)))
ib_dict['ib_device_info'] = self._parse_key_value_lines(ibv_devinfo)
except Exception:
logger.exception('Error: get ib info failed')
return ib_dict
def __get_nic(self):
def get_nic(self):
"""Get nic info.
Returns:
......@@ -384,8 +390,10 @@ def __get_nic(self):
"""
nic_list = []
try:
lsnic_xml = self.__run_cmd('lshw -c network -xml')
lsnic_xml = self._run_cmd('lshw -c network -xml')
lsnic_list = xmltodict.parse(lsnic_xml).get('list', {}).get('node', [])
if not isinstance(lsnic_list, list):
lsnic_list = [lsnic_list]
lsnic_list = list(filter(lambda x: 'logicalname' in x, lsnic_list))
for nic in lsnic_list:
......@@ -404,15 +412,15 @@ def __get_nic(self):
'driverversion', ''
)
nic_info['firmware'] = configuration_dict.get('firmware', '')
speed = self.__run_cmd('cat /sys/class/net/{}/speed'.format(nic_info['logical_name'])).strip()
speed = self._run_cmd('cat /sys/class/net/{}/speed'.format(nic_info['logical_name'])).strip()
if speed.isdigit():
nic_info['speed'] = str(int(speed) / 1000) + ' Gbit/s'
except Exception:
print('Error: get nic device {} info failed'.format(nic_info['logical_name']))
logger.exception('Error: get nic device {} info failed')
nic_list.append(nic_info)
except Exception:
print('Error: get nic info failed')
return nic_list
logger.exception('Error: get nic info failed')
def get_network(self):
"""Get network info, including nic info, ib info and ofed version.
......@@ -421,15 +429,21 @@ def get_network(self):
dict: dict of network info.
"""
network_dict = {}
network_dict['nic'] = self.__get_nic()
network_dict['ib'] = self.__get_ib()
ofed_version = self.__run_cmd('ofed_info -s').strip()
try:
network_dict['nic'] = self.get_nic()
network_dict['ib'] = self.get_ib()
ofed_version = self._run_cmd('ofed_info -s').strip()
network_dict['ofed_version'] = ofed_version
except Exception:
logger.exception('Error: get network info failed')
return network_dict
def get_all(self):
"""Get all system info and save them to file in json format."""
sum_dict = {}
if os.geteuid() != 0:
logger.error('You need to be as a root user to run this tool.')
return sum_dict
sum_dict['System'] = self.get_system()
sum_dict['CPU'] = self.get_cpu()
sum_dict['Memory'] = self.get_memory()
......
......@@ -81,3 +81,7 @@ def test_sb_run_nonexist_host_file(self):
"""Test sb run, --host-file does not exist, should fail."""
result = self.cmd('sb run --host-file ./nonexist.yaml', expect_failure=True)
self.assertEqual(result.exit_code, 1)
def test_sb_node_info(self):
"""Test sb node info, should fail."""
self.cmd('sb node info', expect_failure=False)
......@@ -116,8 +116,6 @@ def test_get_mode_command(self):
'expected_command': (
'python3 -m torch.distributed.launch '
'--use_env --no_python --nproc_per_node=8 '
'--nnodes=1 --node_rank=$NODE_RANK '
'--master_addr=$MASTER_ADDR --master_port=$MASTER_PORT '
f'sb exec --output-dir {self.sb_output_dir} -c sb.config.yaml -C superbench.enable=foo '
'superbench.benchmarks.foo.parameters.distributed_impl=ddp '
'superbench.benchmarks.foo.parameters.distributed_backend=nccl'
......
......@@ -8,7 +8,6 @@ MPI_HOME ?= /usr/local/mpi
HIP_HOME ?= /opt/rocm/hip
RCCL_HOME ?= /opt/rocm/rccl
ROCM_VERSION ?= rocm-$(shell dpkg -l | grep 'rocm-dev ' | awk '{print $$3}' | cut -d '.' -f1-3)
ROCM_ARCH ?= $(shell rocminfo | grep " gfx" | uniq | awk '{print $$2}')
.PHONY: all cuda rocm common cuda_cutlass cuda_bandwidthTest cuda_nccl_tests cuda_perftest rocm_perftest fio rocm_rccl_tests rocm_rocblas rocm_bandwidthTest
......@@ -66,7 +65,7 @@ ifneq (,$(wildcard fio/Makefile))
cd ./fio && ./configure --prefix=$(SB_MICRO_PATH) && make -j && make install
endif
# Build rccl-tests from commit cc34c5 of develop branch (default branch).
# Build rccl-tests from commit dc1ad48 of develop branch (default branch).
rocm_rccl_tests: sb_micro_path
ifneq (, $(wildcard rccl-tests/Makefile))
cd ./rccl-tests && make MPI=1 MPI_HOME=$(MPI_HOME) HIP_HOME=$(HIP_HOME) RCCL_HOME=$(RCCL_HOME) -j
......@@ -81,21 +80,14 @@ rocm_rocblas: sb_micro_path
ifeq (, $(wildcard $(SB_MICRO_PATH)/bin/rocblas-bench))
if [ -d rocBLAS ]; then rm -rf rocBLAS; fi
git clone -b ${ROCM_VERSION} https://github.com/ROCmSoftwarePlatform/rocBLAS.git ./rocBLAS
ifeq (${ROCM_VERSION}, rocm-4.0.0)
sed -i '/CMAKE_MATCH_1/a\ get_filename_component(HIP_CLANG_ROOT "$${HIP_CLANG_ROOT}" DIRECTORY)' /opt/rocm/hip/lib/cmake/hip/hip-config.cmake
cd ./rocBLAS && HIPCC_COMPILE_FLAGS_APPEND="-D_OPENMP=201811 -O3 -Wno-format-nonliteral -DCMAKE_HAVE_LIBC_PTHREAD -parallel-jobs=2" HIPCC_LINK_FLAGS_APPEND="-lpthread -O3 -parallel-jobs=2" ./install.sh -idc -a ${ROCM_ARCH}
else
cd ./rocBLAS && ./install.sh -idc
endif
cd ./rocBLAS && ./install.sh --dependencies --clients-only
cp -v ./rocBLAS/build/release/clients/staging/rocblas-bench $(SB_MICRO_PATH)/bin/
endif
# Build hipBusBandwidth.
# HIP is released with rocm, like rocm-4.2.0 and so on.
# The version we use is the released tag which is consistent with the rocm version in the environment or docker.
rocm_bandwidthTest:
cp -r -v $(shell hipconfig -p) ./
ifneq (, $(wildcard hip/samples/1_Utils/hipBusBandwidth/CMakeLists.txt))
cd ./hip/samples/1_Utils/hipBusBandwidth/ && mkdir -p build && cd build && cmake .. && make
cp -v ./hip/samples/1_Utils/hipBusBandwidth/build/hipBusBandwidth $(SB_MICRO_PATH)/bin/
endif
rocm_bandwidthTest: sb_micro_path
cp -r -v $(shell hipconfig -p)/samples/1_Utils/hipBusBandwidth ./
cd ./hipBusBandwidth/ && mkdir -p build && cd build && cmake .. && make
cp -v ./hipBusBandwidth/build/hipBusBandwidth $(SB_MICRO_PATH)/bin/
Subproject commit cc34c545098145bc148e5035e4c8e767b4d71ece
Subproject commit dc1ad4853d7ec738387d42a75a58a98d7af00c7b
---
slug: release-sb-v0.3
title: Releasing SuperBench v0.3
author: Peng Cheng
author_title: SuperBench Team
author_url: https://github.com/cp5555
author_image_url: https://github.com/cp5555.png
tags: [superbench, announcement, release]
---
We are very happy to announce that **SuperBench 0.3.0 version** is officially released today!
You can install and try superbench by following [Getting Started Tutorial](https://microsoft.github.io/superbenchmark/docs/getting-started/installation).
## SuperBench 0.3.0 Release Notes
### SuperBench Framework
#### Runner
- Implement MPI mode.
#### Benchmarks
- Support Docker benchmark.
### Single-node Validation
#### Micro Benchmarks
1. Memory (Tool: NVIDIA/AMD Bandwidth Test Tool)
| Metrics | Unit | Description |
|----------------|------|-------------------------------------|
| H2D_Mem_BW_GPU | GB/s | host-to-GPU bandwidth for each GPU |
| D2H_Mem_BW_GPU | GB/s | GPU-to-host bandwidth for each GPU |
2. IBLoopback (Tool: PerfTest – Standard RDMA Test Tool)
| Metrics | Unit | Description |
|----------|------|---------------------------------------------------------------|
| IB_Write | MB/s | The IB write loopback throughput with different message sizes |
| IB_Read | MB/s | The IB read loopback throughput with different message sizes |
| IB_Send | MB/s | The IB send loopback throughput with different message sizes |
3. NCCL/RCCL (Tool: NCCL/RCCL Tests)
| Metrics | Unit | Description |
|---------------------|------|-----------------------------------------------------------------|
| NCCL_AllReduce | GB/s | The NCCL AllReduce performance with different message sizes |
| NCCL_AllGather | GB/s | The NCCL AllGather performance with different message sizes |
| NCCL_broadcast | GB/s | The NCCL Broadcast performance with different message sizes |
| NCCL_reduce | GB/s | The NCCL Reduce performance with different message sizes |
| NCCL_reduce_scatter | GB/s | The NCCL ReduceScatter performance with different message sizes |
4. Disk (Tool: FIO – Standard Disk Performance Tool)
| Metrics | Unit | Description |
|----------------|------|---------------------------------------------------------------------------------|
| Seq_Read | MB/s | Sequential read performance |
| Seq_Write | MB/s | Sequential write performance |
| Rand_Read | MB/s | Random read performance |
| Rand_Write | MB/s | Random write performance |
| Seq_R/W_Read | MB/s | Read performance in sequential read/write, fixed measurement (read:write = 4:1) |
| Seq_R/W_Write | MB/s | Write performance in sequential read/write (read:write = 4:1) |
| Rand_R/W_Read | MB/s | Read performance in random read/write (read:write = 4:1) |
| Rand_R/W_Write | MB/s | Write performance in random read/write (read:write = 4:1) |
5. H2D/D2H SM Transmission Bandwidth (Tool: MSR-A build)
| Metrics | Unit | Description |
|---------------|------|-----------------------------------------------------|
| H2D_SM_BW_GPU | GB/s | host-to-GPU bandwidth using GPU kernel for each GPU |
| D2H_SM_BW_GPU | GB/s | GPU-to-host bandwidth using GPU kernel for each GPU |
### AMD GPU Support
#### Docker Image Support
- ROCm 4.2 PyTorch 1.7.0
- ROCm 4.0 PyTorch 1.7.0
#### Micro Benchmarks
1. Kernel Launch (Tool: MSR-A build)
| Metrics | Unit | Description |
|--------------------------|-----------|--------------------------------------------------------------|
| Kernel_Launch_Event_Time | Time (ms) | Dispatch latency measured in GPU time using hipEventRecord() |
| Kernel_Launch_Wall_Time | Time (ms) | Dispatch latency measured in CPU time |
2. GEMM FLOPS (Tool: AMD rocblas-bench Tool)
| Metrics | Unit | Description |
|----------|--------|-------------------------------|
| FP64 | GFLOPS | FP64 FLOPS without MatrixCore |
| FP32(MC) | GFLOPS | TF32 FLOPS with MatrixCore |
| FP16(MC) | GFLOPS | FP16 FLOPS with MatrixCore |
| BF16(MC) | GFLOPS | BF16 FLOPS with MatrixCore |
| INT8(MC) | GOPS | INT8 FLOPS with MatrixCore |
#### E2E Benchmarks
1. CNN models -- Use PyTorch torchvision models
- ResNet: ResNet-50, ResNet-101, ResNet-152
- DenseNet: DenseNet-169, DenseNet-201
- VGG: VGG-11, VGG-13, VGG-16, VGG-19​
2. BERT -- Use huggingface Transformers
- BERT
- BERT Large
3. LSTM -- Use PyTorch
4. GPT-2 -- Use huggingface Transformers
### Bug Fix
- VGG models failed on A100 GPU with batch_size=128
### Other Improvement
1. Contribution related
- Contribute rule
- System information collection
2. Document
- Add release process doc
- Add design documents
- Add developer guide doc for coding style
- Add contribution rules
- Add docker image list
- Add initial validation results
......@@ -101,7 +101,7 @@ module.exports = {
announcementBar: {
id: 'supportus',
content:
'📢 <a href="https://microsoft.github.io/superbenchmark/blog/release-sb-v0.2">v0.2.1</a> has been released! ' +
'📢 <a href="https://microsoft.github.io/superbenchmark/blog/release-sb-v0.3">v0.3.0</a> has been released! ' +
'⭐️ If you like SuperBench, give it a star on <a target="_blank" rel="noopener noreferrer" href="https://github.com/microsoft/superbenchmark">GitHub</a>! ⭐️',
},
algolia: {
......
{
"name": "superbench-website",
"version": "0.2.1",
"version": "0.3.0",
"lockfileVersion": 1,
"requires": true,
"dependencies": {
......
{
"name": "superbench-website",
"version": "0.2.1",
"version": "0.3.0",
"private": true,
"scripts": {
"docusaurus": "docusaurus",
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment