[QAAS] timestamp = ip-172-31-46-37.ec2.internal app_name = llama.cpp git_commit = dataset_name = Llama-3.1-8B-Q4_0 PP=128 NPL=1 run_cmd = -m meta-llama-3.1-8b-instruct-Q4_0.gguf -t -b 2048 -ub 512 -npp 128 -ntg 0 -npl 1 -c 16384 --seed 0 --output-format jsonl LANG = C/CXX [REPORTS] figure_of_merit_type = RATE figure_of_merit_unit = tokens/s compiler_default = armclang multicompiler_report = qaas_compilers.csv mpi_scaling = no openmp_scaling = strong scalability_report = scalability_reference_line = [SYSTEM] machine = ip-172-31-46-37.ec2.internal model_name = Unknown CPU model name ISA = aarch64 architecture = ARM_NEOVERSE_V1 number_of_cpus = 64 number_of_cores = 64 number_of_sockets = 1 number_of_cores_per_socket = 64 number_of_numa_domains = 1 frequency_driver = Unknown frequency driver frequency_governor = Unknown frequency governor scaling_max_frequency = Unknown scaling max frequency scaling_min_frequency = Unknown scaling min frequency advertized_frequency = unsupported maximal_frequency = Unknown maximal frequency huge_pages = gcc_version = 11.5.0 armclang_version = 21.1 mpi_provider = OpenMPI mpi_version = 5.0.9 [TIME] initial_profile = 00H03M11S build_binaries = 00H05M12S multicompiler = 00H01M39S