[Tarantool-patches] [PATCH v1 luajit 37/41] perf: add CMake infrastructure
Sergey Bronnikov
sergeyb at tarantool.org
Tue Nov 18 15:21:21 MSK 2025
Hi, Sergey,
thanks for the patch! See my comments.
Sergey
On 10/24/25 14:00, Sergey Kaplun wrote:
> This commit introduces CMake building scripts for the benches introduced
> before. The benchmarks are enabled only if `LUAJIT_ENABLE_PERF` option
> is set. For each suite (LuaJIT-benches in this patch set)
> `AddBenchTarget()` macro generates 2 targets:
> * Target to run all benches and store results in the
> perf/output/<suite_name> directory.
> * Target to run all benches via CTest and inspect results in the
> console.
>
> For the LuaJIT-benches there are 2 generated files:
> * FASTA_5000000 -- is used as an input for <k-nukleotide.lua> and
> <revcomp.lua>.
> * SUMCOLL_5000.txt -- is used as an input for <sum-file.lua>.
>
> These files and <perf/output> directory are added to the .gitignore files.
> ---
> .gitignore | 5 ++
> CMakeLists.txt | 11 ++++
> perf/CMakeLists.txt | 99 ++++++++++++++++++++++++++++++
> perf/LuaJIT-benches/CMakeLists.txt | 52 ++++++++++++++++
> 4 files changed, 167 insertions(+)
> create mode 100644 perf/CMakeLists.txt
> create mode 100644 perf/LuaJIT-benches/CMakeLists.txt
>
> diff --git a/.gitignore b/.gitignore
> index c26a7eb8..bfc7d401 100644
> --- a/.gitignore
> +++ b/.gitignore
> @@ -28,3 +28,8 @@ luajit-parse-memprof
> luajit-parse-sysprof
> luajit.pc
> *.c_test
> +
> +# Generated by the performance tests.
> +FASTA_5000000
> +SUMCOL_5000.txt
> +perf/output/
> diff --git a/CMakeLists.txt b/CMakeLists.txt
> index c0da4362..73f46835 100644
> --- a/CMakeLists.txt
> +++ b/CMakeLists.txt
> @@ -464,6 +464,17 @@ if(LUAJIT_USE_TEST)
> endif()
> add_subdirectory(test)
>
> +# --- Benchmarks source tree ---------------------------------------------------
> +
> +# The option to enable performance tests for the LuaJIT.
> +# Disabled by default, since commonly it is used only by LuaJIT
> +# developers and run in the CI with the specially set-up machine.
> +option(LUAJIT_ENABLE_PERF "Generate <perf> target" OFF)
> +
> +if(LUAJIT_ENABLE_PERF)
option name confuses a bit due to `perf` utility.
I would rename to something like "LUAJIT_ENABLE_PERF_TESTS".
Feel free to ignore.
> + add_subdirectory(perf)
> +endif()
> +
> # --- Misc rules ---------------------------------------------------------------
>
> # XXX: Implement <uninstall> target using the following recipe:
> diff --git a/perf/CMakeLists.txt b/perf/CMakeLists.txt
> new file mode 100644
> index 00000000..cc3c312f
> --- /dev/null
> +++ b/perf/CMakeLists.txt
> @@ -0,0 +1,99 @@
> +# Running various bench suites against LuaJIT.
> +
> +include(MakeLuaPath)
> +
> +if(CMAKE_BUILD_TYPE STREQUAL "Debug")
> + message(WARNING "LuaJIT and perf tests are built in the Debug mode."
s/./. /
missed whitespace after dot
> + "Timings may be affected.")
> +endif()
> +
> +set(PERF_OUTPUT_DIR ${PROJECT_BINARY_DIR}/perf/output)
> +file(MAKE_DIRECTORY ${PERF_OUTPUT_DIR})
> +
> +# List of paths that will be used for each suite.
> +make_lua_path(LUA_PATH_BENCH_BASE
> + PATHS
> + # Use of the bench module.
> + ${CMAKE_CURRENT_SOURCE_DIR}/utils/?.lua
> + # Simple usage with `jit.dump()`, etc.
> + ${LUAJIT_SOURCE_DIR}/?.lua
> + ${LUAJIT_BINARY_DIR}/?.lua
> +)
> +
> +make_lua_path(LUA_CPATH
> + PATHS
> + # XXX: Some arches may have installed the cjson module here.
> + /usr/lib64/lua/5.1/?.so
> +)
> +
> +# Produce the pair:
> +# Target to run for reporting and target to inspect from the
> +# console, runnable by the CTest.
> +macro(AddBenchTarget perf_suite)
> + file(MAKE_DIRECTORY "${PERF_OUTPUT_DIR}/${perf_suite}/")
> + message(STATUS "Add perf suite ${perf_suite}")
> + add_custom_target(${perf_suite})
> + add_custom_target(${perf_suite}-console
> + COMMAND ${CMAKE_CTEST_COMMAND}
> + -L ${perf_suite}
> + --parallel 1
> + --verbose
> + --output-on-failure
> + --no-tests=error
may be --schedule-random, --timeout XXX (default timeout is 10000000)?
> + )
> + add_dependencies(${perf_suite}-console luajit-main)
> +endmacro()
> +
> +# Add the bench to the pair of targets created by the call above.
> +macro(AddBench bench_name bench_path perf_suite LUA_PATH)
> + set(bench_title "perf/${perf_suite}/${bench_name}")
> + get_filename_component(bench_name_stripped ${bench_name} NAME_WE)
> + set(bench_out_file
> + ${PERF_OUTPUT_DIR}/${perf_suite}/${bench_name_stripped}.json
> + )
> + set(bench_command "${LUAJIT_BINARY} ${bench_path}")
> + if(${ARGC} GREATER 4)
> + set(input_file ${ARGV4})
> + set(bench_command "${bench_command} < ${input_file}")
> + endif()
> + set(BENCH_FLAGS
> + "--benchmark_out_format=json --benchmark_out=${bench_out_file}"
> + )
> + set(bench_command_flags ${bench_command} ${BENCH_FLAGS})
> + separate_arguments(bench_command_separated UNIX_COMMAND ${bench_command})
> + add_custom_command(
> + COMMAND ${CMAKE_COMMAND} -E env
> + LUA_PATH="${LUA_PATH}"
> + LUA_CPATH="${LUA_CPATH}"
> + ${bench_command_separated}
> + --benchmark_out_format=json
> + --benchmark_out="${bench_out_file}"
previous two lines can be replaced with ${BENCH_FLAGS}, right?
> + OUTPUT ${bench_out_file}
> + DEPENDS luajit-main
> + COMMENT
> + "Running benchmark ${bench_title} saving results in ${bench_out_file}."
> + )
> + add_custom_target(${bench_name} DEPENDS ${bench_out_file})
> + add_dependencies(${perf_suite} ${bench_name})
> +
> + # Report in the console.
> + add_test(NAME ${bench_title}
> + COMMAND sh -c "${bench_command}"
> + )
> + set_tests_properties(${bench_title} PROPERTIES
> + ENVIRONMENT "LUA_PATH=${LUA_PATH}"
> + LABELS ${perf_suite}
> + DEPENDS luajit-main
> + )
> + unset(input_file)
> +endmacro()
> +
> +add_subdirectory(LuaJIT-benches)
> +
> +add_custom_target(${PROJECT_NAME}-perf
> + DEPENDS LuaJIT-benches
missed a COMMENT field
> +)
> +
> +add_custom_target(${PROJECT_NAME}-perf-console
> + DEPENDS LuaJIT-benches-console
missed a COMMENT field
> +)
> diff --git a/perf/LuaJIT-benches/CMakeLists.txt b/perf/LuaJIT-benches/CMakeLists.txt
> new file mode 100644
> index 00000000..d9909f36
> --- /dev/null
> +++ b/perf/LuaJIT-benches/CMakeLists.txt
> @@ -0,0 +1,52 @@
> +set(PERF_SUITE_NAME LuaJIT-benches)
> +set(LUA_BENCH_SUFFIX .lua)
it is not a bench-specific suffix. May be LUA_SUFFIX?
> +
> +AddBenchTarget(${PERF_SUITE_NAME})
> +
> +# Input for the k-nucleotide and revcomp benchmarks.
> +set(FASTA_NAME ${CMAKE_CURRENT_BINARY_DIR}/FASTA_5000000)
> +add_custom_target(FASTA_5000000
> + COMMAND ${LUAJIT_BINARY}
> + ${CMAKE_CURRENT_SOURCE_DIR}/libs/fasta.lua 5000000 > ${FASTA_NAME}
FASTA_5000000 is a plain text file. I propose to add extension .txt for
its full name and
probably postfix "_autogenerated". Like we do this for SUMCOL_5000 and
SUMCOL_1.
> + OUTPUT ${FASTA_NAME}
> + DEPENDS luajit-main
> + COMMENT "Generate ${FASTA_NAME}."
> +)
> +
> +make_lua_path(LUA_PATH
> + PATHS
> + ${LUA_PATH_BENCH_BASE}
> + ${CMAKE_CURRENT_SOURCE_DIR}/libs/?.lua
> +)
> +
> +# Input for the <sum-file.lua> benchmark.
> +set(SUM_NAME ${CMAKE_CURRENT_BINARY_DIR}/SUMCOL_5000.txt)
> +# Remove possibly existing file.
> +file(REMOVE ${SUM_NAME})
Why do we need generate file after every cmake configuration?
I propose to skip generation if file already exist or regenerate if
SHA256 is not the same.
> +
> +set(SUMCOL_FILE ${CMAKE_CURRENT_SOURCE_DIR}/SUMCOL_1.txt)
> +file(READ ${SUMCOL_FILE} SUMCOL_CONTENT)
> +foreach(_unused RANGE 4999)
> + file(APPEND ${SUM_NAME} "${SUMCOL_CONTENT}")
> +endforeach()
> +
> +file(GLOB benches "${CMAKE_CURRENT_SOURCE_DIR}/*${LUA_BENCH_SUFFIX}")
> +foreach(bench_path ${benches})
> + file(RELATIVE_PATH bench_name ${CMAKE_CURRENT_SOURCE_DIR} ${bench_path})
> + set(bench_title "perf/${PERF_SUITE_NAME}/${bench_name}")
> + if(bench_name MATCHES "k-nucleotide" OR bench_name MATCHES "revcomp")
> + AddBench(${bench_name}
> + ${bench_path} ${PERF_SUITE_NAME} "${LUA_PATH}" ${FASTA_NAME}
> + )
> + add_dependencies(${bench_name} FASTA_5000000)
> + elseif(bench_name MATCHES "sum-file")
> + AddBench(${bench_name}
> + ${bench_path} ${PERF_SUITE_NAME} "${LUA_PATH}" ${SUM_NAME}
> + )
> + else()
> + AddBench(${bench_name} ${bench_path} ${PERF_SUITE_NAME} "${LUA_PATH}")
> + endif()
> +endforeach()
> +
> +# We need to generate the file before we run tests.
> +add_dependencies(${PERF_SUITE_NAME}-console FASTA_5000000)
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.tarantool.org/pipermail/tarantool-patches/attachments/20251118/62f38777/attachment.htm>
More information about the Tarantool-patches
mailing list