mirror of
https://github.com/enso-org/enso.git
synced 2024-12-23 03:21:44 +03:00
Small wins in Engine CI job execution (#7846)
Reorder steps of Engine tests, run dry-run benchmarks only on Linux. --------- Co-authored-by: Pavel Marek <pavel.marek@enso.org> Co-authored-by: Michał W. Urbańczyk <mwu-tow@gazeta.pl>
This commit is contained in:
parent
7a31dcdd58
commit
df4183e78e
@ -127,7 +127,7 @@ pub struct BuildConfigurationFlags {
|
||||
pub build_benchmarks: bool,
|
||||
/// Whether the Enso-written benchmarks should be checked whether they compile.
|
||||
///
|
||||
/// Note that this does not benchmark, only ensures that they are buildable.
|
||||
/// Note that this does not run benchmark, only ensures that they are buildable.
|
||||
/// Also, this does nothing if `execute_benchmarks` contains `Benchmarks::Enso`.
|
||||
pub check_enso_benchmarks: bool,
|
||||
/// Which benchmarks should be run.
|
||||
|
@ -290,6 +290,8 @@ impl RunContext {
|
||||
self.config.build_engine_package() && big_memory_machine && TARGET_OS != OS::Windows;
|
||||
|
||||
|
||||
// === Build project-manager distribution and native image ===
|
||||
debug!("Bulding project-manager distribution and Native Image");
|
||||
if big_memory_machine {
|
||||
let mut tasks = vec![];
|
||||
|
||||
@ -314,30 +316,7 @@ impl RunContext {
|
||||
if self.config.build_launcher_package() {
|
||||
tasks.push("buildLauncherDistribution");
|
||||
}
|
||||
|
||||
// This just compiles benchmarks, not run them. At least we'll know that they can be
|
||||
// run. Actually running them, as part of this routine, would be too heavy.
|
||||
// TODO [mwu] It should be possible to run them through context config option.
|
||||
if self.config.build_benchmarks {
|
||||
tasks.extend([
|
||||
"runtime/Benchmark/compile",
|
||||
"language-server/Benchmark/compile",
|
||||
"searcher/Benchmark/compile",
|
||||
"std-benchmarks/Benchmark/compile",
|
||||
]);
|
||||
}
|
||||
|
||||
// We want benchmarks to run only after the other build tasks are done, as they are
|
||||
// really CPU-heavy.
|
||||
let build_command = (!tasks.is_empty()).then_some(Sbt::concurrent_tasks(tasks));
|
||||
let benchmark_tasks = self.config.execute_benchmarks.iter().flat_map(|b| b.sbt_task());
|
||||
let command_sequence = build_command.as_deref().into_iter().chain(benchmark_tasks);
|
||||
let final_command = Sbt::sequential_tasks(command_sequence);
|
||||
if !final_command.is_empty() {
|
||||
sbt.call_arg(final_command).await?;
|
||||
} else {
|
||||
debug!("No SBT tasks to run.");
|
||||
}
|
||||
sbt.call_arg(Sbt::concurrent_tasks(tasks)).await?;
|
||||
} else {
|
||||
// If we are run on a weak machine (like GH-hosted runner), we need to build things one
|
||||
// by one.
|
||||
@ -363,27 +342,8 @@ impl RunContext {
|
||||
|
||||
// Prepare Project Manager Distribution
|
||||
sbt.call_arg("buildProjectManagerDistribution").await?;
|
||||
|
||||
if self.config.build_benchmarks {
|
||||
// Check Runtime Benchmark Compilation
|
||||
sbt.call_arg("runtime/Benchmark/compile").await?;
|
||||
|
||||
// Check Language Server Benchmark Compilation
|
||||
sbt.call_arg("language-server/Benchmark/compile").await?;
|
||||
|
||||
// Check Searcher Benchmark Compilation
|
||||
sbt.call_arg("searcher/Benchmark/compile").await?;
|
||||
|
||||
// Check Enso JMH benchmark compilation
|
||||
sbt.call_arg("std-benchmarks/Benchmark/compile").await?;
|
||||
}
|
||||
|
||||
for benchmark in &self.config.execute_benchmarks {
|
||||
if let Some(task) = benchmark.sbt_task() {
|
||||
sbt.call_arg(task).await?;
|
||||
}
|
||||
}
|
||||
} // End of Sbt run.
|
||||
}
|
||||
// === End of Build project-manager distribution and native image ===
|
||||
|
||||
let ret = self.expected_artifacts();
|
||||
|
||||
@ -414,13 +374,87 @@ impl RunContext {
|
||||
}
|
||||
|
||||
let enso = BuiltEnso { paths: self.paths.clone() };
|
||||
|
||||
// === Unit tests and Enso tests ===
|
||||
debug!("Running unit tests and Enso tests.");
|
||||
if self.config.test_scala {
|
||||
// Run unit tests
|
||||
sbt.call_arg("set Global / parallelExecution := false; test").await?;
|
||||
}
|
||||
if self.config.test_standard_library {
|
||||
enso.run_tests(IrCaches::No, &sbt, PARALLEL_ENSO_TESTS).await?;
|
||||
}
|
||||
// If we are run in CI conditions and we prepared some test results, we want to upload
|
||||
// them as a separate artifact to ease debugging.
|
||||
if let Some(test_results_dir) = test_results_dir && is_in_env() {
|
||||
// Each platform gets its own log results, so we need to generate unique names.
|
||||
let name = format!("Test_Results_{TARGET_OS}");
|
||||
if let Err(err) = ide_ci::actions::artifacts::upload_compressed_directory(&test_results_dir, name)
|
||||
.await {
|
||||
// We wouldn't want to fail the whole build if we can't upload the test results.
|
||||
// Still, it should be somehow visible in the build summary.
|
||||
ide_ci::actions::workflow::message(MessageLevel::Warning, format!("Failed to upload test results: {err}"));
|
||||
}
|
||||
}
|
||||
|
||||
perhaps_test_java_generated_from_rust_job.await.transpose()?;
|
||||
|
||||
// === Run benchmarks ===
|
||||
debug!("Running benchmarks.");
|
||||
if big_memory_machine {
|
||||
let mut tasks = vec![];
|
||||
// This just compiles benchmarks, not run them. At least we'll know that they can be
|
||||
// run. Actually running them, as part of this routine, would be too heavy.
|
||||
// TODO [mwu] It should be possible to run them through context config option.
|
||||
if self.config.build_benchmarks {
|
||||
tasks.extend([
|
||||
"runtime/Benchmark/compile",
|
||||
"language-server/Benchmark/compile",
|
||||
"searcher/Benchmark/compile",
|
||||
"std-benchmarks/Benchmark/compile",
|
||||
]);
|
||||
}
|
||||
|
||||
let build_command = (!tasks.is_empty()).then_some(Sbt::concurrent_tasks(tasks));
|
||||
|
||||
// We want benchmarks to run only after the other build tasks are done, as they are
|
||||
// really CPU-heavy.
|
||||
let benchmark_tasks = self.config.execute_benchmarks.iter().flat_map(|b| b.sbt_task());
|
||||
let command_sequence = build_command.as_deref().into_iter().chain(benchmark_tasks);
|
||||
let final_command = Sbt::sequential_tasks(command_sequence);
|
||||
if !final_command.is_empty() {
|
||||
sbt.call_arg(final_command).await?;
|
||||
} else {
|
||||
debug!("No SBT tasks to run.");
|
||||
}
|
||||
} else {
|
||||
if self.config.build_benchmarks {
|
||||
// Check Runtime Benchmark Compilation
|
||||
sbt.call_arg("runtime/Benchmark/compile").await?;
|
||||
|
||||
// Check Language Server Benchmark Compilation
|
||||
sbt.call_arg("language-server/Benchmark/compile").await?;
|
||||
|
||||
// Check Searcher Benchmark Compilation
|
||||
sbt.call_arg("searcher/Benchmark/compile").await?;
|
||||
|
||||
// Check Enso JMH benchmark compilation
|
||||
sbt.call_arg("std-benchmarks/Benchmark/compile").await?;
|
||||
}
|
||||
|
||||
for benchmark in &self.config.execute_benchmarks {
|
||||
if let Some(task) = benchmark.sbt_task() {
|
||||
sbt.call_arg(task).await?;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if self.config.execute_benchmarks.contains(&Benchmarks::Enso) {
|
||||
enso.run_benchmarks(BenchmarkOptions { dry_run: false }).await?;
|
||||
} else if self.config.check_enso_benchmarks {
|
||||
enso.run_benchmarks(BenchmarkOptions { dry_run: true }).await?;
|
||||
}
|
||||
|
||||
|
||||
// If we were running any benchmarks, they are complete by now. Upload the report.
|
||||
if is_in_env() {
|
||||
for bench in &self.config.execute_benchmarks {
|
||||
@ -462,18 +496,9 @@ impl RunContext {
|
||||
}
|
||||
}
|
||||
|
||||
if self.config.test_scala {
|
||||
// Test Enso
|
||||
sbt.call_arg("set Global / parallelExecution := false; test").await?;
|
||||
}
|
||||
|
||||
perhaps_test_java_generated_from_rust_job.await.transpose()?;
|
||||
|
||||
// === Build Distribution ===
|
||||
if self.config.test_standard_library {
|
||||
enso.run_tests(IrCaches::No, &sbt, PARALLEL_ENSO_TESTS).await?;
|
||||
}
|
||||
|
||||
debug!("Building distribution");
|
||||
if self.config.build_engine_package() {
|
||||
let std_libs =
|
||||
&self.repo_root.built_distribution.enso_engine_triple.engine_package.lib.standard;
|
||||
@ -487,22 +512,6 @@ impl RunContext {
|
||||
}
|
||||
}
|
||||
|
||||
if self.config.test_standard_library {
|
||||
enso.run_tests(IrCaches::Yes, &sbt, PARALLEL_ENSO_TESTS).await?;
|
||||
}
|
||||
|
||||
// If we are run in CI conditions and we prepared some test results, we want to upload
|
||||
// them as a separate artifact to ease debugging.
|
||||
if let Some(test_results_dir) = test_results_dir && is_in_env() {
|
||||
// Each platform gets its own log results, so we need to generate unique names.
|
||||
let name = format!("Test_Results_{TARGET_OS}");
|
||||
if let Err(err) = ide_ci::actions::artifacts::upload_compressed_directory(&test_results_dir, name)
|
||||
.await {
|
||||
// We wouldn't want to fail the whole build if we can't upload the test results.
|
||||
// Still, it should be somehow visible in the build summary.
|
||||
ide_ci::actions::workflow::message(MessageLevel::Warning, format!("Failed to upload test results: {err}"));
|
||||
}
|
||||
}
|
||||
|
||||
// if build_native_runner {
|
||||
// let factorial_input = "6";
|
||||
|
@ -424,7 +424,14 @@ impl Processor {
|
||||
test_standard_library: true,
|
||||
test_java_generated_from_rust: true,
|
||||
build_benchmarks: true,
|
||||
execute_benchmarks: once(Benchmarks::Runtime).collect(),
|
||||
execute_benchmarks: {
|
||||
// Run benchmarks only on Linux.
|
||||
let mut ret = BTreeSet::new();
|
||||
if TARGET_OS == OS::Linux {
|
||||
ret.insert(Benchmarks::Runtime);
|
||||
}
|
||||
ret
|
||||
},
|
||||
execute_benchmarks_once: true,
|
||||
check_enso_benchmarks: true,
|
||||
verify_packages: true,
|
||||
|
Loading…
Reference in New Issue
Block a user