mirror of
https://github.com/rui314/mold.git
synced 2024-12-26 01:44:29 +03:00
576 lines
17 KiB
C++
576 lines
17 KiB
C++
// This file implements the Identical Code Folding feature which can
|
|
// reduce the output file size of a typical program by a few percent.
|
|
// ICF identifies read-only input sections that happen to be identical
|
|
// and thus can be used interchangeably. ICF leaves one of them and discards
|
|
// the others.
|
|
//
|
|
// ICF is usually used in combination with -ffunction-sections and
|
|
// -fdata-sections compiler options, so that object files have one section
|
|
// for each function or variable instead of having one large .text or .data.
|
|
// The unit of ICF merging is section.
|
|
//
|
|
// Two sections are considered identical by ICF if they have the exact
|
|
// same contents, metadata such as section flags, exception handling
|
|
// records, and relocations. The last one is interesting because two
|
|
// relocations are considered identical if they point to the _same_
|
|
// section in terms of ICF.
|
|
//
|
|
// To see what that means, consider two sections, A and B, which are
|
|
// identical except for one pair of relocations. Say, A has a relocation to
|
|
// section C, and B has a relocation to D. In this case, A and B are
|
|
// considered identical if C and D are considered identical. C and D can be
|
|
// either really the same section or two different sections that are
|
|
// considered identical by ICF. Below is an example of such inputs, A, B, C
|
|
// and D:
|
|
//
|
|
// void A() { C(); }
|
|
// void B() { D(); }
|
|
// void C() { A(); }
|
|
// void D() { B(); }
|
|
//
|
|
// If we assume A and B are mergeable, we can merge C and D, which makes A
|
|
// and B mergeable. There's no contradiction in our assumption, so we can
|
|
// conclude that A and B as well as C and D are mergeable.
|
|
//
|
|
// This problem boils down to one in graph theory. Input to ICF can be
|
|
// considered as a directed graph in which vertices are sections and edges
|
|
// are relocations. Vertices have labels (section contents, etc.), and so
|
|
// are edges (relocation offsets, etc.). Given this formulation, we want to
|
|
// find as many isomorphic subgraphs as possible.
|
|
//
|
|
// Solving such problem is computationally intensive, but mold is quite fast.
|
|
// For Chromium, mold's ICF finishes in less than 1 second with 20 threads.
|
|
// This is contrary to lld and gold, which take about 5 and 50 seconds to
|
|
// run ICF under the same condition, respectively.
|
|
//
|
|
// mold's ICF is faster because we are using a better algorithm.
|
|
// It's actually me who developed and implemented the lld's ICF algorithm,
|
|
// and I can say that mold's algorithm is better than that in all aspects.
|
|
// It scales better for number of available cores, require less overall
|
|
// computation, and has a smaller working set. So, it's better with a single
|
|
// thread and even better with multiple threads.
|
|
|
|
#include "mold.h"
|
|
|
|
#include <array>
|
|
#include <openssl/sha.h>
|
|
#include <tbb/concurrent_unordered_map.h>
|
|
#include <tbb/concurrent_vector.h>
|
|
#include <tbb/enumerable_thread_specific.h>
|
|
#include <tbb/parallel_for.h>
|
|
#include <tbb/parallel_for_each.h>
|
|
#include <tbb/parallel_sort.h>
|
|
|
|
static constexpr i64 HASH_SIZE = 16;
|
|
|
|
typedef std::array<u8, HASH_SIZE> Digest;
|
|
|
|
namespace tbb {
|
|
template<> struct tbb_hash<Digest> {
|
|
size_t operator()(const Digest &k) const {
|
|
return *(i64 *)&k[0];
|
|
}
|
|
};
|
|
}
|
|
|
|
template <typename E>
|
|
static void uniquify_cies(Context<E> &ctx) {
|
|
Timer t(ctx, "uniquify_cies");
|
|
std::vector<CieRecord<E> *> cies;
|
|
|
|
for (ObjectFile<E> *file : ctx.objs) {
|
|
for (CieRecord<E> &cie : file->cies) {
|
|
for (i64 i = 0; i < cies.size(); i++) {
|
|
if (cie.equals(*cies[i])) {
|
|
cie.icf_idx = i;
|
|
goto found;
|
|
}
|
|
}
|
|
cie.icf_idx = cies.size();
|
|
cies.push_back(&cie);
|
|
found:;
|
|
}
|
|
}
|
|
}
|
|
|
|
template <typename E>
|
|
static bool is_eligible(InputSection<E> &isec) {
|
|
const ElfShdr<E> &shdr = isec.shdr;
|
|
std::string_view name = isec.name();
|
|
|
|
bool is_alloc = (shdr.sh_flags & SHF_ALLOC);
|
|
bool is_executable = (shdr.sh_flags & SHF_EXECINSTR);
|
|
bool is_relro = (name == ".data.rel.ro" ||
|
|
name.starts_with(".data.rel.ro."));
|
|
bool is_readonly = !(shdr.sh_flags & SHF_WRITE) || is_relro;
|
|
bool is_bss = (shdr.sh_type == SHT_NOBITS);
|
|
bool is_empty = (shdr.sh_size == 0);
|
|
bool is_init = (shdr.sh_type == SHT_INIT_ARRAY || name == ".init");
|
|
bool is_fini = (shdr.sh_type == SHT_FINI_ARRAY || name == ".fini");
|
|
bool is_enumerable = is_c_identifier(name);
|
|
|
|
return is_alloc && is_executable && is_readonly && !is_bss &&
|
|
!is_empty && !is_init && !is_fini && !is_enumerable;
|
|
}
|
|
|
|
static Digest digest_final(SHA256_CTX &sha) {
|
|
u8 buf[SHA256_SIZE];
|
|
assert(SHA256_Final(buf, &sha) == 1);
|
|
|
|
Digest digest;
|
|
memcpy(digest.data(), buf, HASH_SIZE);
|
|
return digest;
|
|
}
|
|
|
|
template <typename E>
|
|
static bool is_leaf(Context<E> &ctx, InputSection<E> &isec) {
|
|
if (!isec.get_rels(ctx).empty())
|
|
return false;
|
|
|
|
for (FdeRecord<E> &fde : isec.get_fdes())
|
|
if (fde.get_rels().size() > 1)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static size_t combine_hash(size_t a, size_t b) {
|
|
return a ^ (b + 0x9e3779b9 + (a << 6) + (a >> 2));
|
|
}
|
|
|
|
template <typename E>
|
|
struct LeafHasher {
|
|
size_t operator()(const InputSection<E> *isec) const {
|
|
size_t h = hash_string(isec->contents);
|
|
for (FdeRecord<E> &fde : isec->get_fdes()) {
|
|
size_t h2 = hash_string(fde.get_contents().substr(8));
|
|
h = combine_hash(h, h2);
|
|
}
|
|
return h;
|
|
}
|
|
};
|
|
|
|
template <typename E>
|
|
struct LeafEq {
|
|
bool operator()(const InputSection<E> *a, const InputSection<E> *b) const {
|
|
if (a->contents != b->contents)
|
|
return false;
|
|
|
|
std::span<FdeRecord<E>> x = a->get_fdes();
|
|
std::span<FdeRecord<E>> y = b->get_fdes();
|
|
|
|
if (x.size() != y.size())
|
|
return false;
|
|
|
|
for (i64 i = 0; i < x.size(); i++)
|
|
if (x[i].get_contents().substr(8) != y[i].get_contents().substr(8))
|
|
return false;
|
|
return true;
|
|
}
|
|
};
|
|
|
|
template <typename E>
|
|
static void merge_leaf_nodes(Context<E> &ctx) {
|
|
Timer t(ctx, "merge_leaf_nodes");
|
|
|
|
static Counter eligible("icf_eligibles");
|
|
static Counter non_eligible("icf_non_eligibles");
|
|
static Counter leaf("icf_leaf_nodes");
|
|
|
|
tbb::concurrent_unordered_map<InputSection<E> *, InputSection<E> *,
|
|
LeafHasher<E>, LeafEq<E>> map;
|
|
|
|
tbb::parallel_for((i64)0, (i64)ctx.objs.size(), [&](i64 i) {
|
|
for (std::unique_ptr<InputSection<E>> &isec : ctx.objs[i]->sections) {
|
|
if (!isec || !isec->is_alive)
|
|
continue;
|
|
|
|
if (!is_eligible(*isec)) {
|
|
non_eligible++;
|
|
continue;
|
|
}
|
|
|
|
if (is_leaf(ctx, *isec)) {
|
|
leaf++;
|
|
isec->icf_leaf = true;
|
|
auto [it, inserted] = map.insert({isec.get(), isec.get()});
|
|
if (!inserted && isec->get_priority() < it->second->get_priority())
|
|
it->second = isec.get();
|
|
} else {
|
|
eligible++;
|
|
isec->icf_eligible = true;
|
|
}
|
|
}
|
|
});
|
|
|
|
tbb::parallel_for((i64)0, (i64)ctx.objs.size(), [&](i64 i) {
|
|
for (std::unique_ptr<InputSection<E>> &isec : ctx.objs[i]->sections) {
|
|
if (isec && isec->is_alive && isec->icf_leaf) {
|
|
auto it = map.find(isec.get());
|
|
assert(it != map.end());
|
|
isec->leader = it->second;
|
|
}
|
|
}
|
|
});
|
|
}
|
|
|
|
template <typename E>
|
|
static Digest compute_digest(Context<E> &ctx, InputSection<E> &isec) {
|
|
SHA256_CTX sha;
|
|
SHA256_Init(&sha);
|
|
|
|
u8 *buf = (u8 *)isec.contents.data();
|
|
|
|
auto hash = [&](auto val) {
|
|
SHA256_Update(&sha, &val, sizeof(val));
|
|
};
|
|
|
|
auto hash_string = [&](std::string_view str) {
|
|
hash(str.size());
|
|
SHA256_Update(&sha, str.data(), str.size());
|
|
};
|
|
|
|
auto hash_symbol = [&](Symbol<E> &sym) {
|
|
InputSection<E> *isec = sym.input_section;
|
|
|
|
if (!sym.file) {
|
|
hash('1');
|
|
hash((u64)&sym);
|
|
} else if (SectionFragment<E> *frag = sym.get_frag()) {
|
|
hash('2');
|
|
hash_string(frag->data);
|
|
} else if (!isec) {
|
|
hash('3');
|
|
} else if (isec->leader) {
|
|
hash('4');
|
|
hash((u64)isec->leader);
|
|
} else if (isec->icf_eligible) {
|
|
hash('5');
|
|
} else {
|
|
hash('6');
|
|
hash((u64)isec);
|
|
}
|
|
hash(sym.value);
|
|
};
|
|
|
|
hash_string(isec.contents);
|
|
hash(isec.shdr.sh_flags);
|
|
hash(isec.get_fdes().size());
|
|
hash(isec.get_rels(ctx).size());
|
|
|
|
for (FdeRecord<E> &fde : isec.get_fdes()) {
|
|
hash(fde.cie->icf_idx);
|
|
|
|
// Bytes 0 to 4 contain the length of this record, and
|
|
// bytes 4 to 8 contain an offset to CIE.
|
|
hash_string(fde.get_contents().substr(8));
|
|
|
|
hash(fde.get_rels().size());
|
|
|
|
for (ElfRel<E> &rel : fde.get_rels().subspan(1)) {
|
|
hash_symbol(*isec.file.symbols[rel.r_sym]);
|
|
hash(rel.r_type);
|
|
hash(rel.r_offset - fde.input_offset);
|
|
hash(fde.cie->input_section.get_addend(rel));
|
|
}
|
|
}
|
|
|
|
i64 frag_idx = 0;
|
|
|
|
for (i64 i = 0; i < isec.get_rels(ctx).size(); i++) {
|
|
ElfRel<E> &rel = isec.get_rels(ctx)[i];
|
|
hash(rel.r_offset);
|
|
hash(rel.r_type);
|
|
hash(isec.get_addend(rel));
|
|
|
|
if (isec.rel_fragments && isec.rel_fragments[frag_idx].idx == i) {
|
|
SectionFragmentRef<E> &ref = isec.rel_fragments[frag_idx++];
|
|
hash('a');
|
|
isec.get_addend(rel);
|
|
hash_string(ref.frag->data);
|
|
} else {
|
|
hash_symbol(*isec.file.symbols[rel.r_sym]);
|
|
}
|
|
}
|
|
|
|
return digest_final(sha);
|
|
}
|
|
|
|
template <typename E>
|
|
static std::vector<InputSection<E> *> gather_sections(Context<E> &ctx) {
|
|
Timer t(ctx, "gather_sections");
|
|
|
|
// Count the number of input sections for each input file.
|
|
std::vector<i64> num_sections(ctx.objs.size());
|
|
|
|
tbb::parallel_for((i64)0, (i64)ctx.objs.size(), [&](i64 i) {
|
|
for (std::unique_ptr<InputSection<E>> &isec : ctx.objs[i]->sections)
|
|
if (isec && isec->is_alive && isec->icf_eligible)
|
|
num_sections[i]++;
|
|
});
|
|
|
|
std::vector<i64> section_indices(ctx.objs.size());
|
|
for (i64 i = 0; i < ctx.objs.size() - 1; i++)
|
|
section_indices[i + 1] = section_indices[i] + num_sections[i];
|
|
|
|
std::vector<InputSection<E> *> sections(
|
|
section_indices.back() + num_sections.back());
|
|
|
|
// Fill `sections` contents.
|
|
tbb::parallel_for((i64)0, (i64)ctx.objs.size(), [&](i64 i) {
|
|
i64 idx = section_indices[i];
|
|
for (std::unique_ptr<InputSection<E>> &isec : ctx.objs[i]->sections)
|
|
if (isec && isec->is_alive && isec->icf_eligible)
|
|
sections[idx++] = isec.get();
|
|
});
|
|
|
|
tbb::parallel_for((i64)0, (i64)sections.size(), [&](i64 i) {
|
|
sections[i]->icf_idx = i;
|
|
});
|
|
|
|
return sections;
|
|
}
|
|
|
|
template <typename E>
|
|
static std::vector<Digest>
|
|
compute_digests(Context<E> &ctx, std::span<InputSection<E> *> sections) {
|
|
Timer t(ctx, "compute_digests");
|
|
|
|
std::vector<Digest> digests(sections.size());
|
|
tbb::parallel_for((i64)0, (i64)sections.size(), [&](i64 i) {
|
|
digests[i] = compute_digest(ctx, *sections[i]);
|
|
});
|
|
return digests;
|
|
}
|
|
|
|
template <typename E>
|
|
static void gather_edges(Context<E> &ctx,
|
|
std::span<InputSection<E> *> sections,
|
|
std::vector<u32> &edges,
|
|
std::vector<u32> &edge_indices) {
|
|
Timer t(ctx, "gather_edges");
|
|
|
|
std::vector<i64> num_edges(sections.size());
|
|
edge_indices.resize(sections.size());
|
|
|
|
tbb::parallel_for((i64)0, (i64)sections.size(), [&](i64 i) {
|
|
InputSection<E> &isec = *sections[i];
|
|
assert(isec.icf_eligible);
|
|
i64 frag_idx = 0;
|
|
|
|
for (i64 j = 0; j < isec.get_rels(ctx).size(); j++) {
|
|
if (isec.rel_fragments && isec.rel_fragments[frag_idx].idx == j) {
|
|
frag_idx++;
|
|
} else {
|
|
ElfRel<E> &rel = isec.get_rels(ctx)[j];
|
|
Symbol<E> &sym = *isec.file.symbols[rel.r_sym];
|
|
if (!sym.get_frag() && sym.input_section && sym.input_section->icf_eligible)
|
|
num_edges[i]++;
|
|
}
|
|
}
|
|
});
|
|
|
|
for (i64 i = 0; i < num_edges.size() - 1; i++)
|
|
edge_indices[i + 1] = edge_indices[i] + num_edges[i];
|
|
|
|
edges.resize(edge_indices.back() + num_edges.back());
|
|
|
|
tbb::parallel_for((i64)0, (i64)num_edges.size(), [&](i64 i) {
|
|
InputSection<E> &isec = *sections[i];
|
|
i64 frag_idx = 0;
|
|
i64 idx = edge_indices[i];
|
|
|
|
for (i64 j = 0; j < isec.get_rels(ctx).size(); j++) {
|
|
if (isec.rel_fragments && isec.rel_fragments[frag_idx].idx == j) {
|
|
frag_idx++;
|
|
ElfRel<E> &rel = isec.get_rels(ctx)[j];
|
|
Symbol<E> &sym = *isec.file.symbols[rel.r_sym];
|
|
if (!sym.get_frag() && sym.input_section && sym.input_section->icf_eligible)
|
|
edges[idx++] = sym.input_section->icf_idx;
|
|
}
|
|
}
|
|
});
|
|
}
|
|
|
|
template <typename E>
|
|
static i64 propagate(std::span<std::vector<Digest>> digests,
|
|
std::span<u32> edges, std::span<u32> edge_indices,
|
|
bool &slot, tbb::affinity_partitioner &ap) {
|
|
static Counter round("icf_round");
|
|
round++;
|
|
|
|
i64 num_digests = digests[0].size();
|
|
tbb::enumerable_thread_specific<i64> changed;
|
|
|
|
tbb::parallel_for((i64)0, num_digests, [&](i64 i) {
|
|
if (digests[slot][i] == digests[!slot][i])
|
|
return;
|
|
|
|
SHA256_CTX sha;
|
|
SHA256_Init(&sha);
|
|
SHA256_Update(&sha, digests[2][i].data(), HASH_SIZE);
|
|
|
|
i64 begin = edge_indices[i];
|
|
i64 end = (i + 1 == num_digests) ? edges.size() : edge_indices[i + 1];
|
|
|
|
for (i64 j = begin; j < end; j++)
|
|
SHA256_Update(&sha, digests[slot][edges[j]].data(), HASH_SIZE);
|
|
|
|
digests[!slot][i] = digest_final(sha);
|
|
|
|
if (digests[slot][i] != digests[!slot][i])
|
|
changed.local()++;
|
|
}, ap);
|
|
|
|
slot = !slot;
|
|
return changed.combine(std::plus());
|
|
}
|
|
|
|
template <typename E>
|
|
static i64 count_num_classes(std::span<Digest> digests,
|
|
tbb::affinity_partitioner &ap) {
|
|
std::vector<Digest> vec(digests.begin(), digests.end());
|
|
tbb::parallel_sort(vec);
|
|
|
|
tbb::enumerable_thread_specific<i64> num_classes;
|
|
tbb::parallel_for((i64)0, (i64)vec.size() - 1, [&](i64 i) {
|
|
if (vec[i] != vec[i + 1])
|
|
num_classes.local()++;
|
|
}, ap);
|
|
return num_classes.combine(std::plus());
|
|
}
|
|
|
|
template <typename E>
|
|
static void print_icf_sections(Context<E> &ctx) {
|
|
tbb::concurrent_vector<InputSection<E> *> leaders;
|
|
tbb::concurrent_unordered_multimap<InputSection<E> *, InputSection<E> *> map;
|
|
|
|
tbb::parallel_for_each(ctx.objs, [&](ObjectFile<E> *file) {
|
|
for (std::unique_ptr<InputSection<E>> &isec : file->sections) {
|
|
if (isec && isec->is_alive && isec->leader) {
|
|
if (isec.get() == isec->leader)
|
|
leaders.push_back(isec.get());
|
|
else
|
|
map.insert({isec->leader, isec.get()});
|
|
}
|
|
}
|
|
});
|
|
|
|
tbb::parallel_sort(leaders.begin(), leaders.end(),
|
|
[&](InputSection<E> *a, InputSection<E> *b) {
|
|
return a->get_priority() < b->get_priority();
|
|
});
|
|
|
|
i64 saved_bytes = 0;
|
|
|
|
for (InputSection<E> *leader : leaders) {
|
|
auto [begin, end] = map.equal_range(leader);
|
|
if (begin == end)
|
|
continue;
|
|
|
|
SyncOut(ctx) << "selected section " << *leader;
|
|
|
|
i64 n = 0;
|
|
for (auto it = begin; it != end; it++) {
|
|
SyncOut(ctx) << " removing identical section " << *it->second;
|
|
n++;
|
|
}
|
|
saved_bytes += leader->contents.size() * n;
|
|
}
|
|
|
|
SyncOut(ctx) << "ICF saved " << saved_bytes << " bytes";
|
|
}
|
|
|
|
template <typename E>
|
|
void icf_sections(Context<E> &ctx) {
|
|
Timer t(ctx, "icf");
|
|
|
|
uniquify_cies(ctx);
|
|
merge_leaf_nodes(ctx);
|
|
|
|
// Prepare for the propagation rounds.
|
|
std::vector<InputSection<E> *> sections = gather_sections(ctx);
|
|
|
|
std::vector<std::vector<Digest>> digests(3);
|
|
digests[0] = compute_digests<E>(ctx, sections);
|
|
digests[1].resize(digests[0].size());
|
|
digests[2] = digests[0];
|
|
|
|
std::vector<u32> edges;
|
|
std::vector<u32> edge_indices;
|
|
gather_edges<E>(ctx, sections, edges, edge_indices);
|
|
|
|
bool slot = 0;
|
|
|
|
// Execute the propagation rounds until convergence is obtained.
|
|
{
|
|
Timer t(ctx, "propagate");
|
|
tbb::affinity_partitioner ap;
|
|
|
|
i64 num_changed = -1;
|
|
for (;;) {
|
|
i64 n = propagate<E>(digests, edges, edge_indices, slot, ap);
|
|
if (n == num_changed)
|
|
break;
|
|
num_changed = n;
|
|
}
|
|
|
|
i64 num_classes = -1;
|
|
for (;;) {
|
|
for (i64 i = 0; i < 10; i++)
|
|
propagate<E>(digests, edges, edge_indices, slot, ap);
|
|
|
|
i64 n = count_num_classes<E>(digests[slot], ap);
|
|
if (n == num_classes)
|
|
break;
|
|
num_classes = n;
|
|
}
|
|
}
|
|
|
|
// Group sections by SHA digest.
|
|
{
|
|
Timer t(ctx, "group");
|
|
|
|
auto *map = new tbb::concurrent_unordered_map<Digest, InputSection<E> *>;
|
|
std::span<Digest> digest = digests[slot];
|
|
|
|
tbb::parallel_for((i64)0, (i64)sections.size(), [&](i64 i) {
|
|
InputSection<E> *isec = sections[i];
|
|
auto [it, inserted] = map->insert({digest[i], isec});
|
|
if (!inserted && isec->get_priority() < it->second->get_priority())
|
|
it->second = isec;
|
|
});
|
|
|
|
tbb::parallel_for((i64)0, (i64)sections.size(), [&](i64 i) {
|
|
auto it = map->find(digest[i]);
|
|
assert(it != map->end());
|
|
sections[i]->leader = it->second;
|
|
});
|
|
|
|
// Since free'ing the map is slow, postpone it.
|
|
ctx.on_exit.push_back([=]() { delete map; });
|
|
}
|
|
|
|
if (ctx.arg.print_icf_sections)
|
|
print_icf_sections(ctx);
|
|
|
|
// Re-assign input sections to symbols.
|
|
{
|
|
Timer t(ctx, "reassign");
|
|
tbb::parallel_for_each(ctx.objs, [](ObjectFile<E> *file) {
|
|
for (Symbol<E> *sym : file->symbols) {
|
|
if (sym->file != file)
|
|
continue;
|
|
InputSection<E> *isec = sym->input_section;
|
|
if (isec && isec->leader && isec->leader != isec) {
|
|
sym->input_section = isec->leader;
|
|
isec->kill();
|
|
}
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
template void icf_sections(Context<X86_64> &ctx);
|
|
template void icf_sections(Context<I386> &ctx);
|