1
1
mirror of https://github.com/rui314/mold.git synced 2024-09-19 08:57:39 +03:00
This commit is contained in:
Rui Ueyama 2024-04-26 15:50:46 +09:00
parent 53ebcd80d8
commit 31969a2756
4 changed files with 17 additions and 22 deletions

View File

@ -524,6 +524,14 @@ inline bool remove_prefix(std::string_view &s, std::string_view prefix) {
// Concurrent Map
//
static inline void pause() {
#if defined(__x86_64__)
asm volatile("pause");
#elif defined(__arm__) || defined(__aarch64__)
asm volatile("yield");
#endif
}
// This is an implementation of a fast concurrent hash map. Unlike
// ordinary hash tables, this impl just aborts if it becomes full.
// So you need to give a correct estimation of the final size before
@ -576,7 +584,7 @@ public:
#endif
}
std::pair<T *, bool> insert(std::string_view key, u64 hash, const T &val) {
std::pair<T *, bool> insert(std::string_view key, u32 hash, const T &val) {
assert(has_single_bit(nbuckets));
i64 begin = hash & (nbuckets - 1);
@ -586,8 +594,8 @@ public:
i64 idx = (begin & ~mask) | ((begin + i) & mask);
Entry &ent = entries[idx];
// It seems avoiding compare-and-exchange is faster overall at
// least on my Zen4 machine, so do it.
// It seems avoiding compare-and-swap is faster overall at least
// on my Zen4 machine, so do it.
if (const char *ptr = ent.key.load(std::memory_order_acquire);
ptr != nullptr && ptr != (char *)-1) {
if (key == std::string_view(ptr, ent.keylen))
@ -686,15 +694,6 @@ public:
Entry *entries = nullptr;
i64 nbuckets = 0;
private:
static void pause() {
#if defined(__x86_64__)
asm volatile("pause");
#elif defined(__aarch64__)
asm volatile("yield");
#endif
}
};
//

View File

@ -699,7 +699,7 @@ static size_t find_null(std::string_view data, i64 pos, i64 entsize) {
template <typename E>
static std::unique_ptr<MergeableSection<E>>
split_section(Context<E> &ctx, InputSection<E> &sec) {
if (!sec.is_alive || sec.relsec_idx != -1)
if (!sec.is_alive || sec.relsec_idx != -1 || sec.sh_size == 0)
return nullptr;
const ElfShdr<E> &shdr = sec.shdr();
@ -719,12 +719,9 @@ split_section(Context<E> &ctx, InputSection<E> &sec) {
std::unique_ptr<MergeableSection<E>> m(new MergeableSection<E>);
m->parent = MergedSection<E>::get_instance(ctx, sec.name(), shdr.sh_type,
shdr.sh_flags, entsize, addralign);
shdr.sh_flags, entsize, addralign);
m->p2align = sec.p2align;
if (sec.sh_size == 0)
return m;
// If thes section contents are compressed, uncompress them.
sec.uncompress(ctx);

View File

@ -103,12 +103,12 @@ void OutputShdr<E>::copy_buf(Context<E> &ctx) {
ElfShdr<E> *hdr = (ElfShdr<E> *)(ctx.buf + this->shdr.sh_offset);
memset(hdr, 0, this->shdr.sh_size);
if (ctx.shstrtab && SHN_LORESERVE <= ctx.shstrtab->shndx)
hdr[0].sh_link = ctx.shstrtab->shndx;
i64 shnum = ctx.shdr->shdr.sh_size / sizeof(ElfShdr<E>);
if (UINT16_MAX < shnum)
hdr->sh_size = shnum;
if (ctx.shstrtab && SHN_LORESERVE <= ctx.shstrtab->shndx)
hdr->sh_link = ctx.shstrtab->shndx;
hdr[0].sh_size = shnum;
for (Chunk<E> *chunk : ctx.chunks)
if (chunk->shndx)

View File

@ -2955,7 +2955,6 @@ void fix_synthetic_symbols(Context<E> &ctx) {
}
}
// --section-order symbols
for (SectionOrder &ord : ctx.arg.section_order)
if (ord.type == SectionOrder::SYMBOL)