[v3] elf: Properly align PT_LOAD segments [BZ #28676]
Checks
Context |
Check |
Description |
dj/TryBot-apply_patch |
success
|
Patch applied to master at the time it was sent
|
dj/TryBot-32bit |
success
|
Build for i686
|
Commit Message
On Thu, Dec 09, 2021 at 07:14:10AM -0800, H.J. Lu wrote:
> On Wed, Dec 8, 2021 at 9:57 PM Rongwei Wang
> <rongwei.wang@linux.alibaba.com> wrote:
> >
> > Now, ld.so always map the LOAD segments and aligned by base
> > page size (e.g. 4k in x86 or 4k, 16k and 64k in arm64). This
> > is a bug, and had been reported:
> >
> > https://sourceware.org/bugzilla/show_bug.cgi?id=28676
> >
> > This patch mainly to fix it. In this patch, ld.so can align
> > the mapping address of the first LOAD segment with p_align
> > when p_align is greater than the current base page size.
> >
> > A testcase:
> > main.c:
> >
> > extern void dso_test(void);
> > int main(void)
> > {
> > dso_test();
> > getchar();
> >
> > return 0;
> > }
> >
> > load.c, used to generate libload.so:
> >
> > int foo __attribute__((aligned(0x200000))) = 1;
> > void dso_test(void)
> > {
> > printf("dso test\n");
> > printf("foo: %p\n", &foo);
> > }
> >
> > The steps:
> > $ gcc -O2 -fPIC -c -o load.o load.c
> > $ gcc -shared -Wl,-z,max-page-size=0x200000 -o libload.so load.o
> > $ gcc -no-pie -Wl,-z,max-page-size=0x200000 -O2 -o dso main.c libload.so -Wl,-R,.
> >
> > Before fixing:
> > $ ./dso
> > dso test
> > foo: 0xffff88ae2000
> >
> > After fixed:
> > $ ./dso
> > dso test
> > foo: 0xffff9e000000
> >
> > And this fix can help code segments use huge pages become
> > simple and available.
>
> Please include a testcase, like
>
> https://gitlab.com/x86-glibc/glibc/-/commits/users/hjl/pr28676/master
>
> > Signed-off-by: Xu Yu <xuyu@linux.alibaba.com>
> > Signed-off-by: Rongwei Wang <rongwei.wang@linux.alibaba.com>
> > ---
> > elf/dl-load.c | 1 +
> > elf/dl-map-segments.h | 63 +++++++++++++++++++++++++++++++++++++++----
> > include/link.h | 3 +++
> > 3 files changed, 62 insertions(+), 5 deletions(-)
> >
> > diff --git a/elf/dl-load.c b/elf/dl-load.c
> > index e39980fb19..136cfe2fa8 100644
> > --- a/elf/dl-load.c
> > +++ b/elf/dl-load.c
> > @@ -1154,6 +1154,7 @@ _dl_map_object_from_fd (const char *name, const char *origname, int fd,
> > c->dataend = ph->p_vaddr + ph->p_filesz;
> > c->allocend = ph->p_vaddr + ph->p_memsz;
> > c->mapoff = ALIGN_DOWN (ph->p_offset, GLRO(dl_pagesize));
> > + l->l_load_align = ph->p_align;
>
> Can you add an alignment field to
>
> /* This structure describes one PT_LOAD command.
> Its details have been expanded out and converted. */
> struct loadcmd
> {
> ElfW(Addr) mapstart, mapend, dataend, allocend;
> ElfW(Off) mapoff;
> int prot; /* PROT_* bits. */
> };
>
> instead?
>
Hi,
I updated your patch. Please take a look.
H.J.
--
When PT_LOAD segment alignment > the page size, allocate enough space
to ensure that the segment can be properly aligned.
This fixes [BZ #28676].
---
elf/dl-load.c | 1 +
elf/dl-load.h | 2 +-
elf/dl-map-segments.h | 51 +++++++++++++++++++++++++++++++++++++++----
3 files changed, 49 insertions(+), 5 deletions(-)
@@ -1150,6 +1150,7 @@ _dl_map_object_from_fd (const char *name, const char *origname, int fd,
c->mapend = ALIGN_UP (ph->p_vaddr + ph->p_filesz, GLRO(dl_pagesize));
c->dataend = ph->p_vaddr + ph->p_filesz;
c->allocend = ph->p_vaddr + ph->p_memsz;
+ c->mapalign = ph->p_align;
c->mapoff = ALIGN_DOWN (ph->p_offset, GLRO(dl_pagesize));
/* Determine whether there is a gap between the last segment
@@ -74,7 +74,7 @@ ELF_PREFERRED_ADDRESS_DATA;
Its details have been expanded out and converted. */
struct loadcmd
{
- ElfW(Addr) mapstart, mapend, dataend, allocend;
+ ElfW(Addr) mapstart, mapend, dataend, allocend, mapalign;
ElfW(Off) mapoff;
int prot; /* PROT_* bits. */
};
@@ -18,6 +18,52 @@
#include <dl-load.h>
+/* Map a segment and align it properly. */
+
+static __always_inline ElfW(Addr)
+_dl_map_segment (const struct loadcmd *c, ElfW(Addr) mappref,
+ const size_t maplength, int fd)
+{
+ if (c->mapalign > GLRO(dl_pagesize))
+ {
+ /* If the segment alignment > the page size, allocate enough space
+ to ensure that the segment can be properly aligned. */
+ ElfW(Addr) maplen = (maplength >= c->mapalign
+ ? (maplength + c->mapalign)
+ : (2 * c->mapalign));
+ ElfW(Addr) map_start
+ = (ElfW(Addr)) __mmap ((void *) mappref, maplen,
+ PROT_NONE, MAP_ANONYMOUS|MAP_PRIVATE,
+ -1, 0);
+ if (__glibc_unlikely ((void *) map_start == MAP_FAILED))
+ return map_start;
+
+ ElfW(Addr) map_start_aligned = ALIGN_UP (map_start, c->mapalign);
+ ElfW(Addr) map_end = map_start_aligned + maplength;
+ map_start_aligned
+ = (ElfW(Addr)) __mmap ((void *) map_start_aligned,
+ maplength, c->prot,
+ MAP_COPY|MAP_FILE|MAP_FIXED,
+ fd, c->mapoff);
+ if (__glibc_likely ((void *) map_start_aligned != MAP_FAILED))
+ {
+ /* Unmap the unused regions. */
+ ElfW(Addr) delta = map_start_aligned - map_start;
+ if (delta)
+ __munmap ((void *) map_start, delta);
+ delta = map_start + maplen - map_end;
+ if (delta)
+ __munmap ((void *) map_end, delta);
+ }
+
+ return map_start_aligned;
+ }
+
+ return (ElfW(Addr)) __mmap ((void *) mappref, maplength,
+ c->prot, MAP_COPY|MAP_FILE,
+ fd, c->mapoff);
+}
+
/* This implementation assumes (as does the corresponding implementation
of _dl_unmap_segments, in dl-unmap-segments.h) that shared objects
are always laid out with all segments contiguous (or with gaps
@@ -53,10 +99,7 @@ _dl_map_segments (struct link_map *l, int fd,
- MAP_BASE_ADDR (l));
/* Remember which part of the address space this object uses. */
- l->l_map_start = (ElfW(Addr)) __mmap ((void *) mappref, maplength,
- c->prot,
- MAP_COPY|MAP_FILE,
- fd, c->mapoff);
+ l->l_map_start = _dl_map_segment (c, mappref, maplength, fd);
if (__glibc_unlikely ((void *) l->l_map_start == MAP_FAILED))
return DL_MAP_SEGMENTS_ERROR_MAP_SEGMENT;