2007-07-21 22:10:01 +07:00
|
|
|
#
|
2008-01-30 19:30:42 +07:00
|
|
|
# Building vDSO images for x86.
|
2007-07-21 22:10:01 +07:00
|
|
|
#
|
|
|
|
|
2014-02-08 15:01:05 +07:00
|
|
|
KBUILD_CFLAGS += $(DISABLE_LTO)
|
2015-02-14 05:39:25 +07:00
|
|
|
KASAN_SANITIZE := n
|
2014-02-08 15:01:05 +07:00
|
|
|
|
2008-01-30 19:30:42 +07:00
|
|
|
VDSO64-$(CONFIG_X86_64) := y
|
2012-02-20 02:38:06 +07:00
|
|
|
VDSOX32-$(CONFIG_X86_X32_ABI) := y
|
2008-01-30 19:30:42 +07:00
|
|
|
VDSO32-$(CONFIG_X86_32) := y
|
|
|
|
VDSO32-$(CONFIG_COMPAT) := y
|
|
|
|
|
2007-07-21 22:10:01 +07:00
|
|
|
# files to link into the vdso
|
2014-07-11 08:13:16 +07:00
|
|
|
vobjs-y := vdso-note.o vclock_gettime.o vgetcpu.o
|
2012-02-20 02:38:06 +07:00
|
|
|
|
2007-07-21 22:10:01 +07:00
|
|
|
# files to link into kernel
|
2014-03-18 05:22:08 +07:00
|
|
|
obj-y += vma.o
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
|
|
|
|
# vDSO images to build
|
|
|
|
vdso_img-$(VDSO64-y) += 64
|
|
|
|
vdso_img-$(VDSOX32-y) += x32
|
|
|
|
vdso_img-$(VDSO32-y) += 32-int80
|
|
|
|
vdso_img-$(CONFIG_COMPAT) += 32-syscall
|
|
|
|
vdso_img-$(VDSO32-y) += 32-sysenter
|
|
|
|
|
|
|
|
obj-$(VDSO32-y) += vdso32-setup.o
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2014-06-13 07:53:12 +07:00
|
|
|
vobjs := $(foreach F,$(vobjs-y),$(obj)/$F)
|
2007-07-21 22:10:01 +07:00
|
|
|
|
|
|
|
$(obj)/vdso.o: $(obj)/vdso.so
|
|
|
|
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
targets += vdso.lds $(vobjs-y)
|
|
|
|
|
|
|
|
# Build the vDSO image C files and link them in.
|
|
|
|
vdso_img_objs := $(vdso_img-y:%=vdso-image-%.o)
|
|
|
|
vdso_img_cfiles := $(vdso_img-y:%=vdso-image-%.c)
|
|
|
|
vdso_img_sodbg := $(vdso_img-y:%=vdso%.so.dbg)
|
|
|
|
obj-y += $(vdso_img_objs)
|
|
|
|
targets += $(vdso_img_cfiles)
|
|
|
|
targets += $(vdso_img_sodbg)
|
2014-07-11 08:13:16 +07:00
|
|
|
.SECONDARY: $(vdso_img-y:%=$(obj)/vdso-image-%.c) \
|
|
|
|
$(vdso_img-y:%=$(obj)/vdso%.so)
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2007-11-13 02:14:19 +07:00
|
|
|
export CPPFLAGS_vdso.lds += -P -C
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2010-12-14 07:01:38 +07:00
|
|
|
VDSO_LDFLAGS_vdso.lds = -m64 -Wl,-soname=linux-vdso.so.1 \
|
2011-07-13 20:24:12 +07:00
|
|
|
-Wl,--no-undefined \
|
2014-02-08 15:01:05 +07:00
|
|
|
-Wl,-z,max-page-size=4096 -Wl,-z,common-page-size=4096 \
|
|
|
|
$(DISABLE_LTO)
|
2007-07-21 22:10:01 +07:00
|
|
|
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
$(obj)/vdso64.so.dbg: $(src)/vdso.lds $(vobjs) FORCE
|
2008-01-30 19:30:44 +07:00
|
|
|
$(call if_changed,vdso)
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2015-05-11 15:15:50 +07:00
|
|
|
HOST_EXTRACFLAGS += -I$(srctree)/tools/include -I$(srctree)/include/uapi -I$(srctree)/arch/x86/include/uapi
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
hostprogs-y += vdso2c
|
|
|
|
|
|
|
|
quiet_cmd_vdso2c = VDSO2C $@
|
|
|
|
define cmd_vdso2c
|
2014-07-11 08:13:16 +07:00
|
|
|
$(obj)/vdso2c $< $(<:%.dbg=%) $@
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
endef
|
|
|
|
|
2014-07-11 08:13:16 +07:00
|
|
|
$(obj)/vdso-image-%.c: $(obj)/vdso%.so.dbg $(obj)/vdso%.so $(obj)/vdso2c FORCE
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
$(call if_changed,vdso2c)
|
2007-10-17 23:04:32 +07:00
|
|
|
|
2011-05-23 20:31:29 +07:00
|
|
|
#
|
|
|
|
# Don't omit frame pointers for ease of userspace debugging, but do
|
|
|
|
# optimize sibling calls.
|
|
|
|
#
|
2008-04-15 02:19:30 +07:00
|
|
|
CFL := $(PROFILING) -mcmodel=small -fPIC -O2 -fasynchronous-unwind-tables -m64 \
|
2011-05-23 20:31:29 +07:00
|
|
|
$(filter -g%,$(KBUILD_CFLAGS)) $(call cc-option, -fno-stack-protector) \
|
2014-06-25 03:46:52 +07:00
|
|
|
-fno-omit-frame-pointer -foptimize-sibling-calls \
|
|
|
|
-DDISABLE_BRANCH_PROFILING
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2008-03-20 00:25:53 +07:00
|
|
|
$(vobjs): KBUILD_CFLAGS += $(CFL)
|
2007-07-21 22:10:01 +07:00
|
|
|
|
2011-05-23 20:31:29 +07:00
|
|
|
#
|
|
|
|
# vDSO code runs in userspace and -pg doesn't help with profiling anyway.
|
|
|
|
#
|
|
|
|
CFLAGS_REMOVE_vdso-note.o = -pg
|
|
|
|
CFLAGS_REMOVE_vclock_gettime.o = -pg
|
|
|
|
CFLAGS_REMOVE_vgetcpu.o = -pg
|
|
|
|
CFLAGS_REMOVE_vvar.o = -pg
|
|
|
|
|
2012-02-20 02:38:06 +07:00
|
|
|
#
|
|
|
|
# X32 processes use x32 vDSO to access 64bit kernel data.
|
|
|
|
#
|
|
|
|
# Build x32 vDSO image:
|
|
|
|
# 1. Compile x32 vDSO as 64bit.
|
|
|
|
# 2. Convert object files to x32.
|
|
|
|
# 3. Build x32 VDSO image with x32 objects, which contains 64bit codes
|
|
|
|
# so that it can reach 64bit address space with 64bit pointers.
|
|
|
|
#
|
|
|
|
|
|
|
|
CPPFLAGS_vdsox32.lds = $(CPPFLAGS_vdso.lds)
|
|
|
|
VDSO_LDFLAGS_vdsox32.lds = -Wl,-m,elf32_x86_64 \
|
|
|
|
-Wl,-soname=linux-vdso.so.1 \
|
|
|
|
-Wl,-z,max-page-size=4096 \
|
|
|
|
-Wl,-z,common-page-size=4096
|
|
|
|
|
2014-06-13 07:53:12 +07:00
|
|
|
# 64-bit objects to re-brand as x32
|
|
|
|
vobjs64-for-x32 := $(filter-out $(vobjs-nox32),$(vobjs-y))
|
|
|
|
|
|
|
|
# x32-rebranded versions
|
|
|
|
vobjx32s-y := $(vobjs64-for-x32:.o=-x32.o)
|
|
|
|
|
|
|
|
# same thing, but in the output directory
|
2012-02-20 02:38:06 +07:00
|
|
|
vobjx32s := $(foreach F,$(vobjx32s-y),$(obj)/$F)
|
|
|
|
|
|
|
|
# Convert 64bit object file to x32 for x32 vDSO.
|
|
|
|
quiet_cmd_x32 = X32 $@
|
|
|
|
cmd_x32 = $(OBJCOPY) -O elf32-x86-64 $< $@
|
|
|
|
|
|
|
|
$(obj)/%-x32.o: $(obj)/%.o FORCE
|
|
|
|
$(call if_changed,x32)
|
|
|
|
|
x86, vdso: Reimplement vdso.so preparation in build-time C
Currently, vdso.so files are prepared and analyzed by a combination
of objcopy, nm, some linker script tricks, and some simple ELF
parsers in the kernel. Replace all of that with plain C code that
runs at build time.
All five vdso images now generate .c files that are compiled and
linked in to the kernel image.
This should cause only one userspace-visible change: the loaded vDSO
images are stripped more heavily than they used to be. Everything
outside the loadable segment is dropped. In particular, this causes
the section table and section name strings to be missing. This
should be fine: real dynamic loaders don't load or inspect these
tables anyway. The result is roughly equivalent to eu-strip's
--strip-sections option.
The purpose of this change is to enable the vvar and hpet mappings
to be moved to the page following the vDSO load segment. Currently,
it is possible for the section table to extend into the page after
the load segment, so, if we map it, it risks overlapping the vvar or
hpet page. This happens whenever the load segment is just under a
multiple of PAGE_SIZE.
The only real subtlety here is that the old code had a C file with
inline assembler that did 'call VDSO32_vsyscall' and a linker script
that defined 'VDSO32_vsyscall = __kernel_vsyscall'. This most
likely worked by accident: the linker script entry defines a symbol
associated with an address as opposed to an alias for the real
dynamic symbol __kernel_vsyscall. That caused ld to relocate the
reference at link time instead of leaving an interposable dynamic
relocation. Since the VDSO32_vsyscall hack is no longer needed, I
now use 'call __kernel_vsyscall', and I added -Bsymbolic to make it
work. vdso2c will generate an error and abort the build if the
resulting image contains any dynamic relocations, so we won't
silently generate bad vdso images.
(Dynamic relocations are a problem because nothing will even attempt
to relocate the vdso.)
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Link: http://lkml.kernel.org/r/2c4fcf45524162a34d87fdda1eb046b2a5cecee7.1399317206.git.luto@amacapital.net
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2014-05-06 02:19:34 +07:00
|
|
|
targets += vdsox32.lds $(vobjx32s-y)
|
2012-02-20 02:38:06 +07:00
|
|
|
|
2014-07-11 08:13:16 +07:00
|
|
|
$(obj)/%.so: OBJCOPYFLAGS := -S
|
|
|
|
$(obj)/%.so: $(obj)/%.so.dbg
|
|
|
|
$(call if_changed,objcopy)
|
|
|
|
|
2012-02-20 02:38:06 +07:00
|
|
|
$(obj)/vdsox32.so.dbg: $(src)/vdsox32.lds $(vobjx32s) FORCE
|
|
|
|
$(call if_changed,vdso)
|
|
|
|
|
2008-01-30 19:30:42 +07:00
|
|
|
#
|
|
|
|
# Build multiple 32-bit vDSO images to choose from at boot time.
|
|
|
|
#
|
2008-07-12 16:22:00 +07:00
|
|
|
vdso32.so-$(VDSO32-y) += int80
|
2008-01-30 19:30:43 +07:00
|
|
|
vdso32.so-$(CONFIG_COMPAT) += syscall
|
2008-01-30 19:30:42 +07:00
|
|
|
vdso32.so-$(VDSO32-y) += sysenter
|
|
|
|
|
2008-02-12 05:38:51 +07:00
|
|
|
vdso32-images = $(vdso32.so-y:%=vdso32-%.so)
|
|
|
|
|
2008-01-30 19:30:42 +07:00
|
|
|
CPPFLAGS_vdso32.lds = $(CPPFLAGS_vdso.lds)
|
2014-02-08 15:01:05 +07:00
|
|
|
VDSO_LDFLAGS_vdso32.lds = -m32 -Wl,-m,elf_i386 -Wl,-soname=linux-gate.so.1
|
2008-01-30 19:30:42 +07:00
|
|
|
|
|
|
|
# This makes sure the $(obj) subdirectory exists even though vdso32/
|
|
|
|
# is not a kbuild sub-make subdirectory.
|
|
|
|
override obj-dirs = $(dir $(obj)) $(obj)/vdso32/
|
|
|
|
|
|
|
|
targets += vdso32/vdso32.lds
|
2014-03-26 06:25:53 +07:00
|
|
|
targets += vdso32/note.o vdso32/vclock_gettime.o $(vdso32.so-y:%=vdso32/%.o)
|
2014-07-11 08:13:16 +07:00
|
|
|
targets += vdso32/vclock_gettime.o
|
2008-01-30 19:30:42 +07:00
|
|
|
|
2008-02-12 05:38:51 +07:00
|
|
|
$(obj)/vdso32.o: $(vdso32-images:%=$(obj)/%)
|
2008-01-30 19:30:42 +07:00
|
|
|
|
2008-01-30 19:30:42 +07:00
|
|
|
KBUILD_AFLAGS_32 := $(filter-out -m64,$(KBUILD_AFLAGS))
|
2008-02-12 05:38:51 +07:00
|
|
|
$(vdso32-images:%=$(obj)/%.dbg): KBUILD_AFLAGS = $(KBUILD_AFLAGS_32)
|
|
|
|
$(vdso32-images:%=$(obj)/%.dbg): asflags-$(CONFIG_X86_64) += -m32
|
2008-01-30 19:30:42 +07:00
|
|
|
|
2014-03-18 05:22:09 +07:00
|
|
|
KBUILD_CFLAGS_32 := $(filter-out -m64,$(KBUILD_CFLAGS))
|
|
|
|
KBUILD_CFLAGS_32 := $(filter-out -mcmodel=kernel,$(KBUILD_CFLAGS_32))
|
|
|
|
KBUILD_CFLAGS_32 := $(filter-out -fno-pic,$(KBUILD_CFLAGS_32))
|
|
|
|
KBUILD_CFLAGS_32 := $(filter-out -mfentry,$(KBUILD_CFLAGS_32))
|
|
|
|
KBUILD_CFLAGS_32 += -m32 -msoft-float -mregparm=0 -fpic
|
2014-03-18 05:22:12 +07:00
|
|
|
KBUILD_CFLAGS_32 += $(call cc-option, -fno-stack-protector)
|
|
|
|
KBUILD_CFLAGS_32 += $(call cc-option, -foptimize-sibling-calls)
|
|
|
|
KBUILD_CFLAGS_32 += -fno-omit-frame-pointer
|
2014-06-25 03:46:52 +07:00
|
|
|
KBUILD_CFLAGS_32 += -DDISABLE_BRANCH_PROFILING
|
2014-03-18 05:22:09 +07:00
|
|
|
$(vdso32-images:%=$(obj)/%.dbg): KBUILD_CFLAGS = $(KBUILD_CFLAGS_32)
|
|
|
|
|
2008-02-12 05:38:51 +07:00
|
|
|
$(vdso32-images:%=$(obj)/%.dbg): $(obj)/vdso32-%.so.dbg: FORCE \
|
|
|
|
$(obj)/vdso32/vdso32.lds \
|
2014-03-18 05:22:09 +07:00
|
|
|
$(obj)/vdso32/vclock_gettime.o \
|
2008-02-12 05:38:51 +07:00
|
|
|
$(obj)/vdso32/note.o \
|
|
|
|
$(obj)/vdso32/%.o
|
2008-01-30 19:30:42 +07:00
|
|
|
$(call if_changed,vdso)
|
|
|
|
|
|
|
|
#
|
|
|
|
# The DSO images are built using a special linker script.
|
|
|
|
#
|
|
|
|
quiet_cmd_vdso = VDSO $@
|
|
|
|
cmd_vdso = $(CC) -nostdlib -o $@ \
|
|
|
|
$(VDSO_LDFLAGS) $(VDSO_LDFLAGS_$(filter %.lds,$(^F))) \
|
2010-06-19 04:36:26 +07:00
|
|
|
-Wl,-T,$(filter %.lds,$^) $(filter %.o,$^) && \
|
|
|
|
sh $(srctree)/$(src)/checkundef.sh '$(NM)' '$@'
|
2008-01-30 19:30:42 +07:00
|
|
|
|
2014-02-08 15:01:05 +07:00
|
|
|
VDSO_LDFLAGS = -fPIC -shared $(call cc-ldoption, -Wl$(comma)--hash-style=sysv) \
|
2014-06-21 02:20:44 +07:00
|
|
|
$(call cc-ldoption, -Wl$(comma)--build-id) -Wl,-Bsymbolic $(LTO_CFLAGS)
|
2009-06-18 06:28:09 +07:00
|
|
|
GCOV_PROFILE := n
|
2008-01-30 19:30:42 +07:00
|
|
|
|
|
|
|
#
|
2014-06-21 02:20:44 +07:00
|
|
|
# Install the unstripped copies of vdso*.so. If our toolchain supports
|
|
|
|
# build-id, install .build-id links as well.
|
2008-01-30 19:30:42 +07:00
|
|
|
#
|
2014-06-12 22:28:10 +07:00
|
|
|
quiet_cmd_vdso_install = INSTALL $(@:install_%=%)
|
2014-06-21 02:20:44 +07:00
|
|
|
define cmd_vdso_install
|
|
|
|
cp $< "$(MODLIB)/vdso/$(@:install_%=%)"; \
|
|
|
|
if readelf -n $< |grep -q 'Build ID'; then \
|
|
|
|
buildid=`readelf -n $< |grep 'Build ID' |sed -e 's/^.*Build ID: \(.*\)$$/\1/'`; \
|
|
|
|
first=`echo $$buildid | cut -b-2`; \
|
|
|
|
last=`echo $$buildid | cut -b3-`; \
|
|
|
|
mkdir -p "$(MODLIB)/vdso/.build-id/$$first"; \
|
|
|
|
ln -sf "../../$(@:install_%=%)" "$(MODLIB)/vdso/.build-id/$$first/$$last.debug"; \
|
|
|
|
fi
|
|
|
|
endef
|
2014-06-12 22:28:10 +07:00
|
|
|
|
|
|
|
vdso_img_insttargets := $(vdso_img_sodbg:%.dbg=install_%)
|
|
|
|
|
|
|
|
$(MODLIB)/vdso: FORCE
|
2007-10-17 23:04:32 +07:00
|
|
|
@mkdir -p $(MODLIB)/vdso
|
2014-06-12 22:28:10 +07:00
|
|
|
|
|
|
|
$(vdso_img_insttargets): install_%: $(obj)/%.dbg $(MODLIB)/vdso FORCE
|
2007-10-17 23:04:32 +07:00
|
|
|
$(call cmd,vdso_install)
|
|
|
|
|
2014-06-12 22:28:10 +07:00
|
|
|
PHONY += vdso_install $(vdso_img_insttargets)
|
|
|
|
vdso_install: $(vdso_img_insttargets) FORCE
|
2008-01-30 19:32:27 +07:00
|
|
|
|
2015-03-28 01:48:19 +07:00
|
|
|
clean-files := vdso32-syscall* vdso32-sysenter* vdso32-int80* vdso64* vdso-image-*.c vdsox32.so*
|