dataMonitoring.h 5.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122
  1. #ifndef _ASM_IA64_ELF_H
  2. #define _ASM_IA64_ELF_H
  3. /*
  4. * ELF-specific definitions.
  5. *
  6. * Copyright (C) 1998-1999, 2002-2004 Hewlett-Packard Co
  7. * David Mosberger-Tang <davidm@hpl.hp.com>
  8. */
  9. #include <asm/fpu.h>
  10. #include <asm/page.h>
  11. #include <asm/auxvec.h>
  12. /*
  13. * This is used to ensure we don't load something for the wrong architecture.
  14. */
  15. #define elf_check_arch(x) ((x)->e_machine == EM_IA_64)
  16. /*
  17. * These are used to set parameters in the core dumps.
  18. */
  19. #define ELF_CLASS ELFCLASS64
  20. #define ELF_DATA ELFDATA2LSB
  21. #define ELF_ARCH EM_IA_64
  22. #define CORE_DUMP_USE_REGSET
  23. /* Least-significant four bits of ELF header's e_flags are OS-specific. The bits are
  24. interpreted as follows by Linux: */
  25. #define EF_IA_64_LINUX_EXECUTABLE_STACK 0x1 /* is stack (& heap) executable by default? */
  26. #define ELF_EXEC_PAGESIZE PAGE_SIZE
  27. /*
  28. * This is the location that an ET_DYN program is loaded if exec'ed.
  29. * Typical use of this is to invoke "./ld.so someprog" to test out a
  30. * new version of the loader. We need to make sure that it is out of
  31. * the way of the program that it will "exec", and that there is
  32. * sufficient room for the brk.
  33. */
  34. #define ELF_ET_DYN_BASE (TASK_UNMAPPED_BASE + 0x800000000UL)
  35. #define PT_IA_64_UNWIND 0x70000001
  36. /* IA-64 relocations: */
  37. #define R_IA64_NONE 0x00 /* none */
  38. #define R_IA64_IMM14 0x21 /* symbol + addend, add imm14 */
  39. #define R_IA64_IMM22 0x22 /* symbol + addend, add imm22 */
  40. #define R_IA64_IMM64 0x23 /* symbol + addend, mov imm64 */
  41. #define R_IA64_DIR32MSB 0x24 /* symbol + addend, data4 MSB */
  42. #define R_IA64_DIR32LSB 0x25 /* symbol + addend, data4 LSB */
  43. #define R_IA64_DIR64MSB 0x26 /* symbol + addend, data8 MSB */
  44. #define R_IA64_DIR64LSB 0x27 /* symbol + addend, data8 LSB */
  45. #define R_IA64_GPREL22 0x2a /* @gprel(sym+add), add imm22 */
  46. #define R_IA64_GPREL64I 0x2b /* @gprel(sym+add), mov imm64 */
  47. #define R_IA64_GPREL32MSB 0x2c /* @gprel(sym+add), data4 MSB */
  48. #define R_IA64_GPREL32LSB 0x2d /* @gprel(sym+add), data4 LSB */
  49. #define R_IA64_GPREL64MSB 0x2e /* @gprel(sym+add), data8 MSB */
  50. #define R_IA64_GPREL64LSB 0x2f /* @gprel(sym+add), data8 LSB */
  51. #define R_IA64_LTOFF22 0x32 /* @ltoff(sym+add), add imm22 */
  52. #define R_IA64_LTOFF64I 0x33 /* @ltoff(sym+add), mov imm64 */
  53. #define R_IA64_PLTOFF22 0x3a /* @pltoff(sym+add), add imm22 */
  54. #define R_IA64_PLTOFF64I 0x3b /* @pltoff(sym+add), mov imm64 */
  55. #define R_IA64_PLTOFF64MSB 0x3e /* @pltoff(sym+add), data8 MSB */
  56. #define R_IA64_PLTOFF64LSB 0x3f /* @pltoff(sym+add), data8 LSB */
  57. #define R_IA64_FPTR64I 0x43 /* @fptr(sym+add), mov imm64 */
  58. #define R_IA64_FPTR32MSB 0x44 /* @fptr(sym+add), data4 MSB */
  59. #define R_IA64_FPTR32LSB 0x45 /* @fptr(sym+add), data4 LSB */
  60. #define R_IA64_FPTR64MSB 0x46 /* @fptr(sym+add), data8 MSB */
  61. #define R_IA64_FPTR64LSB 0x47 /* @fptr(sym+add), data8 LSB */
  62. #define R_IA64_PCREL60B 0x48 /* @pcrel(sym+add), brl */
  63. #define R_IA64_PCREL21B 0x49 /* @pcrel(sym+add), ptb, call */
  64. #define R_IA64_PCREL21M 0x4a /* @pcrel(sym+add), chk.s */
  65. #define R_IA64_PCREL21F 0x4b /* @pcrel(sym+add), fchkf */
  66. #define R_IA64_PCREL32MSB 0x4c /* @pcrel(sym+add), data4 MSB */
  67. #define R_IA64_PCREL32LSB 0x4d /* @pcrel(sym+add), data4 LSB */
  68. #define R_IA64_PCREL64MSB 0x4e /* @pcrel(sym+add), data8 MSB */
  69. #define R_IA64_PCREL64LSB 0x4f /* @pcrel(sym+add), data8 LSB */
  70. #define R_IA64_LTOFF_FPTR22 0x52 /* @ltoff(@fptr(s+a)), imm22 */
  71. #define R_IA64_LTOFF_FPTR64I 0x53 /* @ltoff(@fptr(s+a)), imm64 */
  72. #define R_IA64_LTOFF_FPTR32MSB 0x54 /* @ltoff(@fptr(s+a)), 4 MSB */
  73. #define R_IA64_LTOFF_FPTR32LSB 0x55 /* @ltoff(@fptr(s+a)), 4 LSB */
  74. #define R_IA64_LTOFF_FPTR64MSB 0x56 /* @ltoff(@fptr(s+a)), 8 MSB */
  75. #define R_IA64_LTOFF_FPTR64LSB 0x57 /* @ltoff(@fptr(s+a)), 8 LSB */
  76. #define R_IA64_SEGREL32MSB 0x5c /* @segrel(sym+add), data4 MSB */
  77. #define R_IA64_SEGREL32LSB 0x5d /* @segrel(sym+add), data4 LSB */
  78. #define R_IA64_SEGREL64MSB 0x5e /* @segrel(sym+add), data8 MSB */
  79. #define R_IA64_SEGREL64LSB 0x5f /* @segrel(sym+add), data8 LSB */
  80. #define R_IA64_SECREL32MSB 0x64 /* @secrel(sym+add), data4 MSB */
  81. #define R_IA64_SECREL32LSB 0x65 /* @secrel(sym+add), data4 LSB */
  82. #define R_IA64_SECREL64MSB 0x66 /* @secrel(sym+add), data8 MSB */
  83. #define R_IA64_SECREL64LSB 0x67 /* @secrel(sym+add), data8 LSB */
  84. #define R_IA64_REL32MSB 0x6c /* data 4 + REL */
  85. #define R_IA64_REL32LSB 0x6d /* data 4 + REL */
  86. #define R_IA64_REL64MSB 0x6e /* data 8 + REL */
  87. #define R_IA64_REL64LSB 0x6f /* data 8 + REL */
  88. #define R_IA64_LTV32MSB 0x74 /* symbol + addend, data4 MSB */
  89. #define R_IA64_LTV32LSB 0x75 /* symbol + addend, data4 LSB */
  90. #define R_IA64_LTV64MSB 0x76 /* symbol + addend, data8 MSB */
  91. #define R_IA64_LTV64LSB 0x77 /* symbol + addend, data8 LSB */
  92. #define R_IA64_PCREL21BI 0x79 /* @pcrel(sym+add), ptb, call */
  93. #define R_IA64_PCREL22 0x7a /* @pcrel(sym+add), imm22 */
  94. #define R_IA64_PCREL64I 0x7b /* @pcrel(sym+add), imm64 */
  95. #define R_IA64_IPLTMSB 0x80 /* dynamic reloc, imported PLT, MSB */
  96. #define R_IA64_IPLTLSB 0x81 /* dynamic reloc, imported PLT, LSB */
  97. #define R_IA64_COPY 0x84 /* dynamic reloc, data copy */
  98. #define R_IA64_SUB 0x85 /* -symbol + addend, add imm22 */
  99. #define R_IA64_LTOFF22X 0x86 /* LTOFF22, relaxable. */
  100. #define R_IA64_LDXMOV 0x87 /* Use of LTOFF22X. */
  101. #define R_IA64_TPREL14 0x91 /* @tprel(sym+add), add imm14 */
  102. #define R_IA64_TPREL22 0x92 /* @tprel(sym+add), add imm22 */
  103. #define R_IA64_TPREL64I 0x93 /* @tprel(sym+add), add imm64 */
  104. #define R_IA64_TPREL64MSB 0x96 /* @tprel(sym+add), data8 MSB */
  105. #define R_IA64_TPREL64LSB 0x97 /* @tprel(sym+add), data8 LSB */
  106. #define R_IA64_LTOFF_TPREL22 0x9a /* @ltoff(@tprel(s+a)), add imm22 */
  107. #define R_IA64_DTPMOD64MSB 0xa6 /* @dtpmod(sym+add), data8 MSB */
  108. #define R_IA64_DTPMOD64LSB 0xa7 /* @dtpmod(sym+add), data8 LSB */
  109. #define R_IA64_LTOFF_DTPMOD22 0xaa /* @ltoff(@dtpmod(s+a)), imm22 */
  110. #define R_IA64_DTPREL14 0xb1 /* @dtprel(sym+add), imm14 */
  111. #define R_IA64_DTPREL22 0xb2 /* @dtprel(sym+add), imm22 */