1
2#include <asm/cache.h>
3#include <asm/ptrace.h>
4#include <asm/system.h>
5#include <asm/pgtable.h>
6
7#include <asm-generic/vmlinux.lds.h>
8
9OUTPUT_FORMAT("elf64-ia64-little")
10OUTPUT_ARCH(ia64)
11ENTRY(phys_start)
12jiffies = jiffies_64;
13
14PHDRS {
15	code   PT_LOAD;
16	percpu PT_LOAD;
17	data   PT_LOAD;
18	note   PT_NOTE;
19	unwind 0x70000001; /* PT_IA_64_UNWIND, but ld doesn't match the name */
20}
21
22SECTIONS {
23	/*
24	 * unwind exit sections must be discarded before
25	 * the rest of the sections get included.
26	 */
27	/DISCARD/ : {
28		*(.IA_64.unwind.exit.text)
29		*(.IA_64.unwind_info.exit.text)
30		*(.comment)
31		*(.note)
32	}
33
34	v = PAGE_OFFSET; /* this symbol is here to make debugging easier... */
35	phys_start = _start - LOAD_OFFSET;
36
37	code : {
38	} :code
39	. = KERNEL_START;
40
41	_text = .;
42	_stext = .;
43
44	.text : AT(ADDR(.text) - LOAD_OFFSET) {
45		__start_ivt_text = .;
46		*(.text..ivt)
47		__end_ivt_text = .;
48		TEXT_TEXT
49		SCHED_TEXT
50		LOCK_TEXT
51		KPROBES_TEXT
52		*(.gnu.linkonce.t*)
53	}
54
55	.text2 : AT(ADDR(.text2) - LOAD_OFFSET)	{
56		*(.text2)
57	}
58
59#ifdef CONFIG_SMP
60	.text..lock : AT(ADDR(.text..lock) - LOAD_OFFSET) {
61		*(.text..lock)
62	}
63#endif
64	_etext = .;
65
66	/*
67	 * Read-only data
68	 */
69	NOTES :code :note       /* put .notes in text and mark in PT_NOTE  */
70	code_continues : {
71	} : code               /* switch back to regular program...  */
72
73	EXCEPTION_TABLE(16)
74
75	/* MCA table */
76	. = ALIGN(16);
77	__mca_table : AT(ADDR(__mca_table) - LOAD_OFFSET) {
78		__start___mca_table = .;
79		*(__mca_table)
80		__stop___mca_table = .;
81	}
82
83	.data..patch.phys_stack_reg : AT(ADDR(.data..patch.phys_stack_reg) - LOAD_OFFSET) {
84		__start___phys_stack_reg_patchlist = .;
85		*(.data..patch.phys_stack_reg)
86		__end___phys_stack_reg_patchlist = .;
87	}
88
89	/*
90	 * Global data
91	 */
92	_data = .;
93
94	/* Unwind info & table: */
95	. = ALIGN(8);
96	.IA_64.unwind_info : AT(ADDR(.IA_64.unwind_info) - LOAD_OFFSET) {
97		*(.IA_64.unwind_info*)
98	}
99	.IA_64.unwind : AT(ADDR(.IA_64.unwind) - LOAD_OFFSET) {
100		__start_unwind = .;
101		*(.IA_64.unwind*)
102		__end_unwind = .;
103	} :code :unwind
104	code_continues2 : {
105	} : code
106
107	RODATA
108
109	.opd : AT(ADDR(.opd) - LOAD_OFFSET) {
110		*(.opd)
111	}
112
113	/*
114	 * Initialization code and data:
115	 */
116	. = ALIGN(PAGE_SIZE);
117	__init_begin = .;
118
119	INIT_TEXT_SECTION(PAGE_SIZE)
120	INIT_DATA_SECTION(16)
121
122	.data..patch.vtop : AT(ADDR(.data..patch.vtop) - LOAD_OFFSET) {
123		__start___vtop_patchlist = .;
124		*(.data..patch.vtop)
125		__end___vtop_patchlist = .;
126	}
127
128	.data..patch.rse : AT(ADDR(.data..patch.rse) - LOAD_OFFSET) {
129		__start___rse_patchlist = .;
130		*(.data..patch.rse)
131		__end___rse_patchlist = .;
132	}
133
134	.data..patch.mckinley_e9 : AT(ADDR(.data..patch.mckinley_e9) - LOAD_OFFSET) {
135		__start___mckinley_e9_bundles = .;
136		*(.data..patch.mckinley_e9)
137		__end___mckinley_e9_bundles = .;
138	}
139
140#if defined(CONFIG_PARAVIRT)
141	. = ALIGN(16);
142	.paravirt_bundles : AT(ADDR(.paravirt_bundles) - LOAD_OFFSET) {
143		__start_paravirt_bundles = .;
144		*(.paravirt_bundles)
145		__stop_paravirt_bundles = .;
146	}
147	. = ALIGN(16);
148	.paravirt_insts : AT(ADDR(.paravirt_insts) - LOAD_OFFSET) {
149		__start_paravirt_insts = .;
150		*(.paravirt_insts)
151		__stop_paravirt_insts = .;
152	}
153	. = ALIGN(16);
154	.paravirt_branches : AT(ADDR(.paravirt_branches) - LOAD_OFFSET) {
155		__start_paravirt_branches = .;
156		*(.paravirt_branches)
157		__stop_paravirt_branches = .;
158	}
159#endif
160
161#if defined(CONFIG_IA64_GENERIC)
162	/* Machine Vector */
163	. = ALIGN(16);
164	.machvec : AT(ADDR(.machvec) - LOAD_OFFSET) {
165		machvec_start = .;
166		*(.machvec)
167		machvec_end = .;
168	}
169#endif
170
171#ifdef	CONFIG_SMP
172	. = ALIGN(PERCPU_PAGE_SIZE);
173	__cpu0_per_cpu = .;
174	. = . + PERCPU_PAGE_SIZE;   /* cpu0 per-cpu space */
175#endif
176
177	. = ALIGN(PAGE_SIZE);
178	__init_end = .;
179
180	.data..page_aligned : AT(ADDR(.data..page_aligned) - LOAD_OFFSET) {
181		PAGE_ALIGNED_DATA(PAGE_SIZE)
182		. = ALIGN(PAGE_SIZE);
183		__start_gate_section = .;
184		*(.data..gate)
185		__stop_gate_section = .;
186#ifdef CONFIG_XEN
187		. = ALIGN(PAGE_SIZE);
188		__xen_start_gate_section = .;
189		*(.data..gate.xen)
190		__xen_stop_gate_section = .;
191#endif
192	}
193	/*
194	 * make sure the gate page doesn't expose
195	 * kernel data
196	 */
197	. = ALIGN(PAGE_SIZE);
198
199	/* Per-cpu data: */
200	. = ALIGN(PERCPU_PAGE_SIZE);
201	PERCPU_VADDR(SMP_CACHE_BYTES, PERCPU_ADDR, :percpu)
202	__phys_per_cpu_start = __per_cpu_load;
203	/*
204	 * ensure percpu data fits
205	 * into percpu page size
206	 */
207	. = __phys_per_cpu_start + PERCPU_PAGE_SIZE;
208
209	data : {
210	} :data
211	.data : AT(ADDR(.data) - LOAD_OFFSET) {
212		INIT_TASK_DATA(PAGE_SIZE)
213		CACHELINE_ALIGNED_DATA(SMP_CACHE_BYTES)
214		READ_MOSTLY_DATA(SMP_CACHE_BYTES)
215		DATA_DATA
216		*(.data1)
217		*(.gnu.linkonce.d*)
218		CONSTRUCTORS
219	}
220
221	. = ALIGN(16);	/* gp must be 16-byte aligned for exc. table */
222	.got : AT(ADDR(.got) - LOAD_OFFSET) {
223		*(.got.plt)
224		*(.got)
225	}
226	__gp = ADDR(.got) + 0x200000;
227
228	/*
229	 * We want the small data sections together,
230	 * so single-instruction offsets can access
231	 * them all, and initialized data all before
232	 * uninitialized, so we can shorten the
233	 * on-disk segment size.
234	 */
235	.sdata : AT(ADDR(.sdata) - LOAD_OFFSET) {
236		*(.sdata)
237		*(.sdata1)
238		*(.srdata)
239	}
240	_edata  =  .;
241
242	BSS_SECTION(0, 0, 0)
243
244	_end = .;
245
246	code : {
247	} :code
248
249	STABS_DEBUG
250	DWARF_DEBUG
251
252	/* Default discards */
253	DISCARDS
254}
255