blob: 4655ff342c6406498c4640fca7c470333205f4f4 [file] [log] [blame]
Kyle Swenson8d8f6542021-03-15 11:02:55 -06001/*
2 * Copyright (C) 2007-2009 Michal Simek <monstr@monstr.eu>
3 * Copyright (C) 2007-2009 PetaLogix
4 * Copyright (C) 2006 Atmark Techno, Inc.
5 *
6 * MMU code derived from arch/ppc/kernel/head_4xx.S:
7 * Copyright (c) 1995-1996 Gary Thomas <gdt@linuxppc.org>
8 * Initial PowerPC version.
9 * Copyright (c) 1996 Cort Dougan <cort@cs.nmt.edu>
10 * Rewritten for PReP
11 * Copyright (c) 1996 Paul Mackerras <paulus@cs.anu.edu.au>
12 * Low-level exception handers, MMU support, and rewrite.
13 * Copyright (c) 1997 Dan Malek <dmalek@jlc.net>
14 * PowerPC 8xx modifications.
15 * Copyright (c) 1998-1999 TiVo, Inc.
16 * PowerPC 403GCX modifications.
17 * Copyright (c) 1999 Grant Erickson <grant@lcse.umn.edu>
18 * PowerPC 403GCX/405GP modifications.
19 * Copyright 2000 MontaVista Software Inc.
20 * PPC405 modifications
21 * PowerPC 403GCX/405GP modifications.
22 * Author: MontaVista Software, Inc.
23 * frank_rowand@mvista.com or source@mvista.com
24 * debbie_chu@mvista.com
25 *
26 * This file is subject to the terms and conditions of the GNU General Public
27 * License. See the file "COPYING" in the main directory of this archive
28 * for more details.
29 */
30
31#include <linux/init.h>
32#include <linux/linkage.h>
33#include <asm/thread_info.h>
34#include <asm/page.h>
35#include <linux/of_fdt.h> /* for OF_DT_HEADER */
36
37#ifdef CONFIG_MMU
38#include <asm/setup.h> /* COMMAND_LINE_SIZE */
39#include <asm/mmu.h>
40#include <asm/processor.h>
41
42.section .data
43.global empty_zero_page
44.align 12
45empty_zero_page:
46 .space PAGE_SIZE
47.global swapper_pg_dir
48swapper_pg_dir:
49 .space PAGE_SIZE
50
51#endif /* CONFIG_MMU */
52
53.section .rodata
54.align 4
55endian_check:
56 .word 1
57
58 __HEAD
59ENTRY(_start)
60#if CONFIG_KERNEL_BASE_ADDR == 0
61 brai TOPHYS(real_start)
62 .org 0x100
63real_start:
64#endif
65
66 mts rmsr, r0
67/* Disable stack protection from bootloader */
68 mts rslr, r0
69 addi r8, r0, 0xFFFFFFFF
70 mts rshr, r8
71/*
72 * According to Xilinx, msrclr instruction behaves like 'mfs rX,rpc'
73 * if the msrclr instruction is not enabled. We use this to detect
74 * if the opcode is available, by issuing msrclr and then testing the result.
75 * r8 == 0 - msr instructions are implemented
76 * r8 != 0 - msr instructions are not implemented
77 */
78 mfs r1, rmsr
79 msrclr r8, 0 /* clear nothing - just read msr for test */
80 cmpu r8, r8, r1 /* r1 must contain msr reg content */
81
82/* r7 may point to an FDT, or there may be one linked in.
83 if it's in r7, we've got to save it away ASAP.
84 We ensure r7 points to a valid FDT, just in case the bootloader
85 is broken or non-existent */
86 beqi r7, no_fdt_arg /* NULL pointer? don't copy */
87/* Does r7 point to a valid FDT? Load HEADER magic number */
88 /* Run time Big/Little endian platform */
89 /* Save 1 as word and load byte - 0 - BIG, 1 - LITTLE */
90 lbui r11, r0, TOPHYS(endian_check)
91 beqid r11, big_endian /* DO NOT break delay stop dependency */
92 lw r11, r0, r7 /* Big endian load in delay slot */
93 lwr r11, r0, r7 /* Little endian load */
94big_endian:
95 rsubi r11, r11, OF_DT_HEADER /* Check FDT header */
96 beqi r11, _prepare_copy_fdt
97 or r7, r0, r0 /* clear R7 when not valid DTB */
98 bnei r11, no_fdt_arg /* No - get out of here */
99_prepare_copy_fdt:
100 or r11, r0, r0 /* incremment */
101 ori r4, r0, TOPHYS(_fdt_start)
102 ori r3, r0, (0x8000 - 4)
103_copy_fdt:
104 lw r12, r7, r11 /* r12 = r7 + r11 */
105 sw r12, r4, r11 /* addr[r4 + r11] = r12 */
106 addik r11, r11, 4 /* increment counting */
107 bgtid r3, _copy_fdt /* loop for all entries */
108 addik r3, r3, -4 /* descrement loop */
109no_fdt_arg:
110
111#ifdef CONFIG_MMU
112
113#ifndef CONFIG_CMDLINE_BOOL
114/*
115 * handling command line
116 * copy command line directly to cmd_line placed in data section.
117 */
118 beqid r5, skip /* Skip if NULL pointer */
119 or r11, r0, r0 /* incremment */
120 ori r4, r0, cmd_line /* load address of command line */
121 tophys(r4,r4) /* convert to phys address */
122 ori r3, r0, COMMAND_LINE_SIZE - 1 /* number of loops */
123_copy_command_line:
124 /* r2=r5+r6 - r5 contain pointer to command line */
125 lbu r2, r5, r11
126 beqid r2, skip /* Skip if no data */
127 sb r2, r4, r11 /* addr[r4+r6]= r2 */
128 addik r11, r11, 1 /* increment counting */
129 bgtid r3, _copy_command_line /* loop for all entries */
130 addik r3, r3, -1 /* decrement loop */
131 addik r5, r4, 0 /* add new space for command line */
132 tovirt(r5,r5)
133skip:
134#endif /* CONFIG_CMDLINE_BOOL */
135
136#ifdef NOT_COMPILE
137/* save bram context */
138 or r11, r0, r0 /* incremment */
139 ori r4, r0, TOPHYS(_bram_load_start) /* save bram context */
140 ori r3, r0, (LMB_SIZE - 4)
141_copy_bram:
142 lw r7, r0, r11 /* r7 = r0 + r6 */
143 sw r7, r4, r11 /* addr[r4 + r6] = r7 */
144 addik r11, r11, 4 /* increment counting */
145 bgtid r3, _copy_bram /* loop for all entries */
146 addik r3, r3, -4 /* descrement loop */
147#endif
148 /* We have to turn on the MMU right away. */
149
150 /*
151 * Set up the initial MMU state so we can do the first level of
152 * kernel initialization. This maps the first 16 MBytes of memory 1:1
153 * virtual to physical.
154 */
155 nop
156 addik r3, r0, MICROBLAZE_TLB_SIZE -1 /* Invalidate all TLB entries */
157_invalidate:
158 mts rtlbx, r3
159 mts rtlbhi, r0 /* flush: ensure V is clear */
160 mts rtlblo, r0
161 bgtid r3, _invalidate /* loop for all entries */
162 addik r3, r3, -1
163 /* sync */
164
165 /* Setup the kernel PID */
166 mts rpid,r0 /* Load the kernel PID */
167 nop
168 bri 4
169
170 /*
171 * We should still be executing code at physical address area
172 * RAM_BASEADDR at this point. However, kernel code is at
173 * a virtual address. So, set up a TLB mapping to cover this once
174 * translation is enabled.
175 */
176
177 addik r3,r0, CONFIG_KERNEL_START /* Load the kernel virtual address */
178 tophys(r4,r3) /* Load the kernel physical address */
179
180 /* start to do TLB calculation */
181 addik r12, r0, _end
182 rsub r12, r3, r12
183 addik r12, r12, CONFIG_LOWMEM_SIZE >> PTE_SHIFT /* that's the pad */
184
185 or r9, r0, r0 /* TLB0 = 0 */
186 or r10, r0, r0 /* TLB1 = 0 */
187
188 addik r11, r12, -0x1000000
189 bgei r11, GT16 /* size is greater than 16MB */
190 addik r11, r12, -0x0800000
191 bgei r11, GT8 /* size is greater than 8MB */
192 addik r11, r12, -0x0400000
193 bgei r11, GT4 /* size is greater than 4MB */
194 /* size is less than 4MB */
195 addik r11, r12, -0x0200000
196 bgei r11, GT2 /* size is greater than 2MB */
197 addik r9, r0, 0x0100000 /* TLB0 must be 1MB */
198 addik r11, r12, -0x0100000
199 bgei r11, GT1 /* size is greater than 1MB */
200 /* TLB1 is 0 which is setup above */
201 bri tlb_end
202GT4: /* r11 contains the rest - will be either 1 or 4 */
203 ori r9, r0, 0x400000 /* TLB0 is 4MB */
204 bri TLB1
205GT16: /* TLB0 is 16MB */
206 addik r9, r0, 0x1000000 /* means TLB0 is 16MB */
207TLB1:
208 /* must be used r2 because of subtract if failed */
209 addik r2, r11, -0x0400000
210 bgei r2, GT20 /* size is greater than 16MB */
211 /* size is >16MB and <20MB */
212 addik r11, r11, -0x0100000
213 bgei r11, GT17 /* size is greater than 17MB */
214 /* kernel is >16MB and < 17MB */
215GT1:
216 addik r10, r0, 0x0100000 /* means TLB1 is 1MB */
217 bri tlb_end
218GT2: /* TLB0 is 0 and TLB1 will be 4MB */
219GT17: /* TLB1 is 4MB - kernel size <20MB */
220 addik r10, r0, 0x0400000 /* means TLB1 is 4MB */
221 bri tlb_end
222GT8: /* TLB0 is still zero that's why I can use only TLB1 */
223GT20: /* TLB1 is 16MB - kernel size >20MB */
224 addik r10, r0, 0x1000000 /* means TLB1 is 16MB */
225tlb_end:
226
227 /*
228 * Configure and load two entries into TLB slots 0 and 1.
229 * In case we are pinning TLBs, these are reserved in by the
230 * other TLB functions. If not reserving, then it doesn't
231 * matter where they are loaded.
232 */
233 andi r4,r4,0xfffffc00 /* Mask off the real page number */
234 ori r4,r4,(TLB_WR | TLB_EX) /* Set the write and execute bits */
235
236 /*
237 * TLB0 is always used - check if is not zero (r9 stores TLB0 value)
238 * if is use TLB1 value and clear it (r10 stores TLB1 value)
239 */
240 bnei r9, tlb0_not_zero
241 add r9, r10, r0
242 add r10, r0, r0
243tlb0_not_zero:
244
245 /* look at the code below */
246 ori r30, r0, 0x200
247 andi r29, r9, 0x100000
248 bneid r29, 1f
249 addik r30, r30, 0x80
250 andi r29, r9, 0x400000
251 bneid r29, 1f
252 addik r30, r30, 0x80
253 andi r29, r9, 0x1000000
254 bneid r29, 1f
255 addik r30, r30, 0x80
2561:
257 andi r3,r3,0xfffffc00 /* Mask off the effective page number */
258 ori r3,r3,(TLB_VALID)
259 or r3, r3, r30
260
261 /* Load tlb_skip size value which is index to first unused TLB entry */
262 lwi r11, r0, TOPHYS(tlb_skip)
263 mts rtlbx,r11 /* TLB slow 0 */
264
265 mts rtlblo,r4 /* Load the data portion of the entry */
266 mts rtlbhi,r3 /* Load the tag portion of the entry */
267
268 /* Increase tlb_skip size */
269 addik r11, r11, 1
270 swi r11, r0, TOPHYS(tlb_skip)
271
272 /* TLB1 can be zeroes that's why we not setup it */
273 beqi r10, jump_over2
274
275 /* look at the code below */
276 ori r30, r0, 0x200
277 andi r29, r10, 0x100000
278 bneid r29, 1f
279 addik r30, r30, 0x80
280 andi r29, r10, 0x400000
281 bneid r29, 1f
282 addik r30, r30, 0x80
283 andi r29, r10, 0x1000000
284 bneid r29, 1f
285 addik r30, r30, 0x80
2861:
287 addk r4, r4, r9 /* previous addr + TLB0 size */
288 addk r3, r3, r9
289
290 andi r3,r3,0xfffffc00 /* Mask off the effective page number */
291 ori r3,r3,(TLB_VALID)
292 or r3, r3, r30
293
294 lwi r11, r0, TOPHYS(tlb_skip)
295 mts rtlbx, r11 /* r11 is used from TLB0 */
296
297 mts rtlblo,r4 /* Load the data portion of the entry */
298 mts rtlbhi,r3 /* Load the tag portion of the entry */
299
300 /* Increase tlb_skip size */
301 addik r11, r11, 1
302 swi r11, r0, TOPHYS(tlb_skip)
303
304jump_over2:
305 /*
306 * Load a TLB entry for LMB, since we need access to
307 * the exception vectors, using a 4k real==virtual mapping.
308 */
309 /* Use temporary TLB_ID for LMB - clear this temporary mapping later */
310 ori r11, r0, MICROBLAZE_LMB_TLB_ID
311 mts rtlbx,r11
312
313 ori r4,r0,(TLB_WR | TLB_EX)
314 ori r3,r0,(TLB_VALID | TLB_PAGESZ(PAGESZ_4K))
315
316 mts rtlblo,r4 /* Load the data portion of the entry */
317 mts rtlbhi,r3 /* Load the tag portion of the entry */
318
319 /*
320 * We now have the lower 16 Meg of RAM mapped into TLB entries, and the
321 * caches ready to work.
322 */
323turn_on_mmu:
324 ori r15,r0,start_here
325 ori r4,r0,MSR_KERNEL_VMS
326 mts rmsr,r4
327 nop
328 rted r15,0 /* enables MMU */
329 nop
330
331start_here:
332#endif /* CONFIG_MMU */
333
334 /* Initialize small data anchors */
335 addik r13, r0, _KERNEL_SDA_BASE_
336 addik r2, r0, _KERNEL_SDA2_BASE_
337
338 /* Initialize stack pointer */
339 addik r1, r0, init_thread_union + THREAD_SIZE - 4
340
341 /* Initialize r31 with current task address */
342 addik r31, r0, init_task
343
344 /*
345 * Call platform dependent initialize function.
346 * Please see $(ARCH)/mach-$(SUBARCH)/setup.c for
347 * the function.
348 */
349 addik r11, r0, machine_early_init
350 brald r15, r11
351 nop
352
353#ifndef CONFIG_MMU
354 addik r15, r0, machine_halt
355 braid start_kernel
356 nop
357#else
358 /*
359 * Initialize the MMU.
360 */
361 bralid r15, mmu_init
362 nop
363
364 /* Go back to running unmapped so we can load up new values
365 * and change to using our exception vectors.
366 * On the MicroBlaze, all we invalidate the used TLB entries to clear
367 * the old 16M byte TLB mappings.
368 */
369 ori r15,r0,TOPHYS(kernel_load_context)
370 ori r4,r0,MSR_KERNEL
371 mts rmsr,r4
372 nop
373 bri 4
374 rted r15,0
375 nop
376
377 /* Load up the kernel context */
378kernel_load_context:
379 ori r5, r0, MICROBLAZE_LMB_TLB_ID
380 mts rtlbx,r5
381 nop
382 mts rtlbhi,r0
383 nop
384 addi r15, r0, machine_halt
385 ori r17, r0, start_kernel
386 ori r4, r0, MSR_KERNEL_VMS
387 mts rmsr, r4
388 nop
389 rted r17, 0 /* enable MMU and jump to start_kernel */
390 nop
391#endif /* CONFIG_MMU */