1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
|
/*
* arch/arm/mach-tegra/headsmp.S
*
* CPU initialization routines for Tegra SoCs
*
* Copyright (c) 2009-2012, NVIDIA Corporation.
* Copyright (c) 2011 Google, Inc.
* Author: Colin Cross <ccross@android.com>
* Gary King <gking@nvidia.com>
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License.
*
* This program is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
*/
#include <linux/linkage.h>
#include <linux/init.h>
#include <asm/assembler.h>
#include <asm/cache.h>
#include <asm/page.h>
#include <mach/iomap.h>
#include <mach/io.h>
#include "asm_macros.h"
#include "flowctrl.h"
#include "sleep.h"
#include "reset.h"
#define APB_MISC_GP_HIDREV 0x804
#define PMC_SCRATCH41 0x140
#define DEBUG_CPU_RESET_HANDLER 0 /* Non-zero enables debug code */
#define RESET_DATA(x) ((TEGRA_RESET_##x)*4)
#ifdef CONFIG_SMP
/*
* tegra_secondary_startup
*
* Initial secondary processor boot vector; jumps to kernel's
* secondary_startup routine. Used for initial boot and hotplug
* of secondary CPUs.
*/
__CPUINIT
ENTRY(tegra_secondary_startup)
bl __invalidate_cpu_state
/* enable user space perf counter access */
/* only accessible in secure state */
mrc p15, 0, r0, c9, c12, 0
lsr r0, #11
and r0, r0, #0x1f
mov r1, #1
lsl r1, r1, r0
sub r1, r1, #1
movt r1, #0x8000
mcr p15, 0, r1, c9, c14, 2
mov r0, #1
mcr p15, 0, r0, c9, c14, 0
b secondary_startup
ENDPROC(tegra_secondary_startup)
#endif
.section ".text.head", "ax"
#ifdef CONFIG_PM_SLEEP
/*
* tegra_resume
*
* CPU boot vector when restarting the a CPU following
* an LP2 transition. Also branched to by LP0 and LP1 resume after
* re-enabling sdram.
*/
ENTRY(tegra_resume)
#ifdef CONFIG_TRUSTED_FOUNDATIONS
mov32 r1, TEGRA_TMRUS_BASE
ldr r0, [r1]
adr r1, tegra_resume_entry_time
str r0, [r1]
#endif
bl __invalidate_cpu_state
cpu_id r0
#ifndef CONFIG_TEGRA_VIRTUAL_CPUID
cmp r0, #0 @ CPU0?
bne cpu_resume @ no
#endif
#ifndef CONFIG_ARCH_TEGRA_2x_SOC
@ Clear the flow controller flags for this CPU.
cpu_to_csr_reg r1, r0
mov32 r2, TEGRA_FLOW_CTRL_BASE
ldr r1, [r2, r1]
orr r1, r1, #(1 << 15) | (1 << 14) @ write to clear event & intr
movw r0, #0x3FFD @ enable, enable_ext, cluster_switch, immed, & bitmaps
bic r1, r1, r0
str r1, [r2]
#endif
#if defined(CONFIG_HAVE_ARM_SCU)
/* enable SCU */
mov32 r0, TEGRA_ARM_PERIF_BASE
ldr r1, [r0]
orr r1, r1, #1
str r1, [r0]
#endif
#ifdef CONFIG_TRUSTED_FOUNDATIONS
#ifndef CONFIG_ARCH_TEGRA_11x_SOC
mov32 r1, TEGRA_TMRUS_BASE
ldr r0, [r1]
adr r1, tegra_resume_smc_entry_time
str r0, [r1]
/* wake up (should have specified args?) */
bl tegra_generic_smc
mov32 r1, TEGRA_TMRUS_BASE
ldr r0, [r1]
adr r1, tegra_resume_smc_exit_time
str r0, [r1]
#endif
#endif
b cpu_resume
ENDPROC(tegra_resume)
#ifdef CONFIG_TRUSTED_FOUNDATIONS
.globl tegra_resume_timestamps_start
#ifndef CONFIG_ARCH_TEGRA_11x_SOC
.globl tegra_resume_smc_entry_time
.globl tegra_resume_smc_exit_time
#endif
.globl tegra_resume_entry_time
.globl tegra_resume_timestamps_end
tegra_resume_timestamps_start:
#ifndef CONFIG_ARCH_TEGRA_11x_SOC
tegra_resume_smc_entry_time:
.long 0
tegra_resume_smc_exit_time:
.long 0
#endif
tegra_resume_entry_time:
.long 0
tegra_resume_timestamps_end:
ENTRY(__tegra_resume_timestamps_end)
#endif
#endif
/*
* __invalidate_cpu_state
*
* Invalidates volatile CPU state (SCU tags, caches, branch address
* arrays, exclusive monitor, etc.) so that they can be safely enabled
* instruction caching and branch predicition enabled
*
* For tegra chips with CONFIG_HAVE_ARM_SCU undefined, it means there is
* an integrated SCU in L2 memory system, this is true for Cortex-A15
* MP processors. In this case, we only need to set the correct L2 cache
* data RAM latency and enable i-cache/branch prediction
*/
__invalidate_cpu_state:
clrex
mov r0, #0
mcr p15, 0, r0, c1, c0, 1 @ disable SMP, prefetch, broadcast
isb
#if defined(CONFIG_HAVE_ARM_SCU)
mcr p15, 0, r0, c7, c5, 0 @ invalidate BTAC, i-cache
mcr p15, 0, r0, c7, c5, 6 @ invalidate branch pred array
mcr p15, 0, r0, c8, c5, 0 @ invalidate instruction TLB
mcr p15, 0, r0, c8, c6, 0 @ invalidate data TLB
mcr p15, 0, r0, c8, c7, 0 @ invalidate unified TLB
dsb
isb
cpu_id r0
cmp r0, #0
mov32 r1, (TEGRA_ARM_PERIF_BASE + 0xC)
movne r0, r0, lsl #2
movne r2, #0xf
movne r2, r2, lsl r0
strne r2, [r1] @ invalidate SCU tags for CPU
dsb
mov r0, #0x1800
mcr p15, 0, r0, c1, c0, 0 @ enable branch prediction, i-cache
isb
/* fall through */
#else
mrc p15, 0x1, r0, c15, c0, 3 @ L2 prefetch control reg
tst r0, #0x1000
orreq r0, r0, #0x1000 @ disable throttling
mcreq p15, 0x1, r0, c15, c0, 3
/* This is only needed for cluster 0 with integrated L2 cache */
mrc p15, 0, r0, c0, c0, 5
ubfx r0, r0, #8, #4
tst r0, #1
bne __enable_i_cache_branch_pred
mrc p15, 0x1, r0, c9, c0, 2
and r1, r0, #7
cmp r1, #2
beq __enable_i_cache_branch_pred
bic r0, r0, #7
orr r0, r0, #2
mcr p15, 0x1, r0, c9, c0, 2
mrc p15, 0x1, r0, c15, c0, 0 @ L2 ACTLR
orr r0, r0, #0x80 @ hazard detection timeout
mcr p15, 0x1, r0, c15, c0, 0
__enable_i_cache_branch_pred:
mov r0, #0x1800
mcr p15, 0, r0, c1, c0, 0 @ enable branch prediction, i-cache
mov pc, lr
/* no fall through, just return to the caller */
#endif
/*
* tegra_invalidate_cache
*
* Invalidates the L1 data cache (no clean) during initial boot of a cpu
*
* Corrupted registers: r0-r6
*/
tegra_invalidate_cache:
mov r0, #0
mcr p15, 2, r0, c0, c0, 0
mrc p15, 1, r0, c0, c0, 0
movw r1, #0x7fff
and r2, r1, r0, lsr #13
movw r1, #0x3ff
and r3, r1, r0, lsr #3 @ NumWays - 1
add r2, r2, #1 @ NumSets
and r0, r0, #0x7
add r0, r0, #4 @ SetShift
clz r1, r3 @ WayShift
add r4, r3, #1 @ NumWays
1: sub r2, r2, #1 @ NumSets--
mov r3, r4 @ Temp = NumWays
2: subs r3, r3, #1 @ Temp--
mov r5, r3, lsl r1
mov r6, r2, lsl r0
orr r5, r5, r6 @ Reg = (Temp<<WayShift)|(NumSets<<SetShift)
mcr p15, 0, r5, c7, c6, 2
bgt 2b
cmp r2, #0
bgt 1b
dsb
isb
mov pc, lr
/*
* __tegra_cpu_reset_handler_halt_failed:
*
* Alternate entry point for reset handler for cases where the
* WFI halt failed to take effect.
*
*/
.align L1_CACHE_SHIFT
ENTRY(__tegra_cpu_reset_handler_start)
/*
* __tegra_cpu_reset_handler:
*
* Common handler for all CPU reset events.
*
* Register usage within the reset handler:
*
* R7 = CPU present (to the OS) mask
* R8 = CPU in LP1 state mask
* R9 = CPU in LP2 state mask
* R10 = CPU number
* R11 = CPU mask
* R12 = pointer to reset handler data
*
* NOTE: This code is copied to IRAM. All code and data accesses
* must be position-independent.
*/
.align L1_CACHE_SHIFT
ENTRY(__tegra_cpu_reset_handler)
/* DO NOT put any code before the !defined(CONFIG_ARM_SAVE_DEBUG_CONTEXT)
block below. It must be the first thing in this subroutine. */
#if !defined(CONFIG_ARM_SAVE_DEBUG_CONTEXT) || DEBUG_CPU_RESET_HANDLER
/* If Debug Architecture v7.1 or later, unlock the OS lock. */
mrc p15, 0, r0, c0, c1, 2 @ ID_DFR0
and r0, r0, #0xF @ coprocessor debug model
cmp r0, #5 @ debug arch >= v7.1?
movge r0, #0 @ yes, unlock debug
mcrge p14, 0, r0, c1, c0, 4 @ DBGOSLAR
#endif
#if DEBUG_CPU_RESET_HANDLER
b .
#endif
#ifndef CONFIG_TRUSTED_FOUNDATIONS
cpsid aif, 0x13 @ SVC mode, interrupts disabled
mrc p15, 0, r0, c0, c0, 0 @ read main ID register
and r5, r0, #0x00f00000 @ variant
and r6, r0, #0x0000000f @ revision
orr r6, r6, r5, lsr #20-4 @ combine variant and revision
#ifdef CONFIG_ARM_ERRATA_743622
teq r6, #0x20 @ present in r2p0
teqne r6, #0x21 @ present in r2p1
teqne r6, #0x22 @ present in r2p2
teqne r6, #0x27 @ present in r2p7
teqne r6, #0x29 @ present in r2p9
mrceq p15, 0, r10, c15, c0, 1 @ read diagnostic register
orreq r10, r10, #1 << 6 @ set bit #6
mcreq p15, 0, r10, c15, c0, 1 @ write diagnostic register
#endif
#endif
mrc p15, 0, r10, c0, c0, 5 @ MPIDR
and r10, r10, #0x3 @ R10 = CPU number
mov r11, #1
mov r11, r11, lsl r10 @ R11 = CPU mask
adr r12, __tegra_cpu_reset_handler_data
#ifdef CONFIG_SMP
/* Does the OS know about this CPU? */
ldr r7, [r12, #RESET_DATA(MASK_PRESENT)]
tst r7, r11 @ if !present
bleq __die @ CPU not present (to OS)
#endif
#ifdef CONFIG_ARCH_TEGRA_2x_SOC
/* If CPU1, don't let CPU0 reset CPU1 now that CPU1 is coming up. */
mov32 r6, TEGRA_PMC_BASE
mov r0, #0
cmp r10, #0
strne r0, [r6, #PMC_SCRATCH41]
#endif
#ifdef CONFIG_PM_SLEEP
/* Waking up from LP1? */
ldr r8, [r12, #RESET_DATA(MASK_LP1)]
tst r8, r11 @ if in_lp1
beq __is_not_lp1
cmp r10, #0
bne __die @ only CPU0 can be here
ldr lr, [r12, #RESET_DATA(STARTUP_LP1)]
cmp lr, #0
bleq __die @ no LP1 startup handler
bx lr
__is_not_lp1:
#endif
/* Waking up from LP2? */
ldr r9, [r12, #RESET_DATA(MASK_LP2)]
tst r9, r11 @ if in_lp2
beq __is_not_lp2
ldr lr, [r12, #RESET_DATA(STARTUP_LP2)]
cmp lr, #0
bleq __die @ no LP2 startup handler
bx lr
__is_not_lp2:
#ifdef CONFIG_SMP
#ifndef CONFIG_TEGRA_VIRTUAL_CPUID
/* Can only be secondary boot (initial or hotplug) but CPU 0
cannot be here. */
cmp r10, #0
bleq __die @ CPU0 cannot be here
#endif
ldr lr, [r12, #RESET_DATA(STARTUP_SECONDARY)]
cmp lr, #0
bleq __die @ no secondary startup handler
bx lr
#endif
/*
* We don't know why the CPU reset. Just kill it.
* The LR register will contain the address we died at + 4.
*/
__die:
sub lr, lr, #4
mov32 r7, TEGRA_PMC_BASE
str lr, [r7, #PMC_SCRATCH41]
mov32 r7, TEGRA_CLK_RESET_BASE
/* Are we on Tegra20? */
mov32 r6, TEGRA_APB_MISC_BASE
ldr r0, [r6, #APB_MISC_GP_HIDREV]
and r0, r0, #0xff00
cmp r0, #(0x20 << 8)
bne 1f
#ifdef CONFIG_ARCH_TEGRA_2x_SOC
mov32 r0, 0x1111
mov r1, r0, lsl r10
str r1, [r7, #0x340] @ CLK_RST_CPU_CMPLX_SET
#endif
1:
#ifdef CONFIG_ARCH_TEGRA_3x_SOC
mov32 r6, TEGRA_FLOW_CTRL_BASE
cmp r10, #0
moveq r1, #FLOW_CTRL_HALT_CPU0_EVENTS
moveq r2, #FLOW_CTRL_CPU0_CSR
movne r1, r10, lsl #3
addne r2, r1, #(FLOW_CTRL_CPU1_CSR-8)
addne r1, r1, #(FLOW_CTRL_HALT_CPU1_EVENTS-8)
/* Clear CPU "event" and "interrupt" flags and power gate
it when halting but not before it is in the "WFI" state. */
ldr r0, [r6, +r2]
orr r0, r0, #FLOW_CTRL_CSR_INTR_FLAG | FLOW_CTRL_CSR_EVENT_FLAG
orr r0, r0, #FLOW_CTRL_CSR_ENABLE
str r0, [r6, +r2]
/* Unconditionally halt this CPU */
mov r0, #FLOW_CTRL_WAITEVENT
str r0, [r6, +r1]
ldr r0, [r6, +r1] @ memory barrier
dsb
isb
wfi @ CPU should be power gated here
/* If the CPU didn't power gate above just kill it's clock. */
mov r0, r11, lsl #8
str r0, [r7, #348] @ CLK_CPU_CMPLX_SET
#endif
/* If the CPU still isn't dead, just spin here. */
b .
ENDPROC(__tegra_cpu_reset_handler)
.align L1_CACHE_SHIFT
.type __tegra_cpu_reset_handler_data, %object
.globl __tegra_cpu_reset_handler_data
__tegra_cpu_reset_handler_data:
.rept TEGRA_RESET_DATA_SIZE
.long 0
.endr
.size __tegra_cpu_reset_handler_data, . - __tegra_cpu_reset_handler_data
.align L1_CACHE_SHIFT
ENTRY(__tegra_cpu_reset_handler_end)
|