Merge tag 'omap-for-v4.8/legacy-signed' of git://git.kernel.org/pub/scm/linux/kernel...
[deliverable/linux.git] / arch / arm / mach-omap2 / sleep34xx.S
1 /*
2 * (C) Copyright 2007
3 * Texas Instruments
4 * Karthik Dasu <karthik-dp@ti.com>
5 *
6 * (C) Copyright 2004
7 * Texas Instruments, <www.ti.com>
8 * Richard Woodruff <r-woodruff2@ti.com>
9 *
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License as
12 * published by the Free Software Foundation; either version 2 of
13 * the License, or (at your option) any later version.
14 *
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR /PURPOSE. See the
18 * GNU General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston,
23 * MA 02111-1307 USA
24 */
25 #include <linux/linkage.h>
26
27 #include <asm/assembler.h>
28
29 #include "omap34xx.h"
30 #include "iomap.h"
31 #include "cm3xxx.h"
32 #include "prm3xxx.h"
33 #include "sdrc.h"
34 #include "sram.h"
35 #include "control.h"
36
37 /*
38 * Registers access definitions
39 */
40 #define SDRC_SCRATCHPAD_SEM_OFFS 0xc
41 #define SDRC_SCRATCHPAD_SEM_V OMAP343X_SCRATCHPAD_REGADDR\
42 (SDRC_SCRATCHPAD_SEM_OFFS)
43 #define PM_PREPWSTST_CORE_P OMAP3430_PRM_BASE + CORE_MOD +\
44 OMAP3430_PM_PREPWSTST
45 #define PM_PWSTCTRL_MPU_P OMAP3430_PRM_BASE + MPU_MOD + OMAP2_PM_PWSTCTRL
46 #define CM_IDLEST1_CORE_V OMAP34XX_CM_REGADDR(CORE_MOD, CM_IDLEST1)
47 #define CM_IDLEST_CKGEN_V OMAP34XX_CM_REGADDR(PLL_MOD, CM_IDLEST)
48 #define SRAM_BASE_P OMAP3_SRAM_PA
49 #define CONTROL_STAT OMAP343X_CTRL_BASE + OMAP343X_CONTROL_STATUS
50 #define CONTROL_MEM_RTA_CTRL (OMAP343X_CTRL_BASE +\
51 OMAP36XX_CONTROL_MEM_RTA_CTRL)
52
53 /* Move this as correct place is available */
54 #define SCRATCHPAD_MEM_OFFS 0x310
55 #define SCRATCHPAD_BASE_P (OMAP343X_CTRL_BASE +\
56 OMAP343X_CONTROL_MEM_WKUP +\
57 SCRATCHPAD_MEM_OFFS)
58 #define SDRC_POWER_V OMAP34XX_SDRC_REGADDR(SDRC_POWER)
59 #define SDRC_SYSCONFIG_P (OMAP343X_SDRC_BASE + SDRC_SYSCONFIG)
60 #define SDRC_MR_0_P (OMAP343X_SDRC_BASE + SDRC_MR_0)
61 #define SDRC_EMR2_0_P (OMAP343X_SDRC_BASE + SDRC_EMR2_0)
62 #define SDRC_MANUAL_0_P (OMAP343X_SDRC_BASE + SDRC_MANUAL_0)
63 #define SDRC_MR_1_P (OMAP343X_SDRC_BASE + SDRC_MR_1)
64 #define SDRC_EMR2_1_P (OMAP343X_SDRC_BASE + SDRC_EMR2_1)
65 #define SDRC_MANUAL_1_P (OMAP343X_SDRC_BASE + SDRC_MANUAL_1)
66 #define SDRC_DLLA_STATUS_V OMAP34XX_SDRC_REGADDR(SDRC_DLLA_STATUS)
67 #define SDRC_DLLA_CTRL_V OMAP34XX_SDRC_REGADDR(SDRC_DLLA_CTRL)
68
69 /*
70 * This file needs be built unconditionally as ARM to interoperate correctly
71 * with non-Thumb-2-capable firmware.
72 */
73 .arm
74
75 /*
76 * API functions
77 */
78
79 .text
80 /*
81 * L2 cache needs to be toggled for stable OFF mode functionality on 3630.
82 * This function sets up a flag that will allow for this toggling to take
83 * place on 3630. Hopefully some version in the future may not need this.
84 */
85 ENTRY(enable_omap3630_toggle_l2_on_restore)
86 stmfd sp!, {lr} @ save registers on stack
87 /* Setup so that we will disable and enable l2 */
88 mov r1, #0x1
89 adrl r3, l2dis_3630_offset @ may be too distant for plain adr
90 ldr r2, [r3] @ value for offset
91 str r1, [r2, r3] @ write to l2dis_3630
92 ldmfd sp!, {pc} @ restore regs and return
93 ENDPROC(enable_omap3630_toggle_l2_on_restore)
94
95 /*
96 * Function to call rom code to save secure ram context. This gets
97 * relocated to SRAM, so it can be all in .data section. Otherwise
98 * we need to initialize api_params separately.
99 */
100 .data
101 .align 3
102 ENTRY(save_secure_ram_context)
103 stmfd sp!, {r4 - r11, lr} @ save registers on stack
104 adr r3, api_params @ r3 points to parameters
105 str r0, [r3,#0x4] @ r0 has sdram address
106 ldr r12, high_mask
107 and r3, r3, r12
108 ldr r12, sram_phy_addr_mask
109 orr r3, r3, r12
110 mov r0, #25 @ set service ID for PPA
111 mov r12, r0 @ copy secure service ID in r12
112 mov r1, #0 @ set task id for ROM code in r1
113 mov r2, #4 @ set some flags in r2, r6
114 mov r6, #0xff
115 dsb @ data write barrier
116 dmb @ data memory barrier
117 smc #1 @ call SMI monitor (smi #1)
118 nop
119 nop
120 nop
121 nop
122 ldmfd sp!, {r4 - r11, pc}
123 .align
124 sram_phy_addr_mask:
125 .word SRAM_BASE_P
126 high_mask:
127 .word 0xffff
128 api_params:
129 .word 0x4, 0x0, 0x0, 0x1, 0x1
130 ENDPROC(save_secure_ram_context)
131 ENTRY(save_secure_ram_context_sz)
132 .word . - save_secure_ram_context
133
134 .text
135
136 /*
137 * ======================
138 * == Idle entry point ==
139 * ======================
140 */
141
142 /*
143 * Forces OMAP into idle state
144 *
145 * omap34xx_cpu_suspend() - This bit of code saves the CPU context if needed
146 * and executes the WFI instruction. Calling WFI effectively changes the
147 * power domains states to the desired target power states.
148 *
149 *
150 * Notes:
151 * - only the minimum set of functions gets copied to internal SRAM at boot
152 * and after wake-up from OFF mode, cf. omap_push_sram_idle. The function
153 * pointers in SDRAM or SRAM are called depending on the desired low power
154 * target state.
155 * - when the OMAP wakes up it continues at different execution points
156 * depending on the low power mode (non-OFF vs OFF modes),
157 * cf. 'Resume path for xxx mode' comments.
158 */
159 .align 3
160 ENTRY(omap34xx_cpu_suspend)
161 stmfd sp!, {r4 - r11, lr} @ save registers on stack
162
163 /*
164 * r0 contains information about saving context:
165 * 0 - No context lost
166 * 1 - Only L1 and logic lost
167 * 2 - Only L2 lost (Even L1 is retained we clean it along with L2)
168 * 3 - Both L1 and L2 lost and logic lost
169 */
170
171 /*
172 * For OFF mode: save context and jump to WFI in SDRAM (omap3_do_wfi)
173 * For non-OFF modes: jump to the WFI code in SRAM (omap3_do_wfi_sram)
174 */
175 ldr r4, omap3_do_wfi_sram_addr
176 ldr r5, [r4]
177 cmp r0, #0x0 @ If no context save required,
178 bxeq r5 @ jump to the WFI code in SRAM
179
180
181 /* Otherwise fall through to the save context code */
182 save_context_wfi:
183 /*
184 * jump out to kernel flush routine
185 * - reuse that code is better
186 * - it executes in a cached space so is faster than refetch per-block
187 * - should be faster and will change with kernel
188 * - 'might' have to copy address, load and jump to it
189 * Flush all data from the L1 data cache before disabling
190 * SCTLR.C bit.
191 */
192 ldr r1, kernel_flush
193 mov lr, pc
194 bx r1
195
196 /*
197 * Clear the SCTLR.C bit to prevent further data cache
198 * allocation. Clearing SCTLR.C would make all the data accesses
199 * strongly ordered and would not hit the cache.
200 */
201 mrc p15, 0, r0, c1, c0, 0
202 bic r0, r0, #(1 << 2) @ Disable the C bit
203 mcr p15, 0, r0, c1, c0, 0
204 isb
205
206 /*
207 * Invalidate L1 data cache. Even though only invalidate is
208 * necessary exported flush API is used here. Doing clean
209 * on already clean cache would be almost NOP.
210 */
211 ldr r1, kernel_flush
212 blx r1
213 b omap3_do_wfi
214 ENDPROC(omap34xx_cpu_suspend)
215 omap3_do_wfi_sram_addr:
216 .word omap3_do_wfi_sram
217 kernel_flush:
218 .word v7_flush_dcache_all
219
220 /* ===================================
221 * == WFI instruction => Enter idle ==
222 * ===================================
223 */
224
225 /*
226 * Do WFI instruction
227 * Includes the resume path for non-OFF modes
228 *
229 * This code gets copied to internal SRAM and is accessible
230 * from both SDRAM and SRAM:
231 * - executed from SRAM for non-off modes (omap3_do_wfi_sram),
232 * - executed from SDRAM for OFF mode (omap3_do_wfi).
233 */
234 .align 3
235 ENTRY(omap3_do_wfi)
236 ldr r4, sdrc_power @ read the SDRC_POWER register
237 ldr r5, [r4] @ read the contents of SDRC_POWER
238 orr r5, r5, #0x40 @ enable self refresh on idle req
239 str r5, [r4] @ write back to SDRC_POWER register
240
241 /* Data memory barrier and Data sync barrier */
242 dsb
243 dmb
244
245 /*
246 * ===================================
247 * == WFI instruction => Enter idle ==
248 * ===================================
249 */
250 wfi @ wait for interrupt
251
252 /*
253 * ===================================
254 * == Resume path for non-OFF modes ==
255 * ===================================
256 */
257 nop
258 nop
259 nop
260 nop
261 nop
262 nop
263 nop
264 nop
265 nop
266 nop
267
268 /*
269 * This function implements the erratum ID i581 WA:
270 * SDRC state restore before accessing the SDRAM
271 *
272 * Only used at return from non-OFF mode. For OFF
273 * mode the ROM code configures the SDRC and
274 * the DPLL before calling the restore code directly
275 * from DDR.
276 */
277
278 /* Make sure SDRC accesses are ok */
279 wait_sdrc_ok:
280
281 /* DPLL3 must be locked before accessing the SDRC. Maybe the HW ensures this */
282 ldr r4, cm_idlest_ckgen
283 wait_dpll3_lock:
284 ldr r5, [r4]
285 tst r5, #1
286 beq wait_dpll3_lock
287
288 ldr r4, cm_idlest1_core
289 wait_sdrc_ready:
290 ldr r5, [r4]
291 tst r5, #0x2
292 bne wait_sdrc_ready
293 /* allow DLL powerdown upon hw idle req */
294 ldr r4, sdrc_power
295 ldr r5, [r4]
296 bic r5, r5, #0x40
297 str r5, [r4]
298
299 is_dll_in_lock_mode:
300 /* Is dll in lock mode? */
301 ldr r4, sdrc_dlla_ctrl
302 ldr r5, [r4]
303 tst r5, #0x4
304 bne exit_nonoff_modes @ Return if locked
305 /* wait till dll locks */
306 wait_dll_lock_timed:
307 ldr r4, sdrc_dlla_status
308 /* Wait 20uS for lock */
309 mov r6, #8
310 wait_dll_lock:
311 subs r6, r6, #0x1
312 beq kick_dll
313 ldr r5, [r4]
314 and r5, r5, #0x4
315 cmp r5, #0x4
316 bne wait_dll_lock
317 b exit_nonoff_modes @ Return when locked
318
319 /* disable/reenable DLL if not locked */
320 kick_dll:
321 ldr r4, sdrc_dlla_ctrl
322 ldr r5, [r4]
323 mov r6, r5
324 bic r6, #(1<<3) @ disable dll
325 str r6, [r4]
326 dsb
327 orr r6, r6, #(1<<3) @ enable dll
328 str r6, [r4]
329 dsb
330 b wait_dll_lock_timed
331
332 exit_nonoff_modes:
333 /* Re-enable C-bit if needed */
334 mrc p15, 0, r0, c1, c0, 0
335 tst r0, #(1 << 2) @ Check C bit enabled?
336 orreq r0, r0, #(1 << 2) @ Enable the C bit if cleared
337 mcreq p15, 0, r0, c1, c0, 0
338 isb
339
340 /*
341 * ===================================
342 * == Exit point from non-OFF modes ==
343 * ===================================
344 */
345 ldmfd sp!, {r4 - r11, pc} @ restore regs and return
346 ENDPROC(omap3_do_wfi)
347 sdrc_power:
348 .word SDRC_POWER_V
349 cm_idlest1_core:
350 .word CM_IDLEST1_CORE_V
351 cm_idlest_ckgen:
352 .word CM_IDLEST_CKGEN_V
353 sdrc_dlla_status:
354 .word SDRC_DLLA_STATUS_V
355 sdrc_dlla_ctrl:
356 .word SDRC_DLLA_CTRL_V
357 ENTRY(omap3_do_wfi_sz)
358 .word . - omap3_do_wfi
359
360
361 /*
362 * ==============================
363 * == Resume path for OFF mode ==
364 * ==============================
365 */
366
367 /*
368 * The restore_* functions are called by the ROM code
369 * when back from WFI in OFF mode.
370 * Cf. the get_*restore_pointer functions.
371 *
372 * restore_es3: applies to 34xx >= ES3.0
373 * restore_3630: applies to 36xx
374 * restore: common code for 3xxx
375 *
376 * Note: when back from CORE and MPU OFF mode we are running
377 * from SDRAM, without MMU, without the caches and prediction.
378 * Also the SRAM content has been cleared.
379 */
380 ENTRY(omap3_restore_es3)
381 ldr r5, pm_prepwstst_core_p
382 ldr r4, [r5]
383 and r4, r4, #0x3
384 cmp r4, #0x0 @ Check if previous power state of CORE is OFF
385 bne omap3_restore @ Fall through to OMAP3 common code
386 adr r0, es3_sdrc_fix
387 ldr r1, sram_base
388 ldr r2, es3_sdrc_fix_sz
389 mov r2, r2, ror #2
390 copy_to_sram:
391 ldmia r0!, {r3} @ val = *src
392 stmia r1!, {r3} @ *dst = val
393 subs r2, r2, #0x1 @ num_words--
394 bne copy_to_sram
395 ldr r1, sram_base
396 blx r1
397 b omap3_restore @ Fall through to OMAP3 common code
398 ENDPROC(omap3_restore_es3)
399
400 ENTRY(omap3_restore_3630)
401 ldr r1, pm_prepwstst_core_p
402 ldr r2, [r1]
403 and r2, r2, #0x3
404 cmp r2, #0x0 @ Check if previous power state of CORE is OFF
405 bne omap3_restore @ Fall through to OMAP3 common code
406 /* Disable RTA before giving control */
407 ldr r1, control_mem_rta
408 mov r2, #OMAP36XX_RTA_DISABLE
409 str r2, [r1]
410 ENDPROC(omap3_restore_3630)
411
412 /* Fall through to common code for the remaining logic */
413
414 ENTRY(omap3_restore)
415 /*
416 * Read the pwstctrl register to check the reason for mpu reset.
417 * This tells us what was lost.
418 */
419 ldr r1, pm_pwstctrl_mpu
420 ldr r2, [r1]
421 and r2, r2, #0x3
422 cmp r2, #0x0 @ Check if target power state was OFF or RET
423 bne logic_l1_restore
424
425 adr r1, l2dis_3630_offset @ address for offset
426 ldr r0, [r1] @ value for offset
427 ldr r0, [r1, r0] @ value at l2dis_3630
428 cmp r0, #0x1 @ should we disable L2 on 3630?
429 bne skipl2dis
430 mrc p15, 0, r0, c1, c0, 1
431 bic r0, r0, #2 @ disable L2 cache
432 mcr p15, 0, r0, c1, c0, 1
433 skipl2dis:
434 ldr r0, control_stat
435 ldr r1, [r0]
436 and r1, #0x700
437 cmp r1, #0x300
438 beq l2_inv_gp
439 adr r0, l2_inv_api_params_offset
440 ldr r3, [r0]
441 add r3, r3, r0 @ r3 points to dummy parameters
442 mov r0, #40 @ set service ID for PPA
443 mov r12, r0 @ copy secure Service ID in r12
444 mov r1, #0 @ set task id for ROM code in r1
445 mov r2, #4 @ set some flags in r2, r6
446 mov r6, #0xff
447 dsb @ data write barrier
448 dmb @ data memory barrier
449 smc #1 @ call SMI monitor (smi #1)
450 /* Write to Aux control register to set some bits */
451 mov r0, #42 @ set service ID for PPA
452 mov r12, r0 @ copy secure Service ID in r12
453 mov r1, #0 @ set task id for ROM code in r1
454 mov r2, #4 @ set some flags in r2, r6
455 mov r6, #0xff
456 ldr r4, scratchpad_base
457 ldr r3, [r4, #0xBC] @ r3 points to parameters
458 dsb @ data write barrier
459 dmb @ data memory barrier
460 smc #1 @ call SMI monitor (smi #1)
461
462 #ifdef CONFIG_OMAP3_L2_AUX_SECURE_SAVE_RESTORE
463 /* Restore L2 aux control register */
464 @ set service ID for PPA
465 mov r0, #CONFIG_OMAP3_L2_AUX_SECURE_SERVICE_SET_ID
466 mov r12, r0 @ copy service ID in r12
467 mov r1, #0 @ set task ID for ROM code in r1
468 mov r2, #4 @ set some flags in r2, r6
469 mov r6, #0xff
470 ldr r4, scratchpad_base
471 ldr r3, [r4, #0xBC]
472 adds r3, r3, #8 @ r3 points to parameters
473 dsb @ data write barrier
474 dmb @ data memory barrier
475 smc #1 @ call SMI monitor (smi #1)
476 #endif
477 b logic_l1_restore
478
479 .align
480 l2_inv_api_params_offset:
481 .long l2_inv_api_params - .
482 l2_inv_gp:
483 /* Execute smi to invalidate L2 cache */
484 mov r12, #0x1 @ set up to invalidate L2
485 smc #0 @ Call SMI monitor (smieq)
486 /* Write to Aux control register to set some bits */
487 ldr r4, scratchpad_base
488 ldr r3, [r4,#0xBC]
489 ldr r0, [r3,#4]
490 mov r12, #0x3
491 smc #0 @ Call SMI monitor (smieq)
492 ldr r4, scratchpad_base
493 ldr r3, [r4,#0xBC]
494 ldr r0, [r3,#12]
495 mov r12, #0x2
496 smc #0 @ Call SMI monitor (smieq)
497 logic_l1_restore:
498 adr r0, l2dis_3630_offset @ adress for offset
499 ldr r1, [r0] @ value for offset
500 ldr r1, [r0, r1] @ value at l2dis_3630
501 cmp r1, #0x1 @ Test if L2 re-enable needed on 3630
502 bne skipl2reen
503 mrc p15, 0, r1, c1, c0, 1
504 orr r1, r1, #2 @ re-enable L2 cache
505 mcr p15, 0, r1, c1, c0, 1
506 skipl2reen:
507
508 /* Now branch to the common CPU resume function */
509 b cpu_resume
510 ENDPROC(omap3_restore)
511
512 .ltorg
513
514 /*
515 * Local variables
516 */
517 pm_prepwstst_core_p:
518 .word PM_PREPWSTST_CORE_P
519 pm_pwstctrl_mpu:
520 .word PM_PWSTCTRL_MPU_P
521 scratchpad_base:
522 .word SCRATCHPAD_BASE_P
523 sram_base:
524 .word SRAM_BASE_P + 0x8000
525 control_stat:
526 .word CONTROL_STAT
527 control_mem_rta:
528 .word CONTROL_MEM_RTA_CTRL
529 l2dis_3630_offset:
530 .long l2dis_3630 - .
531
532 .data
533 l2dis_3630:
534 .word 0
535
536 .data
537 l2_inv_api_params:
538 .word 0x1, 0x00
539
540 /*
541 * Internal functions
542 */
543
544 /*
545 * This function implements the erratum ID i443 WA, applies to 34xx >= ES3.0
546 * Copied to and run from SRAM in order to reconfigure the SDRC parameters.
547 */
548 .text
549 .align 3
550 ENTRY(es3_sdrc_fix)
551 ldr r4, sdrc_syscfg @ get config addr
552 ldr r5, [r4] @ get value
553 tst r5, #0x100 @ is part access blocked
554 it eq
555 biceq r5, r5, #0x100 @ clear bit if set
556 str r5, [r4] @ write back change
557 ldr r4, sdrc_mr_0 @ get config addr
558 ldr r5, [r4] @ get value
559 str r5, [r4] @ write back change
560 ldr r4, sdrc_emr2_0 @ get config addr
561 ldr r5, [r4] @ get value
562 str r5, [r4] @ write back change
563 ldr r4, sdrc_manual_0 @ get config addr
564 mov r5, #0x2 @ autorefresh command
565 str r5, [r4] @ kick off refreshes
566 ldr r4, sdrc_mr_1 @ get config addr
567 ldr r5, [r4] @ get value
568 str r5, [r4] @ write back change
569 ldr r4, sdrc_emr2_1 @ get config addr
570 ldr r5, [r4] @ get value
571 str r5, [r4] @ write back change
572 ldr r4, sdrc_manual_1 @ get config addr
573 mov r5, #0x2 @ autorefresh command
574 str r5, [r4] @ kick off refreshes
575 bx lr
576
577 /*
578 * Local variables
579 */
580 .align
581 sdrc_syscfg:
582 .word SDRC_SYSCONFIG_P
583 sdrc_mr_0:
584 .word SDRC_MR_0_P
585 sdrc_emr2_0:
586 .word SDRC_EMR2_0_P
587 sdrc_manual_0:
588 .word SDRC_MANUAL_0_P
589 sdrc_mr_1:
590 .word SDRC_MR_1_P
591 sdrc_emr2_1:
592 .word SDRC_EMR2_1_P
593 sdrc_manual_1:
594 .word SDRC_MANUAL_1_P
595 ENDPROC(es3_sdrc_fix)
596 ENTRY(es3_sdrc_fix_sz)
597 .word . - es3_sdrc_fix
This page took 0.045199 seconds and 5 git commands to generate.