]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blob - arch/parisc/kernel/entry.S
Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/scjody...
[mirror_ubuntu-zesty-kernel.git] / arch / parisc / kernel / entry.S
1 /*
2 * Linux/PA-RISC Project (http://www.parisc-linux.org/)
3 *
4 * kernel entry points (interruptions, system call wrappers)
5 * Copyright (C) 1999,2000 Philipp Rumpf
6 * Copyright (C) 1999 SuSE GmbH Nuernberg
7 * Copyright (C) 2000 Hewlett-Packard (John Marvin)
8 * Copyright (C) 1999 Hewlett-Packard (Frank Rowand)
9 *
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License as published by
12 * the Free Software Foundation; either version 2, or (at your option)
13 * any later version.
14 *
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 * GNU General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
23 */
24
25 #include <linux/config.h>
26 #include <asm/asm-offsets.h>
27
28 /* we have the following possibilities to act on an interruption:
29 * - handle in assembly and use shadowed registers only
30 * - save registers to kernel stack and handle in assembly or C */
31
32
33 #include <asm/psw.h>
34 #include <asm/assembly.h> /* for LDREG/STREG defines */
35 #include <asm/pgtable.h>
36 #include <asm/signal.h>
37 #include <asm/unistd.h>
38 #include <asm/thread_info.h>
39
40 #ifdef CONFIG_64BIT
41 #define CMPIB cmpib,*
42 #define CMPB cmpb,*
43 #define COND(x) *x
44
45 .level 2.0w
46 #else
47 #define CMPIB cmpib,
48 #define CMPB cmpb,
49 #define COND(x) x
50
51 .level 2.0
52 #endif
53
54 .import pa_dbit_lock,data
55
56 /* space_to_prot macro creates a prot id from a space id */
57
58 #if (SPACEID_SHIFT) == 0
59 .macro space_to_prot spc prot
60 depd,z \spc,62,31,\prot
61 .endm
62 #else
63 .macro space_to_prot spc prot
64 extrd,u \spc,(64 - (SPACEID_SHIFT)),32,\prot
65 .endm
66 #endif
67
68 /* Switch to virtual mapping, trashing only %r1 */
69 .macro virt_map
70 /* pcxt_ssm_bug */
71 rsm PSW_SM_I, %r0 /* barrier for "Relied upon Translation */
72 mtsp %r0, %sr4
73 mtsp %r0, %sr5
74 mfsp %sr7, %r1
75 or,= %r0,%r1,%r0 /* Only save sr7 in sr3 if sr7 != 0 */
76 mtsp %r1, %sr3
77 tovirt_r1 %r29
78 load32 KERNEL_PSW, %r1
79
80 rsm PSW_SM_QUIET,%r0 /* second "heavy weight" ctl op */
81 mtsp %r0, %sr6
82 mtsp %r0, %sr7
83 mtctl %r0, %cr17 /* Clear IIASQ tail */
84 mtctl %r0, %cr17 /* Clear IIASQ head */
85 mtctl %r1, %ipsw
86 load32 4f, %r1
87 mtctl %r1, %cr18 /* Set IIAOQ tail */
88 ldo 4(%r1), %r1
89 mtctl %r1, %cr18 /* Set IIAOQ head */
90 rfir
91 nop
92 4:
93 .endm
94
95 /*
96 * The "get_stack" macros are responsible for determining the
97 * kernel stack value.
98 *
99 * For Faults:
100 * If sr7 == 0
101 * Already using a kernel stack, so call the
102 * get_stack_use_r30 macro to push a pt_regs structure
103 * on the stack, and store registers there.
104 * else
105 * Need to set up a kernel stack, so call the
106 * get_stack_use_cr30 macro to set up a pointer
107 * to the pt_regs structure contained within the
108 * task pointer pointed to by cr30. Set the stack
109 * pointer to point to the end of the task structure.
110 *
111 * For Interrupts:
112 * If sr7 == 0
113 * Already using a kernel stack, check to see if r30
114 * is already pointing to the per processor interrupt
115 * stack. If it is, call the get_stack_use_r30 macro
116 * to push a pt_regs structure on the stack, and store
117 * registers there. Otherwise, call get_stack_use_cr31
118 * to get a pointer to the base of the interrupt stack
119 * and push a pt_regs structure on that stack.
120 * else
121 * Need to set up a kernel stack, so call the
122 * get_stack_use_cr30 macro to set up a pointer
123 * to the pt_regs structure contained within the
124 * task pointer pointed to by cr30. Set the stack
125 * pointer to point to the end of the task structure.
126 * N.B: We don't use the interrupt stack for the
127 * first interrupt from userland, because signals/
128 * resched's are processed when returning to userland,
129 * and we can sleep in those cases.
130 *
131 * Note that we use shadowed registers for temps until
132 * we can save %r26 and %r29. %r26 is used to preserve
133 * %r8 (a shadowed register) which temporarily contained
134 * either the fault type ("code") or the eirr. We need
135 * to use a non-shadowed register to carry the value over
136 * the rfir in virt_map. We use %r26 since this value winds
137 * up being passed as the argument to either do_cpu_irq_mask
138 * or handle_interruption. %r29 is used to hold a pointer
139 * the register save area, and once again, it needs to
140 * be a non-shadowed register so that it survives the rfir.
141 *
142 * N.B. TASK_SZ_ALGN and PT_SZ_ALGN include space for a stack frame.
143 */
144
145 .macro get_stack_use_cr30
146
147 /* we save the registers in the task struct */
148
149 mfctl %cr30, %r1
150 tophys %r1,%r9
151 LDREG TI_TASK(%r9), %r1 /* thread_info -> task_struct */
152 tophys %r1,%r9
153 ldo TASK_REGS(%r9),%r9
154 STREG %r30, PT_GR30(%r9)
155 STREG %r29,PT_GR29(%r9)
156 STREG %r26,PT_GR26(%r9)
157 copy %r9,%r29
158 mfctl %cr30, %r1
159 ldo THREAD_SZ_ALGN(%r1), %r30
160 .endm
161
162 .macro get_stack_use_r30
163
164 /* we put a struct pt_regs on the stack and save the registers there */
165
166 tophys %r30,%r9
167 STREG %r30,PT_GR30(%r9)
168 ldo PT_SZ_ALGN(%r30),%r30
169 STREG %r29,PT_GR29(%r9)
170 STREG %r26,PT_GR26(%r9)
171 copy %r9,%r29
172 .endm
173
174 .macro rest_stack
175 LDREG PT_GR1(%r29), %r1
176 LDREG PT_GR30(%r29),%r30
177 LDREG PT_GR29(%r29),%r29
178 .endm
179
180 /* default interruption handler
181 * (calls traps.c:handle_interruption) */
182 .macro def code
183 b intr_save
184 ldi \code, %r8
185 .align 32
186 .endm
187
188 /* Interrupt interruption handler
189 * (calls irq.c:do_cpu_irq_mask) */
190 .macro extint code
191 b intr_extint
192 mfsp %sr7,%r16
193 .align 32
194 .endm
195
196 .import os_hpmc, code
197
198 /* HPMC handler */
199 .macro hpmc code
200 nop /* must be a NOP, will be patched later */
201 load32 PA(os_hpmc), %r3
202 bv,n 0(%r3)
203 nop
204 .word 0 /* checksum (will be patched) */
205 .word PA(os_hpmc) /* address of handler */
206 .word 0 /* length of handler */
207 .endm
208
209 /*
210 * Performance Note: Instructions will be moved up into
211 * this part of the code later on, once we are sure
212 * that the tlb miss handlers are close to final form.
213 */
214
215 /* Register definitions for tlb miss handler macros */
216
217 va = r8 /* virtual address for which the trap occured */
218 spc = r24 /* space for which the trap occured */
219
220 #ifndef CONFIG_64BIT
221
222 /*
223 * itlb miss interruption handler (parisc 1.1 - 32 bit)
224 */
225
226 .macro itlb_11 code
227
228 mfctl %pcsq, spc
229 b itlb_miss_11
230 mfctl %pcoq, va
231
232 .align 32
233 .endm
234 #endif
235
236 /*
237 * itlb miss interruption handler (parisc 2.0)
238 */
239
240 .macro itlb_20 code
241 mfctl %pcsq, spc
242 #ifdef CONFIG_64BIT
243 b itlb_miss_20w
244 #else
245 b itlb_miss_20
246 #endif
247 mfctl %pcoq, va
248
249 .align 32
250 .endm
251
252 #ifndef CONFIG_64BIT
253 /*
254 * naitlb miss interruption handler (parisc 1.1 - 32 bit)
255 *
256 * Note: naitlb misses will be treated
257 * as an ordinary itlb miss for now.
258 * However, note that naitlb misses
259 * have the faulting address in the
260 * IOR/ISR.
261 */
262
263 .macro naitlb_11 code
264
265 mfctl %isr,spc
266 b itlb_miss_11
267 mfctl %ior,va
268 /* FIXME: If user causes a naitlb miss, the priv level may not be in
269 * lower bits of va, where the itlb miss handler is expecting them
270 */
271
272 .align 32
273 .endm
274 #endif
275
276 /*
277 * naitlb miss interruption handler (parisc 2.0)
278 *
279 * Note: naitlb misses will be treated
280 * as an ordinary itlb miss for now.
281 * However, note that naitlb misses
282 * have the faulting address in the
283 * IOR/ISR.
284 */
285
286 .macro naitlb_20 code
287
288 mfctl %isr,spc
289 #ifdef CONFIG_64BIT
290 b itlb_miss_20w
291 #else
292 b itlb_miss_20
293 #endif
294 mfctl %ior,va
295 /* FIXME: If user causes a naitlb miss, the priv level may not be in
296 * lower bits of va, where the itlb miss handler is expecting them
297 */
298
299 .align 32
300 .endm
301
302 #ifndef CONFIG_64BIT
303 /*
304 * dtlb miss interruption handler (parisc 1.1 - 32 bit)
305 */
306
307 .macro dtlb_11 code
308
309 mfctl %isr, spc
310 b dtlb_miss_11
311 mfctl %ior, va
312
313 .align 32
314 .endm
315 #endif
316
317 /*
318 * dtlb miss interruption handler (parisc 2.0)
319 */
320
321 .macro dtlb_20 code
322
323 mfctl %isr, spc
324 #ifdef CONFIG_64BIT
325 b dtlb_miss_20w
326 #else
327 b dtlb_miss_20
328 #endif
329 mfctl %ior, va
330
331 .align 32
332 .endm
333
334 #ifndef CONFIG_64BIT
335 /* nadtlb miss interruption handler (parisc 1.1 - 32 bit) */
336
337 .macro nadtlb_11 code
338
339 mfctl %isr,spc
340 b nadtlb_miss_11
341 mfctl %ior,va
342
343 .align 32
344 .endm
345 #endif
346
347 /* nadtlb miss interruption handler (parisc 2.0) */
348
349 .macro nadtlb_20 code
350
351 mfctl %isr,spc
352 #ifdef CONFIG_64BIT
353 b nadtlb_miss_20w
354 #else
355 b nadtlb_miss_20
356 #endif
357 mfctl %ior,va
358
359 .align 32
360 .endm
361
362 #ifndef CONFIG_64BIT
363 /*
364 * dirty bit trap interruption handler (parisc 1.1 - 32 bit)
365 */
366
367 .macro dbit_11 code
368
369 mfctl %isr,spc
370 b dbit_trap_11
371 mfctl %ior,va
372
373 .align 32
374 .endm
375 #endif
376
377 /*
378 * dirty bit trap interruption handler (parisc 2.0)
379 */
380
381 .macro dbit_20 code
382
383 mfctl %isr,spc
384 #ifdef CONFIG_64BIT
385 b dbit_trap_20w
386 #else
387 b dbit_trap_20
388 #endif
389 mfctl %ior,va
390
391 .align 32
392 .endm
393
394 /* The following are simple 32 vs 64 bit instruction
395 * abstractions for the macros */
396 .macro EXTR reg1,start,length,reg2
397 #ifdef CONFIG_64BIT
398 extrd,u \reg1,32+\start,\length,\reg2
399 #else
400 extrw,u \reg1,\start,\length,\reg2
401 #endif
402 .endm
403
404 .macro DEP reg1,start,length,reg2
405 #ifdef CONFIG_64BIT
406 depd \reg1,32+\start,\length,\reg2
407 #else
408 depw \reg1,\start,\length,\reg2
409 #endif
410 .endm
411
412 .macro DEPI val,start,length,reg
413 #ifdef CONFIG_64BIT
414 depdi \val,32+\start,\length,\reg
415 #else
416 depwi \val,\start,\length,\reg
417 #endif
418 .endm
419
420 /* In LP64, the space contains part of the upper 32 bits of the
421 * fault. We have to extract this and place it in the va,
422 * zeroing the corresponding bits in the space register */
423 .macro space_adjust spc,va,tmp
424 #ifdef CONFIG_64BIT
425 extrd,u \spc,63,SPACEID_SHIFT,\tmp
426 depd %r0,63,SPACEID_SHIFT,\spc
427 depd \tmp,31,SPACEID_SHIFT,\va
428 #endif
429 .endm
430
431 .import swapper_pg_dir,code
432
433 /* Get the pgd. For faults on space zero (kernel space), this
434 * is simply swapper_pg_dir. For user space faults, the
435 * pgd is stored in %cr25 */
436 .macro get_pgd spc,reg
437 ldil L%PA(swapper_pg_dir),\reg
438 ldo R%PA(swapper_pg_dir)(\reg),\reg
439 or,COND(=) %r0,\spc,%r0
440 mfctl %cr25,\reg
441 .endm
442
443 /*
444 space_check(spc,tmp,fault)
445
446 spc - The space we saw the fault with.
447 tmp - The place to store the current space.
448 fault - Function to call on failure.
449
450 Only allow faults on different spaces from the
451 currently active one if we're the kernel
452
453 */
454 .macro space_check spc,tmp,fault
455 mfsp %sr7,\tmp
456 or,COND(<>) %r0,\spc,%r0 /* user may execute gateway page
457 * as kernel, so defeat the space
458 * check if it is */
459 copy \spc,\tmp
460 or,COND(=) %r0,\tmp,%r0 /* nullify if executing as kernel */
461 cmpb,COND(<>),n \tmp,\spc,\fault
462 .endm
463
464 /* Look up a PTE in a 2-Level scheme (faulting at each
465 * level if the entry isn't present
466 *
467 * NOTE: we use ldw even for LP64, since the short pointers
468 * can address up to 1TB
469 */
470 .macro L2_ptep pmd,pte,index,va,fault
471 #if PT_NLEVELS == 3
472 EXTR \va,31-ASM_PMD_SHIFT,ASM_BITS_PER_PMD,\index
473 #else
474 EXTR \va,31-ASM_PGDIR_SHIFT,ASM_BITS_PER_PGD,\index
475 #endif
476 DEP %r0,31,PAGE_SHIFT,\pmd /* clear offset */
477 copy %r0,\pte
478 ldw,s \index(\pmd),\pmd
479 bb,>=,n \pmd,_PxD_PRESENT_BIT,\fault
480 DEP %r0,31,PxD_FLAG_SHIFT,\pmd /* clear flags */
481 copy \pmd,%r9
482 #ifdef CONFIG_64BIT
483 shld %r9,PxD_VALUE_SHIFT,\pmd
484 #else
485 shlw %r9,PxD_VALUE_SHIFT,\pmd
486 #endif
487 EXTR \va,31-PAGE_SHIFT,ASM_BITS_PER_PTE,\index
488 DEP %r0,31,PAGE_SHIFT,\pmd /* clear offset */
489 shladd \index,BITS_PER_PTE_ENTRY,\pmd,\pmd
490 LDREG %r0(\pmd),\pte /* pmd is now pte */
491 bb,>=,n \pte,_PAGE_PRESENT_BIT,\fault
492 .endm
493
494 /* Look up PTE in a 3-Level scheme.
495 *
496 * Here we implement a Hybrid L2/L3 scheme: we allocate the
497 * first pmd adjacent to the pgd. This means that we can
498 * subtract a constant offset to get to it. The pmd and pgd
499 * sizes are arranged so that a single pmd covers 4GB (giving
500 * a full LP64 process access to 8TB) so our lookups are
501 * effectively L2 for the first 4GB of the kernel (i.e. for
502 * all ILP32 processes and all the kernel for machines with
503 * under 4GB of memory) */
504 .macro L3_ptep pgd,pte,index,va,fault
505 extrd,u \va,63-ASM_PGDIR_SHIFT,ASM_BITS_PER_PGD,\index
506 copy %r0,\pte
507 extrd,u,*= \va,31,32,%r0
508 ldw,s \index(\pgd),\pgd
509 extrd,u,*= \va,31,32,%r0
510 bb,>=,n \pgd,_PxD_PRESENT_BIT,\fault
511 extrd,u,*= \va,31,32,%r0
512 shld \pgd,PxD_VALUE_SHIFT,\index
513 extrd,u,*= \va,31,32,%r0
514 copy \index,\pgd
515 extrd,u,*<> \va,31,32,%r0
516 ldo ASM_PGD_PMD_OFFSET(\pgd),\pgd
517 L2_ptep \pgd,\pte,\index,\va,\fault
518 .endm
519
520 /* Set the _PAGE_ACCESSED bit of the PTE. Be clever and
521 * don't needlessly dirty the cache line if it was already set */
522 .macro update_ptep ptep,pte,tmp,tmp1
523 ldi _PAGE_ACCESSED,\tmp1
524 or \tmp1,\pte,\tmp
525 and,COND(<>) \tmp1,\pte,%r0
526 STREG \tmp,0(\ptep)
527 .endm
528
529 /* Set the dirty bit (and accessed bit). No need to be
530 * clever, this is only used from the dirty fault */
531 .macro update_dirty ptep,pte,tmp
532 ldi _PAGE_ACCESSED|_PAGE_DIRTY,\tmp
533 or \tmp,\pte,\pte
534 STREG \pte,0(\ptep)
535 .endm
536
537 /* Convert the pte and prot to tlb insertion values. How
538 * this happens is quite subtle, read below */
539 .macro make_insert_tlb spc,pte,prot
540 space_to_prot \spc \prot /* create prot id from space */
541 /* The following is the real subtlety. This is depositing
542 * T <-> _PAGE_REFTRAP
543 * D <-> _PAGE_DIRTY
544 * B <-> _PAGE_DMB (memory break)
545 *
546 * Then incredible subtlety: The access rights are
547 * _PAGE_GATEWAY _PAGE_EXEC _PAGE_READ
548 * See 3-14 of the parisc 2.0 manual
549 *
550 * Finally, _PAGE_READ goes in the top bit of PL1 (so we
551 * trigger an access rights trap in user space if the user
552 * tries to read an unreadable page */
553 depd \pte,8,7,\prot
554
555 /* PAGE_USER indicates the page can be read with user privileges,
556 * so deposit X1|11 to PL1|PL2 (remember the upper bit of PL1
557 * contains _PAGE_READ */
558 extrd,u,*= \pte,_PAGE_USER_BIT+32,1,%r0
559 depdi 7,11,3,\prot
560 /* If we're a gateway page, drop PL2 back to zero for promotion
561 * to kernel privilege (so we can execute the page as kernel).
562 * Any privilege promotion page always denys read and write */
563 extrd,u,*= \pte,_PAGE_GATEWAY_BIT+32,1,%r0
564 depd %r0,11,2,\prot /* If Gateway, Set PL2 to 0 */
565
566 /* Get rid of prot bits and convert to page addr for iitlbt */
567
568 depd %r0,63,PAGE_SHIFT,\pte
569 extrd,u \pte,56,32,\pte
570 .endm
571
572 /* Identical macro to make_insert_tlb above, except it
573 * makes the tlb entry for the differently formatted pa11
574 * insertion instructions */
575 .macro make_insert_tlb_11 spc,pte,prot
576 zdep \spc,30,15,\prot
577 dep \pte,8,7,\prot
578 extru,= \pte,_PAGE_NO_CACHE_BIT,1,%r0
579 depi 1,12,1,\prot
580 extru,= \pte,_PAGE_USER_BIT,1,%r0
581 depi 7,11,3,\prot /* Set for user space (1 rsvd for read) */
582 extru,= \pte,_PAGE_GATEWAY_BIT,1,%r0
583 depi 0,11,2,\prot /* If Gateway, Set PL2 to 0 */
584
585 /* Get rid of prot bits and convert to page addr for iitlba */
586
587 depi 0,31,12,\pte
588 extru \pte,24,25,\pte
589
590 .endm
591
592 /* This is for ILP32 PA2.0 only. The TLB insertion needs
593 * to extend into I/O space if the address is 0xfXXXXXXX
594 * so we extend the f's into the top word of the pte in
595 * this case */
596 .macro f_extend pte,tmp
597 extrd,s \pte,42,4,\tmp
598 addi,<> 1,\tmp,%r0
599 extrd,s \pte,63,25,\pte
600 .endm
601
602 /* The alias region is an 8MB aligned 16MB to do clear and
603 * copy user pages at addresses congruent with the user
604 * virtual address.
605 *
606 * To use the alias page, you set %r26 up with the to TLB
607 * entry (identifying the physical page) and %r23 up with
608 * the from tlb entry (or nothing if only a to entry---for
609 * clear_user_page_asm) */
610 .macro do_alias spc,tmp,tmp1,va,pte,prot,fault
611 cmpib,COND(<>),n 0,\spc,\fault
612 ldil L%(TMPALIAS_MAP_START),\tmp
613 #if defined(CONFIG_64BIT) && (TMPALIAS_MAP_START >= 0x80000000)
614 /* on LP64, ldi will sign extend into the upper 32 bits,
615 * which is behaviour we don't want */
616 depdi 0,31,32,\tmp
617 #endif
618 copy \va,\tmp1
619 DEPI 0,31,23,\tmp1
620 cmpb,COND(<>),n \tmp,\tmp1,\fault
621 ldi (_PAGE_DIRTY|_PAGE_WRITE|_PAGE_READ),\prot
622 depd,z \prot,8,7,\prot
623 /*
624 * OK, it is in the temp alias region, check whether "from" or "to".
625 * Check "subtle" note in pacache.S re: r23/r26.
626 */
627 #ifdef CONFIG_64BIT
628 extrd,u,*= \va,41,1,%r0
629 #else
630 extrw,u,= \va,9,1,%r0
631 #endif
632 or,COND(tr) %r23,%r0,\pte
633 or %r26,%r0,\pte
634 .endm
635
636
637 /*
638 * Align fault_vector_20 on 4K boundary so that both
639 * fault_vector_11 and fault_vector_20 are on the
640 * same page. This is only necessary as long as we
641 * write protect the kernel text, which we may stop
642 * doing once we use large page translations to cover
643 * the static part of the kernel address space.
644 */
645
646 .export fault_vector_20
647
648 .text
649
650 .align 4096
651
652 fault_vector_20:
653 /* First vector is invalid (0) */
654 .ascii "cows can fly"
655 .byte 0
656 .align 32
657
658 hpmc 1
659 def 2
660 def 3
661 extint 4
662 def 5
663 itlb_20 6
664 def 7
665 def 8
666 def 9
667 def 10
668 def 11
669 def 12
670 def 13
671 def 14
672 dtlb_20 15
673 #if 0
674 naitlb_20 16
675 #else
676 def 16
677 #endif
678 nadtlb_20 17
679 def 18
680 def 19
681 dbit_20 20
682 def 21
683 def 22
684 def 23
685 def 24
686 def 25
687 def 26
688 def 27
689 def 28
690 def 29
691 def 30
692 def 31
693
694 #ifndef CONFIG_64BIT
695
696 .export fault_vector_11
697
698 .align 2048
699
700 fault_vector_11:
701 /* First vector is invalid (0) */
702 .ascii "cows can fly"
703 .byte 0
704 .align 32
705
706 hpmc 1
707 def 2
708 def 3
709 extint 4
710 def 5
711 itlb_11 6
712 def 7
713 def 8
714 def 9
715 def 10
716 def 11
717 def 12
718 def 13
719 def 14
720 dtlb_11 15
721 #if 0
722 naitlb_11 16
723 #else
724 def 16
725 #endif
726 nadtlb_11 17
727 def 18
728 def 19
729 dbit_11 20
730 def 21
731 def 22
732 def 23
733 def 24
734 def 25
735 def 26
736 def 27
737 def 28
738 def 29
739 def 30
740 def 31
741
742 #endif
743
744 .import handle_interruption,code
745 .import do_cpu_irq_mask,code
746
747 /*
748 * r26 = function to be called
749 * r25 = argument to pass in
750 * r24 = flags for do_fork()
751 *
752 * Kernel threads don't ever return, so they don't need
753 * a true register context. We just save away the arguments
754 * for copy_thread/ret_ to properly set up the child.
755 */
756
757 #define CLONE_VM 0x100 /* Must agree with <linux/sched.h> */
758 #define CLONE_UNTRACED 0x00800000
759
760 .export __kernel_thread, code
761 .import do_fork
762 __kernel_thread:
763 STREG %r2, -RP_OFFSET(%r30)
764
765 copy %r30, %r1
766 ldo PT_SZ_ALGN(%r30),%r30
767 #ifdef CONFIG_64BIT
768 /* Yo, function pointers in wide mode are little structs... -PB */
769 ldd 24(%r26), %r2
770 STREG %r2, PT_GR27(%r1) /* Store childs %dp */
771 ldd 16(%r26), %r26
772
773 STREG %r22, PT_GR22(%r1) /* save r22 (arg5) */
774 copy %r0, %r22 /* user_tid */
775 #endif
776 STREG %r26, PT_GR26(%r1) /* Store function & argument for child */
777 STREG %r25, PT_GR25(%r1)
778 ldil L%CLONE_UNTRACED, %r26
779 ldo CLONE_VM(%r26), %r26 /* Force CLONE_VM since only init_mm */
780 or %r26, %r24, %r26 /* will have kernel mappings. */
781 ldi 1, %r25 /* stack_start, signals kernel thread */
782 stw %r0, -52(%r30) /* user_tid */
783 #ifdef CONFIG_64BIT
784 ldo -16(%r30),%r29 /* Reference param save area */
785 #endif
786 BL do_fork, %r2
787 copy %r1, %r24 /* pt_regs */
788
789 /* Parent Returns here */
790
791 LDREG -PT_SZ_ALGN-RP_OFFSET(%r30), %r2
792 ldo -PT_SZ_ALGN(%r30), %r30
793 bv %r0(%r2)
794 nop
795
796 /*
797 * Child Returns here
798 *
799 * copy_thread moved args from temp save area set up above
800 * into task save area.
801 */
802
803 .export ret_from_kernel_thread
804 ret_from_kernel_thread:
805
806 /* Call schedule_tail first though */
807 BL schedule_tail, %r2
808 nop
809
810 LDREG TI_TASK-THREAD_SZ_ALGN(%r30), %r1
811 LDREG TASK_PT_GR25(%r1), %r26
812 #ifdef CONFIG_64BIT
813 LDREG TASK_PT_GR27(%r1), %r27
814 LDREG TASK_PT_GR22(%r1), %r22
815 #endif
816 LDREG TASK_PT_GR26(%r1), %r1
817 ble 0(%sr7, %r1)
818 copy %r31, %r2
819
820 #ifdef CONFIG_64BIT
821 ldo -16(%r30),%r29 /* Reference param save area */
822 loadgp /* Thread could have been in a module */
823 #endif
824 #ifndef CONFIG_64BIT
825 b sys_exit
826 #else
827 load32 sys_exit, %r1
828 bv %r0(%r1)
829 #endif
830 ldi 0, %r26
831
832 .import sys_execve, code
833 .export __execve, code
834 __execve:
835 copy %r2, %r15
836 copy %r30, %r16
837 ldo PT_SZ_ALGN(%r30), %r30
838 STREG %r26, PT_GR26(%r16)
839 STREG %r25, PT_GR25(%r16)
840 STREG %r24, PT_GR24(%r16)
841 #ifdef CONFIG_64BIT
842 ldo -16(%r30),%r29 /* Reference param save area */
843 #endif
844 BL sys_execve, %r2
845 copy %r16, %r26
846
847 cmpib,=,n 0,%r28,intr_return /* forward */
848
849 /* yes, this will trap and die. */
850 copy %r15, %r2
851 copy %r16, %r30
852 bv %r0(%r2)
853 nop
854
855 .align 4
856
857 /*
858 * struct task_struct *_switch_to(struct task_struct *prev,
859 * struct task_struct *next)
860 *
861 * switch kernel stacks and return prev */
862 .export _switch_to, code
863 _switch_to:
864 STREG %r2, -RP_OFFSET(%r30)
865
866 callee_save_float
867 callee_save
868
869 load32 _switch_to_ret, %r2
870
871 STREG %r2, TASK_PT_KPC(%r26)
872 LDREG TASK_PT_KPC(%r25), %r2
873
874 STREG %r30, TASK_PT_KSP(%r26)
875 LDREG TASK_PT_KSP(%r25), %r30
876 LDREG TASK_THREAD_INFO(%r25), %r25
877 bv %r0(%r2)
878 mtctl %r25,%cr30
879
880 _switch_to_ret:
881 mtctl %r0, %cr0 /* Needed for single stepping */
882 callee_rest
883 callee_rest_float
884
885 LDREG -RP_OFFSET(%r30), %r2
886 bv %r0(%r2)
887 copy %r26, %r28
888
889 /*
890 * Common rfi return path for interruptions, kernel execve, and
891 * sys_rt_sigreturn (sometimes). The sys_rt_sigreturn syscall will
892 * return via this path if the signal was received when the process
893 * was running; if the process was blocked on a syscall then the
894 * normal syscall_exit path is used. All syscalls for traced
895 * proceses exit via intr_restore.
896 *
897 * XXX If any syscalls that change a processes space id ever exit
898 * this way, then we will need to copy %sr3 in to PT_SR[3..7], and
899 * adjust IASQ[0..1].
900 *
901 */
902
903 .align 4096
904
905 .export syscall_exit_rfi
906 syscall_exit_rfi:
907 mfctl %cr30,%r16
908 LDREG TI_TASK(%r16), %r16 /* thread_info -> task_struct */
909 ldo TASK_REGS(%r16),%r16
910 /* Force iaoq to userspace, as the user has had access to our current
911 * context via sigcontext. Also Filter the PSW for the same reason.
912 */
913 LDREG PT_IAOQ0(%r16),%r19
914 depi 3,31,2,%r19
915 STREG %r19,PT_IAOQ0(%r16)
916 LDREG PT_IAOQ1(%r16),%r19
917 depi 3,31,2,%r19
918 STREG %r19,PT_IAOQ1(%r16)
919 LDREG PT_PSW(%r16),%r19
920 load32 USER_PSW_MASK,%r1
921 #ifdef CONFIG_64BIT
922 load32 USER_PSW_HI_MASK,%r20
923 depd %r20,31,32,%r1
924 #endif
925 and %r19,%r1,%r19 /* Mask out bits that user shouldn't play with */
926 load32 USER_PSW,%r1
927 or %r19,%r1,%r19 /* Make sure default USER_PSW bits are set */
928 STREG %r19,PT_PSW(%r16)
929
930 /*
931 * If we aren't being traced, we never saved space registers
932 * (we don't store them in the sigcontext), so set them
933 * to "proper" values now (otherwise we'll wind up restoring
934 * whatever was last stored in the task structure, which might
935 * be inconsistent if an interrupt occured while on the gateway
936 * page) Note that we may be "trashing" values the user put in
937 * them, but we don't support the the user changing them.
938 */
939
940 STREG %r0,PT_SR2(%r16)
941 mfsp %sr3,%r19
942 STREG %r19,PT_SR0(%r16)
943 STREG %r19,PT_SR1(%r16)
944 STREG %r19,PT_SR3(%r16)
945 STREG %r19,PT_SR4(%r16)
946 STREG %r19,PT_SR5(%r16)
947 STREG %r19,PT_SR6(%r16)
948 STREG %r19,PT_SR7(%r16)
949
950 intr_return:
951 /* NOTE: Need to enable interrupts incase we schedule. */
952 ssm PSW_SM_I, %r0
953
954 /* Check for software interrupts */
955
956 .import irq_stat,data
957
958 load32 irq_stat,%r19
959 #ifdef CONFIG_SMP
960 mfctl %cr30,%r1
961 ldw TI_CPU(%r1),%r1 /* get cpu # - int */
962 /* shift left ____cacheline_aligned (aka L1_CACHE_BYTES) amount
963 ** irq_stat[] is defined using ____cacheline_aligned.
964 */
965 #ifdef CONFIG_64BIT
966 shld %r1, 6, %r20
967 #else
968 shlw %r1, 5, %r20
969 #endif
970 add %r19,%r20,%r19 /* now have &irq_stat[smp_processor_id()] */
971 #endif /* CONFIG_SMP */
972
973 intr_check_resched:
974
975 /* check for reschedule */
976 mfctl %cr30,%r1
977 LDREG TI_FLAGS(%r1),%r19 /* sched.h: TIF_NEED_RESCHED */
978 bb,<,n %r19,31-TIF_NEED_RESCHED,intr_do_resched /* forward */
979
980 intr_check_sig:
981 /* As above */
982 mfctl %cr30,%r1
983 LDREG TI_FLAGS(%r1),%r19 /* sched.h: TIF_SIGPENDING */
984 bb,<,n %r19, 31-TIF_SIGPENDING, intr_do_signal /* forward */
985
986 intr_restore:
987 copy %r16,%r29
988 ldo PT_FR31(%r29),%r1
989 rest_fp %r1
990 rest_general %r29
991
992 /* inverse of virt_map */
993 pcxt_ssm_bug
994 rsm PSW_SM_QUIET,%r0 /* prepare for rfi */
995 tophys_r1 %r29
996
997 /* Restore space id's and special cr's from PT_REGS
998 * structure pointed to by r29
999 */
1000 rest_specials %r29
1001
1002 /* IMPORTANT: rest_stack restores r29 last (we are using it)!
1003 * It also restores r1 and r30.
1004 */
1005 rest_stack
1006
1007 rfi
1008 nop
1009 nop
1010 nop
1011 nop
1012 nop
1013 nop
1014 nop
1015 nop
1016
1017 .import schedule,code
1018 intr_do_resched:
1019 /* Only do reschedule if we are returning to user space */
1020 LDREG PT_IASQ0(%r16), %r20
1021 CMPIB= 0,%r20,intr_restore /* backward */
1022 nop
1023 LDREG PT_IASQ1(%r16), %r20
1024 CMPIB= 0,%r20,intr_restore /* backward */
1025 nop
1026
1027 #ifdef CONFIG_64BIT
1028 ldo -16(%r30),%r29 /* Reference param save area */
1029 #endif
1030
1031 ldil L%intr_check_sig, %r2
1032 #ifndef CONFIG_64BIT
1033 b schedule
1034 #else
1035 load32 schedule, %r20
1036 bv %r0(%r20)
1037 #endif
1038 ldo R%intr_check_sig(%r2), %r2
1039
1040
1041 .import do_signal,code
1042 intr_do_signal:
1043 /*
1044 This check is critical to having LWS
1045 working. The IASQ is zero on the gateway
1046 page and we cannot deliver any signals until
1047 we get off the gateway page.
1048
1049 Only do signals if we are returning to user space
1050 */
1051 LDREG PT_IASQ0(%r16), %r20
1052 CMPIB= 0,%r20,intr_restore /* backward */
1053 nop
1054 LDREG PT_IASQ1(%r16), %r20
1055 CMPIB= 0,%r20,intr_restore /* backward */
1056 nop
1057
1058 copy %r0, %r24 /* unsigned long in_syscall */
1059 copy %r16, %r25 /* struct pt_regs *regs */
1060 #ifdef CONFIG_64BIT
1061 ldo -16(%r30),%r29 /* Reference param save area */
1062 #endif
1063
1064 BL do_signal,%r2
1065 copy %r0, %r26 /* sigset_t *oldset = NULL */
1066
1067 b intr_check_sig
1068 nop
1069
1070 /*
1071 * External interrupts.
1072 */
1073
1074 intr_extint:
1075 CMPIB=,n 0,%r16,1f
1076 get_stack_use_cr30
1077 b,n 3f
1078
1079 1:
1080 #if 0 /* Interrupt Stack support not working yet! */
1081 mfctl %cr31,%r1
1082 copy %r30,%r17
1083 /* FIXME! depi below has hardcoded idea of interrupt stack size (32k)*/
1084 #ifdef CONFIG_64BIT
1085 depdi 0,63,15,%r17
1086 #else
1087 depi 0,31,15,%r17
1088 #endif
1089 CMPB=,n %r1,%r17,2f
1090 get_stack_use_cr31
1091 b,n 3f
1092 #endif
1093 2:
1094 get_stack_use_r30
1095
1096 3:
1097 save_specials %r29
1098 virt_map
1099 save_general %r29
1100
1101 ldo PT_FR0(%r29), %r24
1102 save_fp %r24
1103
1104 loadgp
1105
1106 copy %r29, %r26 /* arg0 is pt_regs */
1107 copy %r29, %r16 /* save pt_regs */
1108
1109 ldil L%intr_return, %r2
1110
1111 #ifdef CONFIG_64BIT
1112 ldo -16(%r30),%r29 /* Reference param save area */
1113 #endif
1114
1115 b do_cpu_irq_mask
1116 ldo R%intr_return(%r2), %r2 /* return to intr_return, not here */
1117
1118
1119 /* Generic interruptions (illegal insn, unaligned, page fault, etc) */
1120
1121 .export intr_save, code /* for os_hpmc */
1122
1123 intr_save:
1124 mfsp %sr7,%r16
1125 CMPIB=,n 0,%r16,1f
1126 get_stack_use_cr30
1127 b 2f
1128 copy %r8,%r26
1129
1130 1:
1131 get_stack_use_r30
1132 copy %r8,%r26
1133
1134 2:
1135 save_specials %r29
1136
1137 /* If this trap is a itlb miss, skip saving/adjusting isr/ior */
1138
1139 /*
1140 * FIXME: 1) Use a #define for the hardwired "6" below (and in
1141 * traps.c.
1142 * 2) Once we start executing code above 4 Gb, we need
1143 * to adjust iasq/iaoq here in the same way we
1144 * adjust isr/ior below.
1145 */
1146
1147 CMPIB=,n 6,%r26,skip_save_ior
1148
1149
1150 mfctl %cr20, %r16 /* isr */
1151 nop /* serialize mfctl on PA 2.0 to avoid 4 cycle penalty */
1152 mfctl %cr21, %r17 /* ior */
1153
1154
1155 #ifdef CONFIG_64BIT
1156 /*
1157 * If the interrupted code was running with W bit off (32 bit),
1158 * clear the b bits (bits 0 & 1) in the ior.
1159 * save_specials left ipsw value in r8 for us to test.
1160 */
1161 extrd,u,*<> %r8,PSW_W_BIT,1,%r0
1162 depdi 0,1,2,%r17
1163
1164 /*
1165 * FIXME: This code has hardwired assumptions about the split
1166 * between space bits and offset bits. This will change
1167 * when we allow alternate page sizes.
1168 */
1169
1170 /* adjust isr/ior. */
1171
1172 extrd,u %r16,63,7,%r1 /* get high bits from isr for ior */
1173 depd %r1,31,7,%r17 /* deposit them into ior */
1174 depdi 0,63,7,%r16 /* clear them from isr */
1175 #endif
1176 STREG %r16, PT_ISR(%r29)
1177 STREG %r17, PT_IOR(%r29)
1178
1179
1180 skip_save_ior:
1181 virt_map
1182 save_general %r29
1183
1184 ldo PT_FR0(%r29), %r25
1185 save_fp %r25
1186
1187 loadgp
1188
1189 copy %r29, %r25 /* arg1 is pt_regs */
1190 #ifdef CONFIG_64BIT
1191 ldo -16(%r30),%r29 /* Reference param save area */
1192 #endif
1193
1194 ldil L%intr_check_sig, %r2
1195 copy %r25, %r16 /* save pt_regs */
1196
1197 b handle_interruption
1198 ldo R%intr_check_sig(%r2), %r2
1199
1200
1201 /*
1202 * Note for all tlb miss handlers:
1203 *
1204 * cr24 contains a pointer to the kernel address space
1205 * page directory.
1206 *
1207 * cr25 contains a pointer to the current user address
1208 * space page directory.
1209 *
1210 * sr3 will contain the space id of the user address space
1211 * of the current running thread while that thread is
1212 * running in the kernel.
1213 */
1214
1215 /*
1216 * register number allocations. Note that these are all
1217 * in the shadowed registers
1218 */
1219
1220 t0 = r1 /* temporary register 0 */
1221 va = r8 /* virtual address for which the trap occured */
1222 t1 = r9 /* temporary register 1 */
1223 pte = r16 /* pte/phys page # */
1224 prot = r17 /* prot bits */
1225 spc = r24 /* space for which the trap occured */
1226 ptp = r25 /* page directory/page table pointer */
1227
1228 #ifdef CONFIG_64BIT
1229
1230 dtlb_miss_20w:
1231 space_adjust spc,va,t0
1232 get_pgd spc,ptp
1233 space_check spc,t0,dtlb_fault
1234
1235 L3_ptep ptp,pte,t0,va,dtlb_check_alias_20w
1236
1237 update_ptep ptp,pte,t0,t1
1238
1239 make_insert_tlb spc,pte,prot
1240
1241 idtlbt pte,prot
1242
1243 rfir
1244 nop
1245
1246 dtlb_check_alias_20w:
1247 do_alias spc,t0,t1,va,pte,prot,dtlb_fault
1248
1249 idtlbt pte,prot
1250
1251 rfir
1252 nop
1253
1254 nadtlb_miss_20w:
1255 space_adjust spc,va,t0
1256 get_pgd spc,ptp
1257 space_check spc,t0,nadtlb_fault
1258
1259 L3_ptep ptp,pte,t0,va,nadtlb_check_flush_20w
1260
1261 update_ptep ptp,pte,t0,t1
1262
1263 make_insert_tlb spc,pte,prot
1264
1265 idtlbt pte,prot
1266
1267 rfir
1268 nop
1269
1270 nadtlb_check_flush_20w:
1271 bb,>=,n pte,_PAGE_FLUSH_BIT,nadtlb_emulate
1272
1273 /* Insert a "flush only" translation */
1274
1275 depdi,z 7,7,3,prot
1276 depdi 1,10,1,prot
1277
1278 /* Get rid of prot bits and convert to page addr for idtlbt */
1279
1280 depdi 0,63,12,pte
1281 extrd,u pte,56,52,pte
1282 idtlbt pte,prot
1283
1284 rfir
1285 nop
1286
1287 #else
1288
1289 dtlb_miss_11:
1290 get_pgd spc,ptp
1291
1292 space_check spc,t0,dtlb_fault
1293
1294 L2_ptep ptp,pte,t0,va,dtlb_check_alias_11
1295
1296 update_ptep ptp,pte,t0,t1
1297
1298 make_insert_tlb_11 spc,pte,prot
1299
1300 mfsp %sr1,t0 /* Save sr1 so we can use it in tlb inserts */
1301 mtsp spc,%sr1
1302
1303 idtlba pte,(%sr1,va)
1304 idtlbp prot,(%sr1,va)
1305
1306 mtsp t0, %sr1 /* Restore sr1 */
1307
1308 rfir
1309 nop
1310
1311 dtlb_check_alias_11:
1312
1313 /* Check to see if fault is in the temporary alias region */
1314
1315 cmpib,<>,n 0,spc,dtlb_fault /* forward */
1316 ldil L%(TMPALIAS_MAP_START),t0
1317 copy va,t1
1318 depwi 0,31,23,t1
1319 cmpb,<>,n t0,t1,dtlb_fault /* forward */
1320 ldi (_PAGE_DIRTY|_PAGE_WRITE|_PAGE_READ),prot
1321 depw,z prot,8,7,prot
1322
1323 /*
1324 * OK, it is in the temp alias region, check whether "from" or "to".
1325 * Check "subtle" note in pacache.S re: r23/r26.
1326 */
1327
1328 extrw,u,= va,9,1,r0
1329 or,tr %r23,%r0,pte /* If "from" use "from" page */
1330 or %r26,%r0,pte /* else "to", use "to" page */
1331
1332 idtlba pte,(va)
1333 idtlbp prot,(va)
1334
1335 rfir
1336 nop
1337
1338 nadtlb_miss_11:
1339 get_pgd spc,ptp
1340
1341 space_check spc,t0,nadtlb_fault
1342
1343 L2_ptep ptp,pte,t0,va,nadtlb_check_flush_11
1344
1345 update_ptep ptp,pte,t0,t1
1346
1347 make_insert_tlb_11 spc,pte,prot
1348
1349
1350 mfsp %sr1,t0 /* Save sr1 so we can use it in tlb inserts */
1351 mtsp spc,%sr1
1352
1353 idtlba pte,(%sr1,va)
1354 idtlbp prot,(%sr1,va)
1355
1356 mtsp t0, %sr1 /* Restore sr1 */
1357
1358 rfir
1359 nop
1360
1361 nadtlb_check_flush_11:
1362 bb,>=,n pte,_PAGE_FLUSH_BIT,nadtlb_emulate
1363
1364 /* Insert a "flush only" translation */
1365
1366 zdepi 7,7,3,prot
1367 depi 1,10,1,prot
1368
1369 /* Get rid of prot bits and convert to page addr for idtlba */
1370
1371 depi 0,31,12,pte
1372 extru pte,24,25,pte
1373
1374 mfsp %sr1,t0 /* Save sr1 so we can use it in tlb inserts */
1375 mtsp spc,%sr1
1376
1377 idtlba pte,(%sr1,va)
1378 idtlbp prot,(%sr1,va)
1379
1380 mtsp t0, %sr1 /* Restore sr1 */
1381
1382 rfir
1383 nop
1384
1385 dtlb_miss_20:
1386 space_adjust spc,va,t0
1387 get_pgd spc,ptp
1388 space_check spc,t0,dtlb_fault
1389
1390 L2_ptep ptp,pte,t0,va,dtlb_check_alias_20
1391
1392 update_ptep ptp,pte,t0,t1
1393
1394 make_insert_tlb spc,pte,prot
1395
1396 f_extend pte,t0
1397
1398 idtlbt pte,prot
1399
1400 rfir
1401 nop
1402
1403 dtlb_check_alias_20:
1404 do_alias spc,t0,t1,va,pte,prot,dtlb_fault
1405
1406 idtlbt pte,prot
1407
1408 rfir
1409 nop
1410
1411 nadtlb_miss_20:
1412 get_pgd spc,ptp
1413
1414 space_check spc,t0,nadtlb_fault
1415
1416 L2_ptep ptp,pte,t0,va,nadtlb_check_flush_20
1417
1418 update_ptep ptp,pte,t0,t1
1419
1420 make_insert_tlb spc,pte,prot
1421
1422 f_extend pte,t0
1423
1424 idtlbt pte,prot
1425
1426 rfir
1427 nop
1428
1429 nadtlb_check_flush_20:
1430 bb,>=,n pte,_PAGE_FLUSH_BIT,nadtlb_emulate
1431
1432 /* Insert a "flush only" translation */
1433
1434 depdi,z 7,7,3,prot
1435 depdi 1,10,1,prot
1436
1437 /* Get rid of prot bits and convert to page addr for idtlbt */
1438
1439 depdi 0,63,12,pte
1440 extrd,u pte,56,32,pte
1441 idtlbt pte,prot
1442
1443 rfir
1444 nop
1445 #endif
1446
1447 nadtlb_emulate:
1448
1449 /*
1450 * Non access misses can be caused by fdc,fic,pdc,lpa,probe and
1451 * probei instructions. We don't want to fault for these
1452 * instructions (not only does it not make sense, it can cause
1453 * deadlocks, since some flushes are done with the mmap
1454 * semaphore held). If the translation doesn't exist, we can't
1455 * insert a translation, so have to emulate the side effects
1456 * of the instruction. Since we don't insert a translation
1457 * we can get a lot of faults during a flush loop, so it makes
1458 * sense to try to do it here with minimum overhead. We only
1459 * emulate fdc,fic,pdc,probew,prober instructions whose base
1460 * and index registers are not shadowed. We defer everything
1461 * else to the "slow" path.
1462 */
1463
1464 mfctl %cr19,%r9 /* Get iir */
1465
1466 /* PA 2.0 Arch Ref. Book pg 382 has a good description of the insn bits.
1467 Checks for fdc,fdce,pdc,"fic,4f",prober,probeir,probew, probeiw */
1468
1469 /* Checks for fdc,fdce,pdc,"fic,4f" only */
1470 ldi 0x280,%r16
1471 and %r9,%r16,%r17
1472 cmpb,<>,n %r16,%r17,nadtlb_probe_check
1473 bb,>=,n %r9,26,nadtlb_nullify /* m bit not set, just nullify */
1474 BL get_register,%r25
1475 extrw,u %r9,15,5,%r8 /* Get index register # */
1476 CMPIB=,n -1,%r1,nadtlb_fault /* have to use slow path */
1477 copy %r1,%r24
1478 BL get_register,%r25
1479 extrw,u %r9,10,5,%r8 /* Get base register # */
1480 CMPIB=,n -1,%r1,nadtlb_fault /* have to use slow path */
1481 BL set_register,%r25
1482 add,l %r1,%r24,%r1 /* doesn't affect c/b bits */
1483
1484 nadtlb_nullify:
1485 mfctl %ipsw,%r8
1486 ldil L%PSW_N,%r9
1487 or %r8,%r9,%r8 /* Set PSW_N */
1488 mtctl %r8,%ipsw
1489
1490 rfir
1491 nop
1492
1493 /*
1494 When there is no translation for the probe address then we
1495 must nullify the insn and return zero in the target regsiter.
1496 This will indicate to the calling code that it does not have
1497 write/read privileges to this address.
1498
1499 This should technically work for prober and probew in PA 1.1,
1500 and also probe,r and probe,w in PA 2.0
1501
1502 WARNING: USE ONLY NON-SHADOW REGISTERS WITH PROBE INSN!
1503 THE SLOW-PATH EMULATION HAS NOT BEEN WRITTEN YET.
1504
1505 */
1506 nadtlb_probe_check:
1507 ldi 0x80,%r16
1508 and %r9,%r16,%r17
1509 cmpb,<>,n %r16,%r17,nadtlb_fault /* Must be probe,[rw]*/
1510 BL get_register,%r25 /* Find the target register */
1511 extrw,u %r9,31,5,%r8 /* Get target register */
1512 CMPIB=,n -1,%r1,nadtlb_fault /* have to use slow path */
1513 BL set_register,%r25
1514 copy %r0,%r1 /* Write zero to target register */
1515 b nadtlb_nullify /* Nullify return insn */
1516 nop
1517
1518
1519 #ifdef CONFIG_64BIT
1520 itlb_miss_20w:
1521
1522 /*
1523 * I miss is a little different, since we allow users to fault
1524 * on the gateway page which is in the kernel address space.
1525 */
1526
1527 space_adjust spc,va,t0
1528 get_pgd spc,ptp
1529 space_check spc,t0,itlb_fault
1530
1531 L3_ptep ptp,pte,t0,va,itlb_fault
1532
1533 update_ptep ptp,pte,t0,t1
1534
1535 make_insert_tlb spc,pte,prot
1536
1537 iitlbt pte,prot
1538
1539 rfir
1540 nop
1541
1542 #else
1543
1544 itlb_miss_11:
1545 get_pgd spc,ptp
1546
1547 space_check spc,t0,itlb_fault
1548
1549 L2_ptep ptp,pte,t0,va,itlb_fault
1550
1551 update_ptep ptp,pte,t0,t1
1552
1553 make_insert_tlb_11 spc,pte,prot
1554
1555 mfsp %sr1,t0 /* Save sr1 so we can use it in tlb inserts */
1556 mtsp spc,%sr1
1557
1558 iitlba pte,(%sr1,va)
1559 iitlbp prot,(%sr1,va)
1560
1561 mtsp t0, %sr1 /* Restore sr1 */
1562
1563 rfir
1564 nop
1565
1566 itlb_miss_20:
1567 get_pgd spc,ptp
1568
1569 space_check spc,t0,itlb_fault
1570
1571 L2_ptep ptp,pte,t0,va,itlb_fault
1572
1573 update_ptep ptp,pte,t0,t1
1574
1575 make_insert_tlb spc,pte,prot
1576
1577 f_extend pte,t0
1578
1579 iitlbt pte,prot
1580
1581 rfir
1582 nop
1583
1584 #endif
1585
1586 #ifdef CONFIG_64BIT
1587
1588 dbit_trap_20w:
1589 space_adjust spc,va,t0
1590 get_pgd spc,ptp
1591 space_check spc,t0,dbit_fault
1592
1593 L3_ptep ptp,pte,t0,va,dbit_fault
1594
1595 #ifdef CONFIG_SMP
1596 CMPIB=,n 0,spc,dbit_nolock_20w
1597 load32 PA(pa_dbit_lock),t0
1598
1599 dbit_spin_20w:
1600 ldcw 0(t0),t1
1601 cmpib,= 0,t1,dbit_spin_20w
1602 nop
1603
1604 dbit_nolock_20w:
1605 #endif
1606 update_dirty ptp,pte,t1
1607
1608 make_insert_tlb spc,pte,prot
1609
1610 idtlbt pte,prot
1611 #ifdef CONFIG_SMP
1612 CMPIB=,n 0,spc,dbit_nounlock_20w
1613 ldi 1,t1
1614 stw t1,0(t0)
1615
1616 dbit_nounlock_20w:
1617 #endif
1618
1619 rfir
1620 nop
1621 #else
1622
1623 dbit_trap_11:
1624
1625 get_pgd spc,ptp
1626
1627 space_check spc,t0,dbit_fault
1628
1629 L2_ptep ptp,pte,t0,va,dbit_fault
1630
1631 #ifdef CONFIG_SMP
1632 CMPIB=,n 0,spc,dbit_nolock_11
1633 load32 PA(pa_dbit_lock),t0
1634
1635 dbit_spin_11:
1636 ldcw 0(t0),t1
1637 cmpib,= 0,t1,dbit_spin_11
1638 nop
1639
1640 dbit_nolock_11:
1641 #endif
1642 update_dirty ptp,pte,t1
1643
1644 make_insert_tlb_11 spc,pte,prot
1645
1646 mfsp %sr1,t1 /* Save sr1 so we can use it in tlb inserts */
1647 mtsp spc,%sr1
1648
1649 idtlba pte,(%sr1,va)
1650 idtlbp prot,(%sr1,va)
1651
1652 mtsp t1, %sr1 /* Restore sr1 */
1653 #ifdef CONFIG_SMP
1654 CMPIB=,n 0,spc,dbit_nounlock_11
1655 ldi 1,t1
1656 stw t1,0(t0)
1657
1658 dbit_nounlock_11:
1659 #endif
1660
1661 rfir
1662 nop
1663
1664 dbit_trap_20:
1665 get_pgd spc,ptp
1666
1667 space_check spc,t0,dbit_fault
1668
1669 L2_ptep ptp,pte,t0,va,dbit_fault
1670
1671 #ifdef CONFIG_SMP
1672 CMPIB=,n 0,spc,dbit_nolock_20
1673 load32 PA(pa_dbit_lock),t0
1674
1675 dbit_spin_20:
1676 ldcw 0(t0),t1
1677 cmpib,= 0,t1,dbit_spin_20
1678 nop
1679
1680 dbit_nolock_20:
1681 #endif
1682 update_dirty ptp,pte,t1
1683
1684 make_insert_tlb spc,pte,prot
1685
1686 f_extend pte,t1
1687
1688 idtlbt pte,prot
1689
1690 #ifdef CONFIG_SMP
1691 CMPIB=,n 0,spc,dbit_nounlock_20
1692 ldi 1,t1
1693 stw t1,0(t0)
1694
1695 dbit_nounlock_20:
1696 #endif
1697
1698 rfir
1699 nop
1700 #endif
1701
1702 .import handle_interruption,code
1703
1704 kernel_bad_space:
1705 b intr_save
1706 ldi 31,%r8 /* Use an unused code */
1707
1708 dbit_fault:
1709 b intr_save
1710 ldi 20,%r8
1711
1712 itlb_fault:
1713 b intr_save
1714 ldi 6,%r8
1715
1716 nadtlb_fault:
1717 b intr_save
1718 ldi 17,%r8
1719
1720 dtlb_fault:
1721 b intr_save
1722 ldi 15,%r8
1723
1724 /* Register saving semantics for system calls:
1725
1726 %r1 clobbered by system call macro in userspace
1727 %r2 saved in PT_REGS by gateway page
1728 %r3 - %r18 preserved by C code (saved by signal code)
1729 %r19 - %r20 saved in PT_REGS by gateway page
1730 %r21 - %r22 non-standard syscall args
1731 stored in kernel stack by gateway page
1732 %r23 - %r26 arg3-arg0, saved in PT_REGS by gateway page
1733 %r27 - %r30 saved in PT_REGS by gateway page
1734 %r31 syscall return pointer
1735 */
1736
1737 /* Floating point registers (FIXME: what do we do with these?)
1738
1739 %fr0 - %fr3 status/exception, not preserved
1740 %fr4 - %fr7 arguments
1741 %fr8 - %fr11 not preserved by C code
1742 %fr12 - %fr21 preserved by C code
1743 %fr22 - %fr31 not preserved by C code
1744 */
1745
1746 .macro reg_save regs
1747 STREG %r3, PT_GR3(\regs)
1748 STREG %r4, PT_GR4(\regs)
1749 STREG %r5, PT_GR5(\regs)
1750 STREG %r6, PT_GR6(\regs)
1751 STREG %r7, PT_GR7(\regs)
1752 STREG %r8, PT_GR8(\regs)
1753 STREG %r9, PT_GR9(\regs)
1754 STREG %r10,PT_GR10(\regs)
1755 STREG %r11,PT_GR11(\regs)
1756 STREG %r12,PT_GR12(\regs)
1757 STREG %r13,PT_GR13(\regs)
1758 STREG %r14,PT_GR14(\regs)
1759 STREG %r15,PT_GR15(\regs)
1760 STREG %r16,PT_GR16(\regs)
1761 STREG %r17,PT_GR17(\regs)
1762 STREG %r18,PT_GR18(\regs)
1763 .endm
1764
1765 .macro reg_restore regs
1766 LDREG PT_GR3(\regs), %r3
1767 LDREG PT_GR4(\regs), %r4
1768 LDREG PT_GR5(\regs), %r5
1769 LDREG PT_GR6(\regs), %r6
1770 LDREG PT_GR7(\regs), %r7
1771 LDREG PT_GR8(\regs), %r8
1772 LDREG PT_GR9(\regs), %r9
1773 LDREG PT_GR10(\regs),%r10
1774 LDREG PT_GR11(\regs),%r11
1775 LDREG PT_GR12(\regs),%r12
1776 LDREG PT_GR13(\regs),%r13
1777 LDREG PT_GR14(\regs),%r14
1778 LDREG PT_GR15(\regs),%r15
1779 LDREG PT_GR16(\regs),%r16
1780 LDREG PT_GR17(\regs),%r17
1781 LDREG PT_GR18(\regs),%r18
1782 .endm
1783
1784 .export sys_fork_wrapper
1785 .export child_return
1786 sys_fork_wrapper:
1787 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
1788 ldo TASK_REGS(%r1),%r1
1789 reg_save %r1
1790 mfctl %cr27, %r3
1791 STREG %r3, PT_CR27(%r1)
1792
1793 STREG %r2,-RP_OFFSET(%r30)
1794 ldo FRAME_SIZE(%r30),%r30
1795 #ifdef CONFIG_64BIT
1796 ldo -16(%r30),%r29 /* Reference param save area */
1797 #endif
1798
1799 /* These are call-clobbered registers and therefore
1800 also syscall-clobbered (we hope). */
1801 STREG %r2,PT_GR19(%r1) /* save for child */
1802 STREG %r30,PT_GR21(%r1)
1803
1804 LDREG PT_GR30(%r1),%r25
1805 copy %r1,%r24
1806 BL sys_clone,%r2
1807 ldi SIGCHLD,%r26
1808
1809 LDREG -RP_OFFSET-FRAME_SIZE(%r30),%r2
1810 wrapper_exit:
1811 ldo -FRAME_SIZE(%r30),%r30 /* get the stackframe */
1812 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1813 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1814
1815 LDREG PT_CR27(%r1), %r3
1816 mtctl %r3, %cr27
1817 reg_restore %r1
1818
1819 /* strace expects syscall # to be preserved in r20 */
1820 ldi __NR_fork,%r20
1821 bv %r0(%r2)
1822 STREG %r20,PT_GR20(%r1)
1823
1824 /* Set the return value for the child */
1825 child_return:
1826 BL schedule_tail, %r2
1827 nop
1828
1829 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE-FRAME_SIZE(%r30), %r1
1830 LDREG TASK_PT_GR19(%r1),%r2
1831 b wrapper_exit
1832 copy %r0,%r28
1833
1834
1835 .export sys_clone_wrapper
1836 sys_clone_wrapper:
1837 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1838 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1839 reg_save %r1
1840 mfctl %cr27, %r3
1841 STREG %r3, PT_CR27(%r1)
1842
1843 STREG %r2,-RP_OFFSET(%r30)
1844 ldo FRAME_SIZE(%r30),%r30
1845 #ifdef CONFIG_64BIT
1846 ldo -16(%r30),%r29 /* Reference param save area */
1847 #endif
1848
1849 /* WARNING - Clobbers r19 and r21, userspace must save these! */
1850 STREG %r2,PT_GR19(%r1) /* save for child */
1851 STREG %r30,PT_GR21(%r1)
1852 BL sys_clone,%r2
1853 copy %r1,%r24
1854
1855 b wrapper_exit
1856 LDREG -RP_OFFSET-FRAME_SIZE(%r30),%r2
1857
1858 .export sys_vfork_wrapper
1859 sys_vfork_wrapper:
1860 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1861 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1862 reg_save %r1
1863 mfctl %cr27, %r3
1864 STREG %r3, PT_CR27(%r1)
1865
1866 STREG %r2,-RP_OFFSET(%r30)
1867 ldo FRAME_SIZE(%r30),%r30
1868 #ifdef CONFIG_64BIT
1869 ldo -16(%r30),%r29 /* Reference param save area */
1870 #endif
1871
1872 STREG %r2,PT_GR19(%r1) /* save for child */
1873 STREG %r30,PT_GR21(%r1)
1874
1875 BL sys_vfork,%r2
1876 copy %r1,%r26
1877
1878 b wrapper_exit
1879 LDREG -RP_OFFSET-FRAME_SIZE(%r30),%r2
1880
1881
1882 .macro execve_wrapper execve
1883 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1884 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1885
1886 /*
1887 * Do we need to save/restore r3-r18 here?
1888 * I don't think so. why would new thread need old
1889 * threads registers?
1890 */
1891
1892 /* %arg0 - %arg3 are already saved for us. */
1893
1894 STREG %r2,-RP_OFFSET(%r30)
1895 ldo FRAME_SIZE(%r30),%r30
1896 #ifdef CONFIG_64BIT
1897 ldo -16(%r30),%r29 /* Reference param save area */
1898 #endif
1899 BL \execve,%r2
1900 copy %r1,%arg0
1901
1902 ldo -FRAME_SIZE(%r30),%r30
1903 LDREG -RP_OFFSET(%r30),%r2
1904
1905 /* If exec succeeded we need to load the args */
1906
1907 ldo -1024(%r0),%r1
1908 cmpb,>>= %r28,%r1,error_\execve
1909 copy %r2,%r19
1910
1911 error_\execve:
1912 bv %r0(%r19)
1913 nop
1914 .endm
1915
1916 .export sys_execve_wrapper
1917 .import sys_execve
1918
1919 sys_execve_wrapper:
1920 execve_wrapper sys_execve
1921
1922 #ifdef CONFIG_64BIT
1923 .export sys32_execve_wrapper
1924 .import sys32_execve
1925
1926 sys32_execve_wrapper:
1927 execve_wrapper sys32_execve
1928 #endif
1929
1930 .export sys_rt_sigreturn_wrapper
1931 sys_rt_sigreturn_wrapper:
1932 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r26
1933 ldo TASK_REGS(%r26),%r26 /* get pt regs */
1934 /* Don't save regs, we are going to restore them from sigcontext. */
1935 STREG %r2, -RP_OFFSET(%r30)
1936 #ifdef CONFIG_64BIT
1937 ldo FRAME_SIZE(%r30), %r30
1938 BL sys_rt_sigreturn,%r2
1939 ldo -16(%r30),%r29 /* Reference param save area */
1940 #else
1941 BL sys_rt_sigreturn,%r2
1942 ldo FRAME_SIZE(%r30), %r30
1943 #endif
1944
1945 ldo -FRAME_SIZE(%r30), %r30
1946 LDREG -RP_OFFSET(%r30), %r2
1947
1948 /* FIXME: I think we need to restore a few more things here. */
1949 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1950 ldo TASK_REGS(%r1),%r1 /* get pt regs */
1951 reg_restore %r1
1952
1953 /* If the signal was received while the process was blocked on a
1954 * syscall, then r2 will take us to syscall_exit; otherwise r2 will
1955 * take us to syscall_exit_rfi and on to intr_return.
1956 */
1957 bv %r0(%r2)
1958 LDREG PT_GR28(%r1),%r28 /* reload original r28 for syscall_exit */
1959
1960 .export sys_sigaltstack_wrapper
1961 sys_sigaltstack_wrapper:
1962 /* Get the user stack pointer */
1963 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1964 ldo TASK_REGS(%r1),%r24 /* get pt regs */
1965 LDREG TASK_PT_GR30(%r24),%r24
1966 STREG %r2, -RP_OFFSET(%r30)
1967 #ifdef CONFIG_64BIT
1968 ldo FRAME_SIZE(%r30), %r30
1969 b,l do_sigaltstack,%r2
1970 ldo -16(%r30),%r29 /* Reference param save area */
1971 #else
1972 bl do_sigaltstack,%r2
1973 ldo FRAME_SIZE(%r30), %r30
1974 #endif
1975
1976 ldo -FRAME_SIZE(%r30), %r30
1977 LDREG -RP_OFFSET(%r30), %r2
1978 bv %r0(%r2)
1979 nop
1980
1981 #ifdef CONFIG_64BIT
1982 .export sys32_sigaltstack_wrapper
1983 sys32_sigaltstack_wrapper:
1984 /* Get the user stack pointer */
1985 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r24
1986 LDREG TASK_PT_GR30(%r24),%r24
1987 STREG %r2, -RP_OFFSET(%r30)
1988 ldo FRAME_SIZE(%r30), %r30
1989 b,l do_sigaltstack32,%r2
1990 ldo -16(%r30),%r29 /* Reference param save area */
1991
1992 ldo -FRAME_SIZE(%r30), %r30
1993 LDREG -RP_OFFSET(%r30), %r2
1994 bv %r0(%r2)
1995 nop
1996 #endif
1997
1998 .export sys_rt_sigsuspend_wrapper
1999 sys_rt_sigsuspend_wrapper:
2000 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
2001 ldo TASK_REGS(%r1),%r24
2002 reg_save %r24
2003
2004 STREG %r2, -RP_OFFSET(%r30)
2005 #ifdef CONFIG_64BIT
2006 ldo FRAME_SIZE(%r30), %r30
2007 b,l sys_rt_sigsuspend,%r2
2008 ldo -16(%r30),%r29 /* Reference param save area */
2009 #else
2010 bl sys_rt_sigsuspend,%r2
2011 ldo FRAME_SIZE(%r30), %r30
2012 #endif
2013
2014 ldo -FRAME_SIZE(%r30), %r30
2015 LDREG -RP_OFFSET(%r30), %r2
2016
2017 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
2018 ldo TASK_REGS(%r1),%r1
2019 reg_restore %r1
2020
2021 bv %r0(%r2)
2022 nop
2023
2024 .export syscall_exit
2025 syscall_exit:
2026
2027 /* NOTE: HP-UX syscalls also come through here
2028 * after hpux_syscall_exit fixes up return
2029 * values. */
2030
2031 /* NOTE: Not all syscalls exit this way. rt_sigreturn will exit
2032 * via syscall_exit_rfi if the signal was received while the process
2033 * was running.
2034 */
2035
2036 /* save return value now */
2037
2038 mfctl %cr30, %r1
2039 LDREG TI_TASK(%r1),%r1
2040 STREG %r28,TASK_PT_GR28(%r1)
2041
2042 #ifdef CONFIG_HPUX
2043
2044 /* <linux/personality.h> cannot be easily included */
2045 #define PER_HPUX 0x10
2046 LDREG TASK_PERSONALITY(%r1),%r19
2047
2048 /* We can't use "CMPIB<> PER_HPUX" since "im5" field is sign extended */
2049 ldo -PER_HPUX(%r19), %r19
2050 CMPIB<>,n 0,%r19,1f
2051
2052 /* Save other hpux returns if personality is PER_HPUX */
2053 STREG %r22,TASK_PT_GR22(%r1)
2054 STREG %r29,TASK_PT_GR29(%r1)
2055 1:
2056
2057 #endif /* CONFIG_HPUX */
2058
2059 /* Seems to me that dp could be wrong here, if the syscall involved
2060 * calling a module, and nothing got round to restoring dp on return.
2061 */
2062 loadgp
2063
2064 syscall_check_bh:
2065
2066 /* Check for software interrupts */
2067
2068 .import irq_stat,data
2069
2070 load32 irq_stat,%r19
2071
2072 #ifdef CONFIG_SMP
2073 /* sched.h: int processor */
2074 /* %r26 is used as scratch register to index into irq_stat[] */
2075 ldw TI_CPU-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r26 /* cpu # */
2076
2077 /* shift left ____cacheline_aligned (aka L1_CACHE_BYTES) bits */
2078 #ifdef CONFIG_64BIT
2079 shld %r26, 6, %r20
2080 #else
2081 shlw %r26, 5, %r20
2082 #endif
2083 add %r19,%r20,%r19 /* now have &irq_stat[smp_processor_id()] */
2084 #endif /* CONFIG_SMP */
2085
2086 syscall_check_resched:
2087
2088 /* check for reschedule */
2089
2090 LDREG TI_FLAGS-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r19 /* long */
2091 bb,<,n %r19, 31-TIF_NEED_RESCHED, syscall_do_resched /* forward */
2092
2093 syscall_check_sig:
2094 LDREG TI_FLAGS-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r19 /* get ti flags */
2095 bb,<,n %r19, 31-TIF_SIGPENDING, syscall_do_signal /* forward */
2096
2097 syscall_restore:
2098 /* Are we being ptraced? */
2099 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
2100
2101 LDREG TASK_PTRACE(%r1), %r19
2102 bb,< %r19,31,syscall_restore_rfi
2103 nop
2104
2105 ldo TASK_PT_FR31(%r1),%r19 /* reload fpregs */
2106 rest_fp %r19
2107
2108 LDREG TASK_PT_SAR(%r1),%r19 /* restore SAR */
2109 mtsar %r19
2110
2111 LDREG TASK_PT_GR2(%r1),%r2 /* restore user rp */
2112 LDREG TASK_PT_GR19(%r1),%r19
2113 LDREG TASK_PT_GR20(%r1),%r20
2114 LDREG TASK_PT_GR21(%r1),%r21
2115 LDREG TASK_PT_GR22(%r1),%r22
2116 LDREG TASK_PT_GR23(%r1),%r23
2117 LDREG TASK_PT_GR24(%r1),%r24
2118 LDREG TASK_PT_GR25(%r1),%r25
2119 LDREG TASK_PT_GR26(%r1),%r26
2120 LDREG TASK_PT_GR27(%r1),%r27 /* restore user dp */
2121 LDREG TASK_PT_GR28(%r1),%r28 /* syscall return value */
2122 LDREG TASK_PT_GR29(%r1),%r29
2123 LDREG TASK_PT_GR31(%r1),%r31 /* restore syscall rp */
2124
2125 /* NOTE: We use rsm/ssm pair to make this operation atomic */
2126 rsm PSW_SM_I, %r0
2127 LDREG TASK_PT_GR30(%r1),%r30 /* restore user sp */
2128 mfsp %sr3,%r1 /* Get users space id */
2129 mtsp %r1,%sr7 /* Restore sr7 */
2130 ssm PSW_SM_I, %r0
2131
2132 /* Set sr2 to zero for userspace syscalls to work. */
2133 mtsp %r0,%sr2
2134 mtsp %r1,%sr4 /* Restore sr4 */
2135 mtsp %r1,%sr5 /* Restore sr5 */
2136 mtsp %r1,%sr6 /* Restore sr6 */
2137
2138 depi 3,31,2,%r31 /* ensure return to user mode. */
2139
2140 #ifdef CONFIG_64BIT
2141 /* decide whether to reset the wide mode bit
2142 *
2143 * For a syscall, the W bit is stored in the lowest bit
2144 * of sp. Extract it and reset W if it is zero */
2145 extrd,u,*<> %r30,63,1,%r1
2146 rsm PSW_SM_W, %r0
2147 /* now reset the lowest bit of sp if it was set */
2148 xor %r30,%r1,%r30
2149 #endif
2150 be,n 0(%sr3,%r31) /* return to user space */
2151
2152 /* We have to return via an RFI, so that PSW T and R bits can be set
2153 * appropriately.
2154 * This sets up pt_regs so we can return via intr_restore, which is not
2155 * the most efficient way of doing things, but it works.
2156 */
2157 syscall_restore_rfi:
2158 ldo -1(%r0),%r2 /* Set recovery cntr to -1 */
2159 mtctl %r2,%cr0 /* for immediate trap */
2160 LDREG TASK_PT_PSW(%r1),%r2 /* Get old PSW */
2161 ldi 0x0b,%r20 /* Create new PSW */
2162 depi -1,13,1,%r20 /* C, Q, D, and I bits */
2163
2164 /* The values of PA_SINGLESTEP_BIT and PA_BLOCKSTEP_BIT are
2165 * set in include/linux/ptrace.h and converted to PA bitmap
2166 * numbers in asm-offsets.c */
2167
2168 /* if ((%r19.PA_SINGLESTEP_BIT)) { %r20.27=1} */
2169 extru,= %r19,PA_SINGLESTEP_BIT,1,%r0
2170 depi -1,27,1,%r20 /* R bit */
2171
2172 /* if ((%r19.PA_BLOCKSTEP_BIT)) { %r20.7=1} */
2173 extru,= %r19,PA_BLOCKSTEP_BIT,1,%r0
2174 depi -1,7,1,%r20 /* T bit */
2175
2176 STREG %r20,TASK_PT_PSW(%r1)
2177
2178 /* Always store space registers, since sr3 can be changed (e.g. fork) */
2179
2180 mfsp %sr3,%r25
2181 STREG %r25,TASK_PT_SR3(%r1)
2182 STREG %r25,TASK_PT_SR4(%r1)
2183 STREG %r25,TASK_PT_SR5(%r1)
2184 STREG %r25,TASK_PT_SR6(%r1)
2185 STREG %r25,TASK_PT_SR7(%r1)
2186 STREG %r25,TASK_PT_IASQ0(%r1)
2187 STREG %r25,TASK_PT_IASQ1(%r1)
2188
2189 /* XXX W bit??? */
2190 /* Now if old D bit is clear, it means we didn't save all registers
2191 * on syscall entry, so do that now. This only happens on TRACEME
2192 * calls, or if someone attached to us while we were on a syscall.
2193 * We could make this more efficient by not saving r3-r18, but
2194 * then we wouldn't be able to use the common intr_restore path.
2195 * It is only for traced processes anyway, so performance is not
2196 * an issue.
2197 */
2198 bb,< %r2,30,pt_regs_ok /* Branch if D set */
2199 ldo TASK_REGS(%r1),%r25
2200 reg_save %r25 /* Save r3 to r18 */
2201
2202 /* Save the current sr */
2203 mfsp %sr0,%r2
2204 STREG %r2,TASK_PT_SR0(%r1)
2205
2206 /* Save the scratch sr */
2207 mfsp %sr1,%r2
2208 STREG %r2,TASK_PT_SR1(%r1)
2209
2210 /* sr2 should be set to zero for userspace syscalls */
2211 STREG %r0,TASK_PT_SR2(%r1)
2212
2213 pt_regs_ok:
2214 LDREG TASK_PT_GR31(%r1),%r2
2215 depi 3,31,2,%r2 /* ensure return to user mode. */
2216 STREG %r2,TASK_PT_IAOQ0(%r1)
2217 ldo 4(%r2),%r2
2218 STREG %r2,TASK_PT_IAOQ1(%r1)
2219 copy %r25,%r16
2220 b intr_restore
2221 nop
2222
2223 .import schedule,code
2224 syscall_do_resched:
2225 BL schedule,%r2
2226 #ifdef CONFIG_64BIT
2227 ldo -16(%r30),%r29 /* Reference param save area */
2228 #else
2229 nop
2230 #endif
2231 b syscall_check_bh /* if resched, we start over again */
2232 nop
2233
2234 .import do_signal,code
2235 syscall_do_signal:
2236 /* Save callee-save registers (for sigcontext).
2237 FIXME: After this point the process structure should be
2238 consistent with all the relevant state of the process
2239 before the syscall. We need to verify this. */
2240 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
2241 ldo TASK_REGS(%r1), %r25 /* struct pt_regs *regs */
2242 reg_save %r25
2243
2244 ldi 1, %r24 /* unsigned long in_syscall */
2245
2246 #ifdef CONFIG_64BIT
2247 ldo -16(%r30),%r29 /* Reference param save area */
2248 #endif
2249 BL do_signal,%r2
2250 copy %r0, %r26 /* sigset_t *oldset = NULL */
2251
2252 LDREG TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
2253 ldo TASK_REGS(%r1), %r20 /* reload pt_regs */
2254 reg_restore %r20
2255
2256 b,n syscall_check_sig
2257
2258 /*
2259 * get_register is used by the non access tlb miss handlers to
2260 * copy the value of the general register specified in r8 into
2261 * r1. This routine can't be used for shadowed registers, since
2262 * the rfir will restore the original value. So, for the shadowed
2263 * registers we put a -1 into r1 to indicate that the register
2264 * should not be used (the register being copied could also have
2265 * a -1 in it, but that is OK, it just means that we will have
2266 * to use the slow path instead).
2267 */
2268
2269 get_register:
2270 blr %r8,%r0
2271 nop
2272 bv %r0(%r25) /* r0 */
2273 copy %r0,%r1
2274 bv %r0(%r25) /* r1 - shadowed */
2275 ldi -1,%r1
2276 bv %r0(%r25) /* r2 */
2277 copy %r2,%r1
2278 bv %r0(%r25) /* r3 */
2279 copy %r3,%r1
2280 bv %r0(%r25) /* r4 */
2281 copy %r4,%r1
2282 bv %r0(%r25) /* r5 */
2283 copy %r5,%r1
2284 bv %r0(%r25) /* r6 */
2285 copy %r6,%r1
2286 bv %r0(%r25) /* r7 */
2287 copy %r7,%r1
2288 bv %r0(%r25) /* r8 - shadowed */
2289 ldi -1,%r1
2290 bv %r0(%r25) /* r9 - shadowed */
2291 ldi -1,%r1
2292 bv %r0(%r25) /* r10 */
2293 copy %r10,%r1
2294 bv %r0(%r25) /* r11 */
2295 copy %r11,%r1
2296 bv %r0(%r25) /* r12 */
2297 copy %r12,%r1
2298 bv %r0(%r25) /* r13 */
2299 copy %r13,%r1
2300 bv %r0(%r25) /* r14 */
2301 copy %r14,%r1
2302 bv %r0(%r25) /* r15 */
2303 copy %r15,%r1
2304 bv %r0(%r25) /* r16 - shadowed */
2305 ldi -1,%r1
2306 bv %r0(%r25) /* r17 - shadowed */
2307 ldi -1,%r1
2308 bv %r0(%r25) /* r18 */
2309 copy %r18,%r1
2310 bv %r0(%r25) /* r19 */
2311 copy %r19,%r1
2312 bv %r0(%r25) /* r20 */
2313 copy %r20,%r1
2314 bv %r0(%r25) /* r21 */
2315 copy %r21,%r1
2316 bv %r0(%r25) /* r22 */
2317 copy %r22,%r1
2318 bv %r0(%r25) /* r23 */
2319 copy %r23,%r1
2320 bv %r0(%r25) /* r24 - shadowed */
2321 ldi -1,%r1
2322 bv %r0(%r25) /* r25 - shadowed */
2323 ldi -1,%r1
2324 bv %r0(%r25) /* r26 */
2325 copy %r26,%r1
2326 bv %r0(%r25) /* r27 */
2327 copy %r27,%r1
2328 bv %r0(%r25) /* r28 */
2329 copy %r28,%r1
2330 bv %r0(%r25) /* r29 */
2331 copy %r29,%r1
2332 bv %r0(%r25) /* r30 */
2333 copy %r30,%r1
2334 bv %r0(%r25) /* r31 */
2335 copy %r31,%r1
2336
2337 /*
2338 * set_register is used by the non access tlb miss handlers to
2339 * copy the value of r1 into the general register specified in
2340 * r8.
2341 */
2342
2343 set_register:
2344 blr %r8,%r0
2345 nop
2346 bv %r0(%r25) /* r0 (silly, but it is a place holder) */
2347 copy %r1,%r0
2348 bv %r0(%r25) /* r1 */
2349 copy %r1,%r1
2350 bv %r0(%r25) /* r2 */
2351 copy %r1,%r2
2352 bv %r0(%r25) /* r3 */
2353 copy %r1,%r3
2354 bv %r0(%r25) /* r4 */
2355 copy %r1,%r4
2356 bv %r0(%r25) /* r5 */
2357 copy %r1,%r5
2358 bv %r0(%r25) /* r6 */
2359 copy %r1,%r6
2360 bv %r0(%r25) /* r7 */
2361 copy %r1,%r7
2362 bv %r0(%r25) /* r8 */
2363 copy %r1,%r8
2364 bv %r0(%r25) /* r9 */
2365 copy %r1,%r9
2366 bv %r0(%r25) /* r10 */
2367 copy %r1,%r10
2368 bv %r0(%r25) /* r11 */
2369 copy %r1,%r11
2370 bv %r0(%r25) /* r12 */
2371 copy %r1,%r12
2372 bv %r0(%r25) /* r13 */
2373 copy %r1,%r13
2374 bv %r0(%r25) /* r14 */
2375 copy %r1,%r14
2376 bv %r0(%r25) /* r15 */
2377 copy %r1,%r15
2378 bv %r0(%r25) /* r16 */
2379 copy %r1,%r16
2380 bv %r0(%r25) /* r17 */
2381 copy %r1,%r17
2382 bv %r0(%r25) /* r18 */
2383 copy %r1,%r18
2384 bv %r0(%r25) /* r19 */
2385 copy %r1,%r19
2386 bv %r0(%r25) /* r20 */
2387 copy %r1,%r20
2388 bv %r0(%r25) /* r21 */
2389 copy %r1,%r21
2390 bv %r0(%r25) /* r22 */
2391 copy %r1,%r22
2392 bv %r0(%r25) /* r23 */
2393 copy %r1,%r23
2394 bv %r0(%r25) /* r24 */
2395 copy %r1,%r24
2396 bv %r0(%r25) /* r25 */
2397 copy %r1,%r25
2398 bv %r0(%r25) /* r26 */
2399 copy %r1,%r26
2400 bv %r0(%r25) /* r27 */
2401 copy %r1,%r27
2402 bv %r0(%r25) /* r28 */
2403 copy %r1,%r28
2404 bv %r0(%r25) /* r29 */
2405 copy %r1,%r29
2406 bv %r0(%r25) /* r30 */
2407 copy %r1,%r30
2408 bv %r0(%r25) /* r31 */
2409 copy %r1,%r31