]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
ppc64: use copyuser_64.S from powerpc/lib
authorStephen Rothwell <sfr@canb.auug.org.au>
Thu, 27 Oct 2005 06:40:05 +0000 (16:40 +1000)
committerStephen Rothwell <sfr@canb.auug.org.au>
Fri, 28 Oct 2005 02:53:37 +0000 (12:53 +1000)
since it is identical to copyuser.S from ppc64/lib.

Signed-off-by: Stephen Rothwell <sfr@canb.auug.org.au>
arch/powerpc/lib/Makefile
arch/ppc64/lib/Makefile
arch/ppc64/lib/copyuser.S [deleted file]

index 5e481944250eca60b9f503de6b24329e872e5163..3474c9309e2f345904a844afd95638fb52b6fe20 100644 (file)
@@ -5,12 +5,12 @@
 ifeq ($(CONFIG_PPC_MERGE),y)
 obj-y                  := strcase.o string.o
 obj-$(CONFIG_PPC32)    += div64.o copy_32.o checksum_32.o
-obj-$(CONFIG_PPC64)    += copyuser_64.o memcpy_64.o \
+obj-$(CONFIG_PPC64)    += memcpy_64.o \
                           usercopy_64.o sstep.o mem_64.o
 ifeq ($(CONFIG_PPC64),y)
 obj-$(CONFIG_SMP)      += locks.o
 endif
 endif
 
-obj-$(CONFIG_PPC64)    += checksum_64.o copypage_64.o
+obj-$(CONFIG_PPC64)    += checksum_64.o copypage_64.o copyuser_64.o
 obj-$(CONFIG_PPC_ISERIES) += e2a.o
index 820021885e1a2d7f0a938deb3100b737befe5e64..f29e76559950e9b98fc8d481187b2b6dbdb84791 100644 (file)
@@ -3,7 +3,7 @@
 #
 
 lib-y := string.o strcase.o
-lib-y += memcpy.o copyuser.o usercopy.o
+lib-y += memcpy.o usercopy.o
 
 # Lock primitives are defined as no-ops in include/linux/spinlock.h
 # for non-SMP configs. Don't build the real versions.
diff --git a/arch/ppc64/lib/copyuser.S b/arch/ppc64/lib/copyuser.S
deleted file mode 100644 (file)
index a0b3fbb..0000000
+++ /dev/null
@@ -1,576 +0,0 @@
-/*
- * arch/ppc64/lib/copyuser.S
- *
- * Copyright (C) 2002 Paul Mackerras, IBM Corp.
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License
- * as published by the Free Software Foundation; either version
- * 2 of the License, or (at your option) any later version.
- */
-#include <asm/processor.h>
-#include <asm/ppc_asm.h>
-
-       .align  7
-_GLOBAL(__copy_tofrom_user)
-       /* first check for a whole page copy on a page boundary */
-       cmpldi  cr1,r5,16
-       cmpdi   cr6,r5,4096
-       or      r0,r3,r4
-       neg     r6,r3           /* LS 3 bits = # bytes to 8-byte dest bdry */
-       andi.   r0,r0,4095
-       std     r3,-24(r1)
-       crand   cr0*4+2,cr0*4+2,cr6*4+2
-       std     r4,-16(r1)
-       std     r5,-8(r1)
-       dcbt    0,r4
-       beq     .Lcopy_page
-       andi.   r6,r6,7
-       mtcrf   0x01,r5
-       blt     cr1,.Lshort_copy
-       bne     .Ldst_unaligned
-.Ldst_aligned:
-       andi.   r0,r4,7
-       addi    r3,r3,-16
-       bne     .Lsrc_unaligned
-       srdi    r7,r5,4
-20:    ld      r9,0(r4)
-       addi    r4,r4,-8
-       mtctr   r7
-       andi.   r5,r5,7
-       bf      cr7*4+0,22f
-       addi    r3,r3,8
-       addi    r4,r4,8
-       mr      r8,r9
-       blt     cr1,72f
-21:    ld      r9,8(r4)
-70:    std     r8,8(r3)
-22:    ldu     r8,16(r4)
-71:    stdu    r9,16(r3)
-       bdnz    21b
-72:    std     r8,8(r3)
-       beq+    3f
-       addi    r3,r3,16
-23:    ld      r9,8(r4)
-.Ldo_tail:
-       bf      cr7*4+1,1f
-       rotldi  r9,r9,32
-73:    stw     r9,0(r3)
-       addi    r3,r3,4
-1:     bf      cr7*4+2,2f
-       rotldi  r9,r9,16
-74:    sth     r9,0(r3)
-       addi    r3,r3,2
-2:     bf      cr7*4+3,3f
-       rotldi  r9,r9,8
-75:    stb     r9,0(r3)
-3:     li      r3,0
-       blr
-
-.Lsrc_unaligned:
-       srdi    r6,r5,3
-       addi    r5,r5,-16
-       subf    r4,r0,r4
-       srdi    r7,r5,4
-       sldi    r10,r0,3
-       cmpldi  cr6,r6,3
-       andi.   r5,r5,7
-       mtctr   r7
-       subfic  r11,r10,64
-       add     r5,r5,r0
-       bt      cr7*4+0,28f
-
-24:    ld      r9,0(r4)        /* 3+2n loads, 2+2n stores */
-25:    ld      r0,8(r4)
-       sld     r6,r9,r10
-26:    ldu     r9,16(r4)
-       srd     r7,r0,r11
-       sld     r8,r0,r10
-       or      r7,r7,r6
-       blt     cr6,79f
-27:    ld      r0,8(r4)
-       b       2f
-
-28:    ld      r0,0(r4)        /* 4+2n loads, 3+2n stores */
-29:    ldu     r9,8(r4)
-       sld     r8,r0,r10
-       addi    r3,r3,-8
-       blt     cr6,5f
-30:    ld      r0,8(r4)
-       srd     r12,r9,r11
-       sld     r6,r9,r10
-31:    ldu     r9,16(r4)
-       or      r12,r8,r12
-       srd     r7,r0,r11
-       sld     r8,r0,r10
-       addi    r3,r3,16
-       beq     cr6,78f
-
-1:     or      r7,r7,r6
-32:    ld      r0,8(r4)
-76:    std     r12,8(r3)
-2:     srd     r12,r9,r11
-       sld     r6,r9,r10
-33:    ldu     r9,16(r4)
-       or      r12,r8,r12
-77:    stdu    r7,16(r3)
-       srd     r7,r0,r11
-       sld     r8,r0,r10
-       bdnz    1b
-
-78:    std     r12,8(r3)
-       or      r7,r7,r6
-79:    std     r7,16(r3)
-5:     srd     r12,r9,r11
-       or      r12,r8,r12
-80:    std     r12,24(r3)
-       bne     6f
-       li      r3,0
-       blr
-6:     cmpwi   cr1,r5,8
-       addi    r3,r3,32
-       sld     r9,r9,r10
-       ble     cr1,.Ldo_tail
-34:    ld      r0,8(r4)
-       srd     r7,r0,r11
-       or      r9,r7,r9
-       b       .Ldo_tail
-
-.Ldst_unaligned:
-       mtcrf   0x01,r6         /* put #bytes to 8B bdry into cr7 */
-       subf    r5,r6,r5
-       li      r7,0
-       cmpldi  r1,r5,16
-       bf      cr7*4+3,1f
-35:    lbz     r0,0(r4)
-81:    stb     r0,0(r3)
-       addi    r7,r7,1
-1:     bf      cr7*4+2,2f
-36:    lhzx    r0,r7,r4
-82:    sthx    r0,r7,r3
-       addi    r7,r7,2
-2:     bf      cr7*4+1,3f
-37:    lwzx    r0,r7,r4
-83:    stwx    r0,r7,r3
-3:     mtcrf   0x01,r5
-       add     r4,r6,r4
-       add     r3,r6,r3
-       b       .Ldst_aligned
-
-.Lshort_copy:
-       bf      cr7*4+0,1f
-38:    lwz     r0,0(r4)
-39:    lwz     r9,4(r4)
-       addi    r4,r4,8
-84:    stw     r0,0(r3)
-85:    stw     r9,4(r3)
-       addi    r3,r3,8
-1:     bf      cr7*4+1,2f
-40:    lwz     r0,0(r4)
-       addi    r4,r4,4
-86:    stw     r0,0(r3)
-       addi    r3,r3,4
-2:     bf      cr7*4+2,3f
-41:    lhz     r0,0(r4)
-       addi    r4,r4,2
-87:    sth     r0,0(r3)
-       addi    r3,r3,2
-3:     bf      cr7*4+3,4f
-42:    lbz     r0,0(r4)
-88:    stb     r0,0(r3)
-4:     li      r3,0
-       blr
-
-/*
- * exception handlers follow
- * we have to return the number of bytes not copied
- * for an exception on a load, we set the rest of the destination to 0
- */
-
-136:
-137:
-       add     r3,r3,r7
-       b       1f
-130:
-131:
-       addi    r3,r3,8
-120:
-122:
-124:
-125:
-126:
-127:
-128:
-129:
-133:
-       addi    r3,r3,8
-121:
-132:
-       addi    r3,r3,8
-123:
-134:
-135:
-138:
-139:
-140:
-141:
-142:
-
-/*
- * here we have had a fault on a load and r3 points to the first
- * unmodified byte of the destination
- */
-1:     ld      r6,-24(r1)
-       ld      r4,-16(r1)
-       ld      r5,-8(r1)
-       subf    r6,r6,r3
-       add     r4,r4,r6
-       subf    r5,r6,r5        /* #bytes left to go */
-
-/*
- * first see if we can copy any more bytes before hitting another exception
- */
-       mtctr   r5
-43:    lbz     r0,0(r4)
-       addi    r4,r4,1
-89:    stb     r0,0(r3)
-       addi    r3,r3,1
-       bdnz    43b
-       li      r3,0            /* huh? all copied successfully this time? */
-       blr
-
-/*
- * here we have trapped again, need to clear ctr bytes starting at r3
- */
-143:   mfctr   r5
-       li      r0,0
-       mr      r4,r3
-       mr      r3,r5           /* return the number of bytes not copied */
-1:     andi.   r9,r4,7
-       beq     3f
-90:    stb     r0,0(r4)
-       addic.  r5,r5,-1
-       addi    r4,r4,1
-       bne     1b
-       blr
-3:     cmpldi  cr1,r5,8
-       srdi    r9,r5,3
-       andi.   r5,r5,7
-       blt     cr1,93f
-       mtctr   r9
-91:    std     r0,0(r4)
-       addi    r4,r4,8
-       bdnz    91b
-93:    beqlr
-       mtctr   r5      
-92:    stb     r0,0(r4)
-       addi    r4,r4,1
-       bdnz    92b
-       blr
-
-/*
- * exception handlers for stores: we just need to work
- * out how many bytes weren't copied
- */
-182:
-183:
-       add     r3,r3,r7
-       b       1f
-180:
-       addi    r3,r3,8
-171:
-177:
-       addi    r3,r3,8
-170:
-172:
-176:
-178:
-       addi    r3,r3,4
-185:
-       addi    r3,r3,4
-173:
-174:
-175:
-179:
-181:
-184:
-186:
-187:
-188:
-189:   
-1:
-       ld      r6,-24(r1)
-       ld      r5,-8(r1)
-       add     r6,r6,r5
-       subf    r3,r3,r6        /* #bytes not copied */
-190:
-191:
-192:
-       blr                     /* #bytes not copied in r3 */
-
-       .section __ex_table,"a"
-       .align  3
-       .llong  20b,120b
-       .llong  21b,121b
-       .llong  70b,170b
-       .llong  22b,122b
-       .llong  71b,171b
-       .llong  72b,172b
-       .llong  23b,123b
-       .llong  73b,173b
-       .llong  74b,174b
-       .llong  75b,175b
-       .llong  24b,124b
-       .llong  25b,125b
-       .llong  26b,126b
-       .llong  27b,127b
-       .llong  28b,128b
-       .llong  29b,129b
-       .llong  30b,130b
-       .llong  31b,131b
-       .llong  32b,132b
-       .llong  76b,176b
-       .llong  33b,133b
-       .llong  77b,177b
-       .llong  78b,178b
-       .llong  79b,179b
-       .llong  80b,180b
-       .llong  34b,134b
-       .llong  35b,135b
-       .llong  81b,181b
-       .llong  36b,136b
-       .llong  82b,182b
-       .llong  37b,137b
-       .llong  83b,183b
-       .llong  38b,138b
-       .llong  39b,139b
-       .llong  84b,184b
-       .llong  85b,185b
-       .llong  40b,140b
-       .llong  86b,186b
-       .llong  41b,141b
-       .llong  87b,187b
-       .llong  42b,142b
-       .llong  88b,188b
-       .llong  43b,143b
-       .llong  89b,189b
-       .llong  90b,190b
-       .llong  91b,191b
-       .llong  92b,192b
-       
-       .text
-
-/*
- * Routine to copy a whole page of data, optimized for POWER4.
- * On POWER4 it is more than 50% faster than the simple loop
- * above (following the .Ldst_aligned label) but it runs slightly
- * slower on POWER3.
- */
-.Lcopy_page:
-       std     r31,-32(1)
-       std     r30,-40(1)
-       std     r29,-48(1)
-       std     r28,-56(1)
-       std     r27,-64(1)
-       std     r26,-72(1)
-       std     r25,-80(1)
-       std     r24,-88(1)
-       std     r23,-96(1)
-       std     r22,-104(1)
-       std     r21,-112(1)
-       std     r20,-120(1)
-       li      r5,4096/32 - 1
-       addi    r3,r3,-8
-       li      r0,5
-0:     addi    r5,r5,-24
-       mtctr   r0
-20:    ld      r22,640(4)
-21:    ld      r21,512(4)
-22:    ld      r20,384(4)
-23:    ld      r11,256(4)
-24:    ld      r9,128(4)
-25:    ld      r7,0(4)
-26:    ld      r25,648(4)
-27:    ld      r24,520(4)
-28:    ld      r23,392(4)
-29:    ld      r10,264(4)
-30:    ld      r8,136(4)
-31:    ldu     r6,8(4)
-       cmpwi   r5,24
-1:
-32:    std     r22,648(3)
-33:    std     r21,520(3)
-34:    std     r20,392(3)
-35:    std     r11,264(3)
-36:    std     r9,136(3)
-37:    std     r7,8(3)
-38:    ld      r28,648(4)
-39:    ld      r27,520(4)
-40:    ld      r26,392(4)
-41:    ld      r31,264(4)
-42:    ld      r30,136(4)
-43:    ld      r29,8(4)
-44:    std     r25,656(3)
-45:    std     r24,528(3)
-46:    std     r23,400(3)
-47:    std     r10,272(3)
-48:    std     r8,144(3)
-49:    std     r6,16(3)
-50:    ld      r22,656(4)
-51:    ld      r21,528(4)
-52:    ld      r20,400(4)
-53:    ld      r11,272(4)
-54:    ld      r9,144(4)
-55:    ld      r7,16(4)
-56:    std     r28,664(3)
-57:    std     r27,536(3)
-58:    std     r26,408(3)
-59:    std     r31,280(3)
-60:    std     r30,152(3)
-61:    stdu    r29,24(3)
-62:    ld      r25,664(4)
-63:    ld      r24,536(4)
-64:    ld      r23,408(4)
-65:    ld      r10,280(4)
-66:    ld      r8,152(4)
-67:    ldu     r6,24(4)
-       bdnz    1b
-68:    std     r22,648(3)
-69:    std     r21,520(3)
-70:    std     r20,392(3)
-71:    std     r11,264(3)
-72:    std     r9,136(3)
-73:    std     r7,8(3)
-74:    addi    r4,r4,640
-75:    addi    r3,r3,648
-       bge     0b
-       mtctr   r5
-76:    ld      r7,0(4)
-77:    ld      r8,8(4)
-78:    ldu     r9,16(4)
-3:
-79:    ld      r10,8(4)
-80:    std     r7,8(3)
-81:    ld      r7,16(4)
-82:    std     r8,16(3)
-83:    ld      r8,24(4)
-84:    std     r9,24(3)
-85:    ldu     r9,32(4)
-86:    stdu    r10,32(3)
-       bdnz    3b
-4:
-87:    ld      r10,8(4)
-88:    std     r7,8(3)
-89:    std     r8,16(3)
-90:    std     r9,24(3)
-91:    std     r10,32(3)
-9:     ld      r20,-120(1)
-       ld      r21,-112(1)
-       ld      r22,-104(1)
-       ld      r23,-96(1)
-       ld      r24,-88(1)
-       ld      r25,-80(1)
-       ld      r26,-72(1)
-       ld      r27,-64(1)
-       ld      r28,-56(1)
-       ld      r29,-48(1)
-       ld      r30,-40(1)
-       ld      r31,-32(1)
-       li      r3,0
-       blr
-
-/*
- * on an exception, reset to the beginning and jump back into the
- * standard __copy_tofrom_user
- */
-100:   ld      r20,-120(1)
-       ld      r21,-112(1)
-       ld      r22,-104(1)
-       ld      r23,-96(1)
-       ld      r24,-88(1)
-       ld      r25,-80(1)
-       ld      r26,-72(1)
-       ld      r27,-64(1)
-       ld      r28,-56(1)
-       ld      r29,-48(1)
-       ld      r30,-40(1)
-       ld      r31,-32(1)
-       ld      r3,-24(r1)
-       ld      r4,-16(r1)
-       li      r5,4096
-       b       .Ldst_aligned
-
-       .section __ex_table,"a"
-       .align  3
-       .llong  20b,100b
-       .llong  21b,100b
-       .llong  22b,100b
-       .llong  23b,100b
-       .llong  24b,100b
-       .llong  25b,100b
-       .llong  26b,100b
-       .llong  27b,100b
-       .llong  28b,100b
-       .llong  29b,100b
-       .llong  30b,100b
-       .llong  31b,100b
-       .llong  32b,100b
-       .llong  33b,100b
-       .llong  34b,100b
-       .llong  35b,100b
-       .llong  36b,100b
-       .llong  37b,100b
-       .llong  38b,100b
-       .llong  39b,100b
-       .llong  40b,100b
-       .llong  41b,100b
-       .llong  42b,100b
-       .llong  43b,100b
-       .llong  44b,100b
-       .llong  45b,100b
-       .llong  46b,100b
-       .llong  47b,100b
-       .llong  48b,100b
-       .llong  49b,100b
-       .llong  50b,100b
-       .llong  51b,100b
-       .llong  52b,100b
-       .llong  53b,100b
-       .llong  54b,100b
-       .llong  55b,100b
-       .llong  56b,100b
-       .llong  57b,100b
-       .llong  58b,100b
-       .llong  59b,100b
-       .llong  60b,100b
-       .llong  61b,100b
-       .llong  62b,100b
-       .llong  63b,100b
-       .llong  64b,100b
-       .llong  65b,100b
-       .llong  66b,100b
-       .llong  67b,100b
-       .llong  68b,100b
-       .llong  69b,100b
-       .llong  70b,100b
-       .llong  71b,100b
-       .llong  72b,100b
-       .llong  73b,100b
-       .llong  74b,100b
-       .llong  75b,100b
-       .llong  76b,100b
-       .llong  77b,100b
-       .llong  78b,100b
-       .llong  79b,100b
-       .llong  80b,100b
-       .llong  81b,100b
-       .llong  82b,100b
-       .llong  83b,100b
-       .llong  84b,100b
-       .llong  85b,100b
-       .llong  86b,100b
-       .llong  87b,100b
-       .llong  88b,100b
-       .llong  89b,100b
-       .llong  90b,100b
-       .llong  91b,100b