Blame view
arch/arm/lib/copy_page.S
1.26 KB
1da177e4c
|
1 2 3 4 5 6 7 8 9 10 11 12 13 |
/* * linux/arch/arm/lib/copypage.S * * Copyright (C) 1995-1999 Russell King * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. * * ASM optimised string functions */ #include <linux/linkage.h> #include <asm/assembler.h> |
e6ae744dd
|
14 |
#include <asm/asm-offsets.h> |
dca230f00
|
15 |
#include <asm/cache.h> |
1da177e4c
|
16 |
|
dca230f00
|
17 |
#define COPY_COUNT (PAGE_SZ / (2 * L1_CACHE_BYTES) PLD( -1 )) |
1da177e4c
|
18 19 20 21 22 23 24 25 26 27 28 29 |
.text .align 5 /* * StrongARM optimised copy_page routine * now 1.78bytes/cycle, was 1.60 bytes/cycle (50MHz bus -> 89MB/s) * Note that we probably achieve closer to the 100MB/s target with * the core clock switching. */ ENTRY(copy_page) stmfd sp!, {r4, lr} @ 2 PLD( pld [r1, #0] ) |
dca230f00
|
30 |
PLD( pld [r1, #L1_CACHE_BYTES] ) |
1da177e4c
|
31 32 |
mov r2, #COPY_COUNT @ 1 ldmia r1!, {r3, r4, ip, lr} @ 4+1 |
dca230f00
|
33 34 35 36 |
1: PLD( pld [r1, #2 * L1_CACHE_BYTES]) PLD( pld [r1, #3 * L1_CACHE_BYTES]) 2: .rept (2 * L1_CACHE_BYTES / 16 - 1) |
1da177e4c
|
37 38 |
stmia r0!, {r3, r4, ip, lr} @ 4 ldmia r1!, {r3, r4, ip, lr} @ 4 |
dca230f00
|
39 |
.endr |
1da177e4c
|
40 41 42 43 44 45 |
subs r2, r2, #1 @ 1 stmia r0!, {r3, r4, ip, lr} @ 4 ldmgtia r1!, {r3, r4, ip, lr} @ 4 bgt 1b @ 1 PLD( ldmeqia r1!, {r3, r4, ip, lr} ) PLD( beq 2b ) |
1b93a7175
|
46 |
ldmfd sp!, {r4, pc} @ 3 |
93ed39701
|
47 |
ENDPROC(copy_page) |