123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758 |
- /*
- * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com)
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License version 2 as
- * published by the Free Software Foundation.
- *
- * vineetg: June 2010
- * -__clear_user( ) called multiple times during elf load was byte loop
- * converted to do as much word clear as possible.
- *
- * vineetg: Dec 2009
- * -Hand crafted constant propagation for "constant" copy sizes
- * -stock kernel shrunk by 33K at -O3
- *
- * vineetg: Sept 2009
- * -Added option to (UN)inline copy_(to|from)_user to reduce code sz
- * -kernel shrunk by 200K even at -O3 (gcc 4.2.1)
- * -Enabled when doing -Os
- *
- * Amit Bhor, Sameer Dhavale: Codito Technologies 2004
- */
- #ifndef _ASM_ARC_UACCESS_H
- #define _ASM_ARC_UACCESS_H
- #include <linux/sched.h>
- #include <asm/errno.h>
- #include <linux/string.h> /* for generic string functions */
- #define __kernel_ok (segment_eq(get_fs(), KERNEL_DS))
- /*
- * Algorthmically, for __user_ok() we want do:
- * (start < TASK_SIZE) && (start+len < TASK_SIZE)
- * where TASK_SIZE could either be retrieved from thread_info->addr_limit or
- * emitted directly in code.
- *
- * This can however be rewritten as follows:
- * (len <= TASK_SIZE) && (start+len < TASK_SIZE)
- *
- * Because it essentially checks if buffer end is within limit and @len is
- * non-ngeative, which implies that buffer start will be within limit too.
- *
- * The reason for rewriting being, for majority of cases, @len is generally
- * compile time constant, causing first sub-expression to be compile time
- * subsumed.
- *
- * The second part would generate weird large LIMMs e.g. (0x6000_0000 - 0x10),
- * so we check for TASK_SIZE using get_fs() since the addr_limit load from mem
- * would already have been done at this call site for __kernel_ok()
- *
- */
- #define __user_ok(addr, sz) (((sz) <= TASK_SIZE) && \
- ((addr) <= (get_fs() - (sz))))
- #define __access_ok(addr, sz) (unlikely(__kernel_ok) || \
- likely(__user_ok((addr), (sz))))
- /*********** Single byte/hword/word copies ******************/
- #define __get_user_fn(sz, u, k) \
- ({ \
- long __ret = 0; /* success by default */ \
- switch (sz) { \
- case 1: __arc_get_user_one(*(k), u, "ldb", __ret); break; \
- case 2: __arc_get_user_one(*(k), u, "ldw", __ret); break; \
- case 4: __arc_get_user_one(*(k), u, "ld", __ret); break; \
- case 8: __arc_get_user_one_64(*(k), u, __ret); break; \
- } \
- __ret; \
- })
- /*
- * Returns 0 on success, -EFAULT if not.
- * @ret already contains 0 - given that errors will be less likely
- * (hence +r asm constraint below).
- * In case of error, fixup code will make it -EFAULT
- */
- #define __arc_get_user_one(dst, src, op, ret) \
- __asm__ __volatile__( \
- "1: "op" %1,[%2]\n" \
- "2: ;nop\n" \
- " .section .fixup, \"ax\"\n" \
- " .align 4\n" \
- "3: # return -EFAULT\n" \
- " mov %0, %3\n" \
- " # zero out dst ptr\n" \
- " mov %1, 0\n" \
- " j 2b\n" \
- " .previous\n" \
- " .section __ex_table, \"a\"\n" \
- " .align 4\n" \
- " .word 1b,3b\n" \
- " .previous\n" \
- \
- : "+r" (ret), "=r" (dst) \
- : "r" (src), "ir" (-EFAULT))
- #define __arc_get_user_one_64(dst, src, ret) \
- __asm__ __volatile__( \
- "1: ld %1,[%2]\n" \
- "4: ld %R1,[%2, 4]\n" \
- "2: ;nop\n" \
- " .section .fixup, \"ax\"\n" \
- " .align 4\n" \
- "3: # return -EFAULT\n" \
- " mov %0, %3\n" \
- " # zero out dst ptr\n" \
- " mov %1, 0\n" \
- " mov %R1, 0\n" \
- " j 2b\n" \
- " .previous\n" \
- " .section __ex_table, \"a\"\n" \
- " .align 4\n" \
- " .word 1b,3b\n" \
- " .word 4b,3b\n" \
- " .previous\n" \
- \
- : "+r" (ret), "=r" (dst) \
- : "r" (src), "ir" (-EFAULT))
- #define __put_user_fn(sz, u, k) \
- ({ \
- long __ret = 0; /* success by default */ \
- switch (sz) { \
- case 1: __arc_put_user_one(*(k), u, "stb", __ret); break; \
- case 2: __arc_put_user_one(*(k), u, "stw", __ret); break; \
- case 4: __arc_put_user_one(*(k), u, "st", __ret); break; \
- case 8: __arc_put_user_one_64(*(k), u, __ret); break; \
- } \
- __ret; \
- })
- #define __arc_put_user_one(src, dst, op, ret) \
- __asm__ __volatile__( \
- "1: "op" %1,[%2]\n" \
- "2: ;nop\n" \
- " .section .fixup, \"ax\"\n" \
- " .align 4\n" \
- "3: mov %0, %3\n" \
- " j 2b\n" \
- " .previous\n" \
- " .section __ex_table, \"a\"\n" \
- " .align 4\n" \
- " .word 1b,3b\n" \
- " .previous\n" \
- \
- : "+r" (ret) \
- : "r" (src), "r" (dst), "ir" (-EFAULT))
- #define __arc_put_user_one_64(src, dst, ret) \
- __asm__ __volatile__( \
- "1: st %1,[%2]\n" \
- "4: st %R1,[%2, 4]\n" \
- "2: ;nop\n" \
- " .section .fixup, \"ax\"\n" \
- " .align 4\n" \
- "3: mov %0, %3\n" \
- " j 2b\n" \
- " .previous\n" \
- " .section __ex_table, \"a\"\n" \
- " .align 4\n" \
- " .word 1b,3b\n" \
- " .word 4b,3b\n" \
- " .previous\n" \
- \
- : "+r" (ret) \
- : "r" (src), "r" (dst), "ir" (-EFAULT))
- static inline unsigned long
- __arc_copy_from_user(void *to, const void __user *from, unsigned long n)
- {
- long res = 0;
- char val;
- unsigned long tmp1, tmp2, tmp3, tmp4;
- unsigned long orig_n = n;
- if (n == 0)
- return 0;
- /* unaligned */
- if (((unsigned long)to & 0x3) || ((unsigned long)from & 0x3)) {
- unsigned char tmp;
- __asm__ __volatile__ (
- " mov.f lp_count, %0 \n"
- " lpnz 2f \n"
- "1: ldb.ab %1, [%3, 1] \n"
- " stb.ab %1, [%2, 1] \n"
- " sub %0,%0,1 \n"
- "2: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "3: j 2b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 3b \n"
- " .previous \n"
- : "+r" (n),
- /*
- * Note as an '&' earlyclobber operand to make sure the
- * temporary register inside the loop is not the same as
- * FROM or TO.
- */
- "=&r" (tmp), "+r" (to), "+r" (from)
- :
- : "lp_count", "memory");
- return n;
- }
- /*
- * Hand-crafted constant propagation to reduce code sz of the
- * laddered copy 16x,8,4,2,1
- */
- if (__builtin_constant_p(orig_n)) {
- res = orig_n;
- if (orig_n / 16) {
- orig_n = orig_n % 16;
- __asm__ __volatile__(
- " lsr lp_count, %7,4 \n"
- " lp 3f \n"
- "1: ld.ab %3, [%2, 4] \n"
- "11: ld.ab %4, [%2, 4] \n"
- "12: ld.ab %5, [%2, 4] \n"
- "13: ld.ab %6, [%2, 4] \n"
- " st.ab %3, [%1, 4] \n"
- " st.ab %4, [%1, 4] \n"
- " st.ab %5, [%1, 4] \n"
- " st.ab %6, [%1, 4] \n"
- " sub %0,%0,16 \n"
- "3: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 3b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .word 11b,4b \n"
- " .word 12b,4b \n"
- " .word 13b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from),
- "=r"(tmp1), "=r"(tmp2), "=r"(tmp3), "=r"(tmp4)
- : "ir"(n)
- : "lp_count", "memory");
- }
- if (orig_n / 8) {
- orig_n = orig_n % 8;
- __asm__ __volatile__(
- "14: ld.ab %3, [%2,4] \n"
- "15: ld.ab %4, [%2,4] \n"
- " st.ab %3, [%1,4] \n"
- " st.ab %4, [%1,4] \n"
- " sub %0,%0,8 \n"
- "31: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 31b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 14b,4b \n"
- " .word 15b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from),
- "=r"(tmp1), "=r"(tmp2)
- :
- : "memory");
- }
- if (orig_n / 4) {
- orig_n = orig_n % 4;
- __asm__ __volatile__(
- "16: ld.ab %3, [%2,4] \n"
- " st.ab %3, [%1,4] \n"
- " sub %0,%0,4 \n"
- "32: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 32b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 16b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- if (orig_n / 2) {
- orig_n = orig_n % 2;
- __asm__ __volatile__(
- "17: ldw.ab %3, [%2,2] \n"
- " stw.ab %3, [%1,2] \n"
- " sub %0,%0,2 \n"
- "33: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 33b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 17b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- if (orig_n & 1) {
- __asm__ __volatile__(
- "18: ldb.ab %3, [%2,2] \n"
- " stb.ab %3, [%1,2] \n"
- " sub %0,%0,1 \n"
- "34: ; nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 34b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 18b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- } else { /* n is NOT constant, so laddered copy of 16x,8,4,2,1 */
- __asm__ __volatile__(
- " mov %0,%3 \n"
- " lsr.f lp_count, %3,4 \n" /* 16x bytes */
- " lpnz 3f \n"
- "1: ld.ab %5, [%2, 4] \n"
- "11: ld.ab %6, [%2, 4] \n"
- "12: ld.ab %7, [%2, 4] \n"
- "13: ld.ab %8, [%2, 4] \n"
- " st.ab %5, [%1, 4] \n"
- " st.ab %6, [%1, 4] \n"
- " st.ab %7, [%1, 4] \n"
- " st.ab %8, [%1, 4] \n"
- " sub %0,%0,16 \n"
- "3: and.f %3,%3,0xf \n" /* stragglers */
- " bz 34f \n"
- " bbit0 %3,3,31f \n" /* 8 bytes left */
- "14: ld.ab %5, [%2,4] \n"
- "15: ld.ab %6, [%2,4] \n"
- " st.ab %5, [%1,4] \n"
- " st.ab %6, [%1,4] \n"
- " sub.f %0,%0,8 \n"
- "31: bbit0 %3,2,32f \n" /* 4 bytes left */
- "16: ld.ab %5, [%2,4] \n"
- " st.ab %5, [%1,4] \n"
- " sub.f %0,%0,4 \n"
- "32: bbit0 %3,1,33f \n" /* 2 bytes left */
- "17: ldw.ab %5, [%2,2] \n"
- " stw.ab %5, [%1,2] \n"
- " sub.f %0,%0,2 \n"
- "33: bbit0 %3,0,34f \n"
- "18: ldb.ab %5, [%2,1] \n" /* 1 byte left */
- " stb.ab %5, [%1,1] \n"
- " sub.f %0,%0,1 \n"
- "34: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 34b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .word 11b,4b \n"
- " .word 12b,4b \n"
- " .word 13b,4b \n"
- " .word 14b,4b \n"
- " .word 15b,4b \n"
- " .word 16b,4b \n"
- " .word 17b,4b \n"
- " .word 18b,4b \n"
- " .previous \n"
- : "=r" (res), "+r"(to), "+r"(from), "+r"(n), "=r"(val),
- "=r"(tmp1), "=r"(tmp2), "=r"(tmp3), "=r"(tmp4)
- :
- : "lp_count", "memory");
- }
- return res;
- }
- extern unsigned long slowpath_copy_to_user(void __user *to, const void *from,
- unsigned long n);
- static inline unsigned long
- __arc_copy_to_user(void __user *to, const void *from, unsigned long n)
- {
- long res = 0;
- char val;
- unsigned long tmp1, tmp2, tmp3, tmp4;
- unsigned long orig_n = n;
- if (n == 0)
- return 0;
- /* unaligned */
- if (((unsigned long)to & 0x3) || ((unsigned long)from & 0x3)) {
- unsigned char tmp;
- __asm__ __volatile__(
- " mov.f lp_count, %0 \n"
- " lpnz 3f \n"
- " ldb.ab %1, [%3, 1] \n"
- "1: stb.ab %1, [%2, 1] \n"
- " sub %0, %0, 1 \n"
- "3: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 3b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .previous \n"
- : "+r" (n),
- /* Note as an '&' earlyclobber operand to make sure the
- * temporary register inside the loop is not the same as
- * FROM or TO.
- */
- "=&r" (tmp), "+r" (to), "+r" (from)
- :
- : "lp_count", "memory");
- return n;
- }
- if (__builtin_constant_p(orig_n)) {
- res = orig_n;
- if (orig_n / 16) {
- orig_n = orig_n % 16;
- __asm__ __volatile__(
- " lsr lp_count, %7,4 \n"
- " lp 3f \n"
- " ld.ab %3, [%2, 4] \n"
- " ld.ab %4, [%2, 4] \n"
- " ld.ab %5, [%2, 4] \n"
- " ld.ab %6, [%2, 4] \n"
- "1: st.ab %3, [%1, 4] \n"
- "11: st.ab %4, [%1, 4] \n"
- "12: st.ab %5, [%1, 4] \n"
- "13: st.ab %6, [%1, 4] \n"
- " sub %0, %0, 16 \n"
- "3:;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 3b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .word 11b,4b \n"
- " .word 12b,4b \n"
- " .word 13b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from),
- "=r"(tmp1), "=r"(tmp2), "=r"(tmp3), "=r"(tmp4)
- : "ir"(n)
- : "lp_count", "memory");
- }
- if (orig_n / 8) {
- orig_n = orig_n % 8;
- __asm__ __volatile__(
- " ld.ab %3, [%2,4] \n"
- " ld.ab %4, [%2,4] \n"
- "14: st.ab %3, [%1,4] \n"
- "15: st.ab %4, [%1,4] \n"
- " sub %0, %0, 8 \n"
- "31:;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 31b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 14b,4b \n"
- " .word 15b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from),
- "=r"(tmp1), "=r"(tmp2)
- :
- : "memory");
- }
- if (orig_n / 4) {
- orig_n = orig_n % 4;
- __asm__ __volatile__(
- " ld.ab %3, [%2,4] \n"
- "16: st.ab %3, [%1,4] \n"
- " sub %0, %0, 4 \n"
- "32:;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 32b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 16b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- if (orig_n / 2) {
- orig_n = orig_n % 2;
- __asm__ __volatile__(
- " ldw.ab %3, [%2,2] \n"
- "17: stw.ab %3, [%1,2] \n"
- " sub %0, %0, 2 \n"
- "33:;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 33b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 17b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- if (orig_n & 1) {
- __asm__ __volatile__(
- " ldb.ab %3, [%2,1] \n"
- "18: stb.ab %3, [%1,1] \n"
- " sub %0, %0, 1 \n"
- "34: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 34b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 18b,4b \n"
- " .previous \n"
- : "+r" (res), "+r"(to), "+r"(from), "=r"(tmp1)
- :
- : "memory");
- }
- } else { /* n is NOT constant, so laddered copy of 16x,8,4,2,1 */
- __asm__ __volatile__(
- " mov %0,%3 \n"
- " lsr.f lp_count, %3,4 \n" /* 16x bytes */
- " lpnz 3f \n"
- " ld.ab %5, [%2, 4] \n"
- " ld.ab %6, [%2, 4] \n"
- " ld.ab %7, [%2, 4] \n"
- " ld.ab %8, [%2, 4] \n"
- "1: st.ab %5, [%1, 4] \n"
- "11: st.ab %6, [%1, 4] \n"
- "12: st.ab %7, [%1, 4] \n"
- "13: st.ab %8, [%1, 4] \n"
- " sub %0, %0, 16 \n"
- "3: and.f %3,%3,0xf \n" /* stragglers */
- " bz 34f \n"
- " bbit0 %3,3,31f \n" /* 8 bytes left */
- " ld.ab %5, [%2,4] \n"
- " ld.ab %6, [%2,4] \n"
- "14: st.ab %5, [%1,4] \n"
- "15: st.ab %6, [%1,4] \n"
- " sub.f %0, %0, 8 \n"
- "31: bbit0 %3,2,32f \n" /* 4 bytes left */
- " ld.ab %5, [%2,4] \n"
- "16: st.ab %5, [%1,4] \n"
- " sub.f %0, %0, 4 \n"
- "32: bbit0 %3,1,33f \n" /* 2 bytes left */
- " ldw.ab %5, [%2,2] \n"
- "17: stw.ab %5, [%1,2] \n"
- " sub.f %0, %0, 2 \n"
- "33: bbit0 %3,0,34f \n"
- " ldb.ab %5, [%2,1] \n" /* 1 byte left */
- "18: stb.ab %5, [%1,1] \n"
- " sub.f %0, %0, 1 \n"
- "34: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: j 34b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .word 11b,4b \n"
- " .word 12b,4b \n"
- " .word 13b,4b \n"
- " .word 14b,4b \n"
- " .word 15b,4b \n"
- " .word 16b,4b \n"
- " .word 17b,4b \n"
- " .word 18b,4b \n"
- " .previous \n"
- : "=r" (res), "+r"(to), "+r"(from), "+r"(n), "=r"(val),
- "=r"(tmp1), "=r"(tmp2), "=r"(tmp3), "=r"(tmp4)
- :
- : "lp_count", "memory");
- }
- return res;
- }
- static inline unsigned long __arc_clear_user(void __user *to, unsigned long n)
- {
- long res = n;
- unsigned char *d_char = to;
- __asm__ __volatile__(
- " bbit0 %0, 0, 1f \n"
- "75: stb.ab %2, [%0,1] \n"
- " sub %1, %1, 1 \n"
- "1: bbit0 %0, 1, 2f \n"
- "76: stw.ab %2, [%0,2] \n"
- " sub %1, %1, 2 \n"
- "2: asr.f lp_count, %1, 2 \n"
- " lpnz 3f \n"
- "77: st.ab %2, [%0,4] \n"
- " sub %1, %1, 4 \n"
- "3: bbit0 %1, 1, 4f \n"
- "78: stw.ab %2, [%0,2] \n"
- " sub %1, %1, 2 \n"
- "4: bbit0 %1, 0, 5f \n"
- "79: stb.ab %2, [%0,1] \n"
- " sub %1, %1, 1 \n"
- "5: \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "3: j 5b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 75b, 3b \n"
- " .word 76b, 3b \n"
- " .word 77b, 3b \n"
- " .word 78b, 3b \n"
- " .word 79b, 3b \n"
- " .previous \n"
- : "+r"(d_char), "+r"(res)
- : "i"(0)
- : "lp_count", "memory");
- return res;
- }
- static inline long
- __arc_strncpy_from_user(char *dst, const char __user *src, long count)
- {
- long res = 0;
- char val;
- if (count == 0)
- return 0;
- __asm__ __volatile__(
- " mov lp_count, %5 \n"
- " lp 3f \n"
- "1: ldb.ab %3, [%2, 1] \n"
- " breq.d %3, 0, 3f \n"
- " stb.ab %3, [%1, 1] \n"
- " add %0, %0, 1 # Num of NON NULL bytes copied \n"
- "3: \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: mov %0, %4 # sets @res as -EFAULT \n"
- " j 3b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .previous \n"
- : "+r"(res), "+r"(dst), "+r"(src), "=r"(val)
- : "g"(-EFAULT), "r"(count)
- : "lp_count", "memory");
- return res;
- }
- static inline long __arc_strnlen_user(const char __user *s, long n)
- {
- long res, tmp1, cnt;
- char val;
- __asm__ __volatile__(
- " mov %2, %1 \n"
- "1: ldb.ab %3, [%0, 1] \n"
- " breq.d %3, 0, 2f \n"
- " sub.f %2, %2, 1 \n"
- " bnz 1b \n"
- " sub %2, %2, 1 \n"
- "2: sub %0, %1, %2 \n"
- "3: ;nop \n"
- " .section .fixup, \"ax\" \n"
- " .align 4 \n"
- "4: mov %0, 0 \n"
- " j 3b \n"
- " .previous \n"
- " .section __ex_table, \"a\" \n"
- " .align 4 \n"
- " .word 1b, 4b \n"
- " .previous \n"
- : "=r"(res), "=r"(tmp1), "=r"(cnt), "=r"(val)
- : "0"(s), "1"(n)
- : "memory");
- return res;
- }
- #ifndef CONFIG_CC_OPTIMIZE_FOR_SIZE
- #define __copy_from_user(t, f, n) __arc_copy_from_user(t, f, n)
- #define __copy_to_user(t, f, n) __arc_copy_to_user(t, f, n)
- #define __clear_user(d, n) __arc_clear_user(d, n)
- #define __strncpy_from_user(d, s, n) __arc_strncpy_from_user(d, s, n)
- #define __strnlen_user(s, n) __arc_strnlen_user(s, n)
- #else
- extern long arc_copy_from_user_noinline(void *to, const void __user * from,
- unsigned long n);
- extern long arc_copy_to_user_noinline(void __user *to, const void *from,
- unsigned long n);
- extern unsigned long arc_clear_user_noinline(void __user *to,
- unsigned long n);
- extern long arc_strncpy_from_user_noinline (char *dst, const char __user *src,
- long count);
- extern long arc_strnlen_user_noinline(const char __user *src, long n);
- #define __copy_from_user(t, f, n) arc_copy_from_user_noinline(t, f, n)
- #define __copy_to_user(t, f, n) arc_copy_to_user_noinline(t, f, n)
- #define __clear_user(d, n) arc_clear_user_noinline(d, n)
- #define __strncpy_from_user(d, s, n) arc_strncpy_from_user_noinline(d, s, n)
- #define __strnlen_user(s, n) arc_strnlen_user_noinline(s, n)
- #endif
- #include <asm-generic/uaccess.h>
- extern int fixup_exception(struct pt_regs *regs);
- #endif
|