Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | /* |
| 2 | * arch/x86_64/lib/csum-partial.c |
| 3 | * |
| 4 | * This file contains network checksum routines that are better done |
| 5 | * in an architecture-specific manner due to speed. |
| 6 | */ |
| 7 | |
| 8 | #include <linux/compiler.h> |
| 9 | #include <linux/module.h> |
| 10 | #include <asm/checksum.h> |
| 11 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 12 | static inline unsigned short from32to16(unsigned a) |
| 13 | { |
| 14 | unsigned short b = a >> 16; |
| 15 | asm("addw %w2,%w0\n\t" |
| 16 | "adcw $0,%w0\n" |
| 17 | : "=r" (b) |
| 18 | : "0" (b), "r" (a)); |
| 19 | return b; |
| 20 | } |
| 21 | |
| 22 | /* |
| 23 | * Do a 64-bit checksum on an arbitrary memory area. |
| 24 | * Returns a 32bit checksum. |
| 25 | * |
| 26 | * This isn't as time critical as it used to be because many NICs |
| 27 | * do hardware checksumming these days. |
| 28 | * |
| 29 | * Things tried and found to not make it faster: |
| 30 | * Manual Prefetching |
| 31 | * Unrolling to an 128 bytes inner loop. |
| 32 | * Using interleaving with more registers to break the carry chains. |
| 33 | */ |
Andi Kleen | b6bcc4b | 2006-12-07 02:14:07 +0100 | [diff] [blame] | 34 | static unsigned do_csum(const unsigned char *buff, unsigned len) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 35 | { |
| 36 | unsigned odd, count; |
| 37 | unsigned long result = 0; |
| 38 | |
| 39 | if (unlikely(len == 0)) |
| 40 | return result; |
| 41 | odd = 1 & (unsigned long) buff; |
| 42 | if (unlikely(odd)) { |
| 43 | result = *buff << 8; |
| 44 | len--; |
| 45 | buff++; |
| 46 | } |
| 47 | count = len >> 1; /* nr of 16-bit words.. */ |
| 48 | if (count) { |
| 49 | if (2 & (unsigned long) buff) { |
| 50 | result += *(unsigned short *)buff; |
| 51 | count--; |
| 52 | len -= 2; |
| 53 | buff += 2; |
| 54 | } |
| 55 | count >>= 1; /* nr of 32-bit words.. */ |
| 56 | if (count) { |
| 57 | unsigned long zero; |
| 58 | unsigned count64; |
| 59 | if (4 & (unsigned long) buff) { |
| 60 | result += *(unsigned int *) buff; |
| 61 | count--; |
| 62 | len -= 4; |
| 63 | buff += 4; |
| 64 | } |
| 65 | count >>= 1; /* nr of 64-bit words.. */ |
| 66 | |
| 67 | /* main loop using 64byte blocks */ |
| 68 | zero = 0; |
| 69 | count64 = count >> 3; |
| 70 | while (count64) { |
| 71 | asm("addq 0*8(%[src]),%[res]\n\t" |
| 72 | "adcq 1*8(%[src]),%[res]\n\t" |
| 73 | "adcq 2*8(%[src]),%[res]\n\t" |
| 74 | "adcq 3*8(%[src]),%[res]\n\t" |
| 75 | "adcq 4*8(%[src]),%[res]\n\t" |
| 76 | "adcq 5*8(%[src]),%[res]\n\t" |
| 77 | "adcq 6*8(%[src]),%[res]\n\t" |
| 78 | "adcq 7*8(%[src]),%[res]\n\t" |
| 79 | "adcq %[zero],%[res]" |
| 80 | : [res] "=r" (result) |
| 81 | : [src] "r" (buff), [zero] "r" (zero), |
| 82 | "[res]" (result)); |
| 83 | buff += 64; |
| 84 | count64--; |
| 85 | } |
| 86 | |
| 87 | /* last upto 7 8byte blocks */ |
| 88 | count %= 8; |
| 89 | while (count) { |
| 90 | asm("addq %1,%0\n\t" |
| 91 | "adcq %2,%0\n" |
| 92 | : "=r" (result) |
| 93 | : "m" (*(unsigned long *)buff), |
| 94 | "r" (zero), "0" (result)); |
| 95 | --count; |
| 96 | buff += 8; |
| 97 | } |
| 98 | result = add32_with_carry(result>>32, |
| 99 | result&0xffffffff); |
| 100 | |
| 101 | if (len & 4) { |
| 102 | result += *(unsigned int *) buff; |
| 103 | buff += 4; |
| 104 | } |
| 105 | } |
| 106 | if (len & 2) { |
| 107 | result += *(unsigned short *) buff; |
| 108 | buff += 2; |
| 109 | } |
| 110 | } |
| 111 | if (len & 1) |
| 112 | result += *buff; |
| 113 | result = add32_with_carry(result>>32, result & 0xffffffff); |
| 114 | if (unlikely(odd)) { |
| 115 | result = from32to16(result); |
| 116 | result = ((result >> 8) & 0xff) | ((result & 0xff) << 8); |
| 117 | } |
| 118 | return result; |
| 119 | } |
| 120 | |
| 121 | /* |
| 122 | * computes the checksum of a memory block at buff, length len, |
| 123 | * and adds in "sum" (32-bit) |
| 124 | * |
| 125 | * returns a 32-bit number suitable for feeding into itself |
| 126 | * or csum_tcpudp_magic |
| 127 | * |
| 128 | * this function must be called with even lengths, except |
| 129 | * for the last fragment, which may be odd |
| 130 | * |
| 131 | * it's best to have buff aligned on a 64-bit boundary |
| 132 | */ |
Al Viro | a4f89fb | 2006-11-14 21:20:08 -0800 | [diff] [blame] | 133 | __wsum csum_partial(const void *buff, int len, __wsum sum) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 134 | { |
Al Viro | a4f89fb | 2006-11-14 21:20:08 -0800 | [diff] [blame] | 135 | return (__force __wsum)add32_with_carry(do_csum(buff, len), |
| 136 | (__force u32)sum); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 137 | } |
| 138 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 139 | /* |
| 140 | * this routine is used for miscellaneous IP-like checksums, mainly |
| 141 | * in icmp.c |
| 142 | */ |
Al Viro | a4f89fb | 2006-11-14 21:20:08 -0800 | [diff] [blame] | 143 | __sum16 ip_compute_csum(const void *buff, int len) |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 144 | { |
| 145 | return csum_fold(csum_partial(buff,len,0)); |
| 146 | } |
Andi Kleen | 2ee60e17 | 2006-06-26 13:59:44 +0200 | [diff] [blame] | 147 | EXPORT_SYMBOL(ip_compute_csum); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 148 | |