/* * MD5 hash in x86-64 assembly * * Copyright (c) 2021 Project Nayuki. (MIT License) * https://www.nayuki.io/page/fast-md5-hash-implementation-in-x86-assembly * * Permission is hereby granted, free of charge, to any person obtaining a copy of * this software and associated documentation files (the "Software"), to deal in * the Software without restriction, including without limitation the rights to * use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of * the Software, and to permit persons to whom the Software is furnished to do so, * subject to the following conditions: * - The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * - The Software is provided "as is", without warranty of any kind, express or * implied, including but not limited to the warranties of merchantability, * fitness for a particular purpose and noninfringement. In no event shall the * authors or copyright holders be liable for any claim, damages or other * liability, whether in an action of contract, tort or otherwise, arising from, * out of or in connection with the Software or the use or other dealings in the * Software. */ /* void md5_compress(const uint8_t block[static 64], uint32_t state[static 4]) */ .globl md5_compress md5_compress: /* * Storage usage: * Bytes Location Description * 4 eax MD5 state variable A * 4 ebx MD5 state variable B * 4 ecx MD5 state variable C * 4 edx MD5 state variable D * 4 esi Temporary for calculation per round * 4 edi Temporary for calculation per round * 8 rbp Base address of block array argument (read-only) * 8 r8 Base address of state array argument (read-only) * 16 xmm0 Caller's value of rbx (only low 64 bits are used) * 16 xmm1 Caller's value of rbp (only low 64 bits are used) */ #define ROUND0(a, b, c, d, k, s, t) \ movl %c, %esi; \ addl (k*4)(%rbp), %a; \ xorl %d, %esi; \ andl %b, %esi; \ xorl %d, %esi; \ leal t(%esi,%a), %a; \ roll $s, %a; \ addl %b, %a; #define ROUND1(a, b, c, d, k, s, t) \ movl %d, %esi; \ movl %d, %edi; \ addl (k*4)(%rbp), %a; \ notl %esi; \ andl %b, %edi; \ andl %c, %esi; \ orl %edi, %esi; \ leal t(%esi,%a), %a; \ roll $s, %a; \ addl %b, %a; #define ROUND2(a, b, c, d, k, s, t) \ movl %c, %esi; \ addl (k*4)(%rbp), %a; \ xorl %d, %esi; \ xorl %b, %esi; \ leal t(%esi,%a), %a; \ roll $s, %a; \ addl %b, %a; #define ROUND3(a, b, c, d, k, s, t) \ movl %d, %esi; \ not %esi; \ addl (k*4)(%rbp), %a; \ orl %b, %esi; \ xorl %c, %esi; \ leal t(%esi,%a), %a; \ roll $s, %a; \ addl %b, %a; /* Save registers */ movq %rbx, %xmm0 movq %rbp, %xmm1 /* Load arguments */ movq %rdi, %rbp movl 0(%rsi), %eax /* a */ movl 4(%rsi), %ebx /* b */ movl 8(%rsi), %ecx /* c */ movl 12(%rsi), %edx /* d */ movq %rsi, %r8 /* 64 rounds of hashing */ ROUND0(eax, ebx, ecx, edx, 0, 7, -0x28955B88) ROUND0(edx, eax, ebx, ecx, 1, 12, -0x173848AA) ROUND0(ecx, edx, eax, ebx, 2, 17, 0x242070DB) ROUND0(ebx, ecx, edx, eax, 3, 22, -0x3E423112) ROUND0(eax, ebx, ecx, edx, 4, 7, -0x0A83F051) ROUND0(edx, eax, ebx, ecx, 5, 12, 0x4787C62A) ROUND0(ecx, edx, eax, ebx, 6, 17, -0x57CFB9ED) ROUND0(ebx, ecx, edx, eax, 7, 22, -0x02B96AFF) ROUND0(eax, ebx, ecx, edx, 8, 7, 0x698098D8) ROUND0(edx, eax, ebx, ecx, 9, 12, -0x74BB0851) ROUND0(ecx, edx, eax, ebx, 10, 17, -0x0000A44F) ROUND0(ebx, ecx, edx, eax, 11, 22, -0x76A32842) ROUND0(eax, ebx, ecx, edx, 12, 7, 0x6B901122) ROUND0(edx, eax, ebx, ecx, 13, 12, -0x02678E6D) ROUND0(ecx, edx, eax, ebx, 14, 17, -0x5986BC72) ROUND0(ebx, ecx, edx, eax, 15, 22, 0x49B40821) ROUND1(eax, ebx, ecx, edx, 1, 5, -0x09E1DA9E) ROUND1(edx, eax, ebx, ecx, 6, 9, -0x3FBF4CC0) ROUND1(ecx, edx, eax, ebx, 11, 14, 0x265E5A51) ROUND1(ebx, ecx, edx, eax, 0, 20, -0x16493856) ROUND1(eax, ebx, ecx, edx, 5, 5, -0x29D0EFA3) ROUND1(edx, eax, ebx, ecx, 10, 9, 0x02441453) ROUND1(ecx, edx, eax, ebx, 15, 14, -0x275E197F) ROUND1(ebx, ecx, edx, eax, 4, 20, -0x182C0438) ROUND1(eax, ebx, ecx, edx, 9, 5, 0x21E1CDE6) ROUND1(edx, eax, ebx, ecx, 14, 9, -0x3CC8F82A) ROUND1(ecx, edx, eax, ebx, 3, 14, -0x0B2AF279) ROUND1(ebx, ecx, edx, eax, 8, 20, 0x455A14ED) ROUND1(eax, ebx, ecx, edx, 13, 5, -0x561C16FB) ROUND1(edx, eax, ebx, ecx, 2, 9, -0x03105C08) ROUND1(ecx, edx, eax, ebx, 7, 14, 0x676F02D9) ROUND1(ebx, ecx, edx, eax, 12, 20, -0x72D5B376) ROUND2(eax, ebx, ecx, edx, 5, 4, -0x0005C6BE) ROUND2(edx, eax, ebx, ecx, 8, 11, -0x788E097F) ROUND2(ecx, edx, eax, ebx, 11, 16, 0x6D9D6122) ROUND2(ebx, ecx, edx, eax, 14, 23, -0x021AC7F4) ROUND2(eax, ebx, ecx, edx, 1, 4, -0x5B4115BC) ROUND2(edx, eax, ebx, ecx, 4, 11, 0x4BDECFA9) ROUND2(ecx, edx, eax, ebx, 7, 16, -0x0944B4A0) ROUND2(ebx, ecx, edx, eax, 10, 23, -0x41404390) ROUND2(eax, ebx, ecx, edx, 13, 4, 0x289B7EC6) ROUND2(edx, eax, ebx, ecx, 0, 11, -0x155ED806) ROUND2(ecx, edx, eax, ebx, 3, 16, -0x2B10CF7B) ROUND2(ebx, ecx, edx, eax, 6, 23, 0x04881D05) ROUND2(eax, ebx, ecx, edx, 9, 4, -0x262B2FC7) ROUND2(edx, eax, ebx, ecx, 12, 11, -0x1924661B) ROUND2(ecx, edx, eax, ebx, 15, 16, 0x1FA27CF8) ROUND2(ebx, ecx, edx, eax, 2, 23, -0x3B53A99B) ROUND3(eax, ebx, ecx, edx, 0, 6, -0x0BD6DDBC) ROUND3(edx, eax, ebx, ecx, 7, 10, 0x432AFF97) ROUND3(ecx, edx, eax, ebx, 14, 15, -0x546BDC59) ROUND3(ebx, ecx, edx, eax, 5, 21, -0x036C5FC7) ROUND3(eax, ebx, ecx, edx, 12, 6, 0x655B59C3) ROUND3(edx, eax, ebx, ecx, 3, 10, -0x70F3336E) ROUND3(ecx, edx, eax, ebx, 10, 15, -0x00100B83) ROUND3(ebx, ecx, edx, eax, 1, 21, -0x7A7BA22F) ROUND3(eax, ebx, ecx, edx, 8, 6, 0x6FA87E4F) ROUND3(edx, eax, ebx, ecx, 15, 10, -0x01D31920) ROUND3(ecx, edx, eax, ebx, 6, 15, -0x5CFEBCEC) ROUND3(ebx, ecx, edx, eax, 13, 21, 0x4E0811A1) ROUND3(eax, ebx, ecx, edx, 4, 6, -0x08AC817E) ROUND3(edx, eax, ebx, ecx, 11, 10, -0x42C50DCB) ROUND3(ecx, edx, eax, ebx, 2, 15, 0x2AD7D2BB) ROUND3(ebx, ecx, edx, eax, 9, 21, -0x14792C6F) /* Save updated state */ addl %eax, 0(%r8) addl %ebx, 4(%r8) addl %ecx, 8(%r8) addl %edx, 12(%r8) /* Restore registers */ movq %xmm0, %rbx movq %xmm1, %rbp retq