- " mov %rdi,%rax\n" /* stack */
- " add %rsi,%rax\n" /* size */
- " andq $-16, %rax\n" /* align stack */
- " movq $0, -8(%rax)\n" /* @return for func */
- " mov %rdx,-16(%rax)\n" /* func */
- " mov %rcx,-24(%rax)\n" /* arg/rdi */
- " movq $0, -32(%rax)\n" /* rsi */
- " movq $0, -40(%rax)\n" /* rdx */
- " movq $0, -48(%rax)\n" /* rcx */
- " movq $0, -56(%rax)\n" /* r8 */
- " movq $0, -64(%rax)\n" /* r9 */
- " movq $0, -72(%rax)\n" /* rbp */
- " movq $0, -80(%rax)\n" /* rbx */
- " movq $0, -88(%rax)\n" /* r12 */
- " movq $0, -96(%rax)\n" /* r13 */
- " movq $0, -104(%rax)\n" /* r14 */
- " movq $0, -112(%rax)\n" /* r15 */
- " sub $112,%rax\n"
- " ret\n"
-);
-
-__asm__ (
+ " mov %rdi,%rax\n" /* stack */
+ " add %rsi,%rax\n" /* size */
+ " andq $-16, %rax\n" /* align stack */
+ " movq $0, -8(%rax)\n" /* @return for func */
+ " mov %rdx,-16(%rax)\n" /* func */
+ " mov %rcx,-24(%rax)\n" /* arg/rdi */
+ " movq $0, -32(%rax)\n" /* rsi */
+ " movq $0, -40(%rax)\n" /* rdx */
+ " movq $0, -48(%rax)\n" /* rcx */
+ " movq $0, -56(%rax)\n" /* r8 */
+ " movq $0, -64(%rax)\n" /* r9 */
+ " movq $0, -72(%rax)\n" /* rbp */
+ " movq $0, -80(%rax)\n" /* rbx */
+ " movq $0, -88(%rax)\n" /* r12 */
+ " movq $0, -96(%rax)\n" /* r13 */
+ " movq $0, -104(%rax)\n" /* r14 */
+ " movq $0, -112(%rax)\n" /* r15 */
+ " sub $112,%rax\n"
+ " ret\n");
+
+__asm__(