Lines Matching defs:esp
86 mov eax, [esp + 12 + 4] // src
87 mov edi, [esp + 12 + 8] // src_stride
88 mov edx, [esp + 12 + 12] // dst
89 mov esi, [esp + 12 + 16] // dst_stride
90 mov ecx, [esp + 12 + 20] // width
181 mov eax, [esp + 16 + 4] // src
182 mov edi, [esp + 16 + 8] // src_stride
183 mov edx, [esp + 16 + 12] // dst_a
184 mov esi, [esp + 16 + 16] // dst_stride_a
185 mov ebx, [esp + 16 + 20] // dst_b
186 mov ebp, [esp + 16 + 24] // dst_stride_b
187 mov ecx, esp
188 sub esp, 4 + 16
189 and esp, ~15
190 mov [esp + 16], ecx
221 movdqa [esp], xmm5 // backup xmm5
242 movdqa xmm5, [esp] // restore xmm5
243 movdqa [esp], xmm6 // backup xmm6
254 movdqa xmm6, [esp] // restore xmm6
291 mov esp, [esp + 16]
401 "mov 0x14(%esp),%eax \n"
402 "mov 0x18(%esp),%edi \n"
403 "mov 0x1c(%esp),%edx \n"
404 "mov 0x20(%esp),%esi \n"
405 "mov 0x24(%esp),%ebx \n"
406 "mov 0x28(%esp),%ebp \n"
407 "mov %esp,%ecx \n"
408 "sub $0x14,%esp \n"
409 "and $0xfffffff0,%esp \n"
410 "mov %ecx,0x10(%esp) \n"
438 "movdqa %xmm5,(%esp) \n"
458 "movdqa (%esp),%xmm5 \n"
459 "movdqa %xmm6,(%esp) \n"
468 "movdqa (%esp),%xmm6 \n"
504 "mov 0x10(%esp),%esp \n"