Lines Matching defs:esp
70 mov eax, [esp + 12 + 4] // src
71 mov edi, [esp + 12 + 8] // src_stride
72 mov edx, [esp + 12 + 12] // dst
73 mov esi, [esp + 12 + 16] // dst_stride
74 mov ecx, [esp + 12 + 20] // width
165 mov eax, [esp + 16 + 4] // src
166 mov edi, [esp + 16 + 8] // src_stride
167 mov edx, [esp + 16 + 12] // dst_a
168 mov esi, [esp + 16 + 16] // dst_stride_a
169 mov ebx, [esp + 16 + 20] // dst_b
170 mov ebp, [esp + 16 + 24] // dst_stride_b
171 mov ecx, esp
172 sub esp, 4 + 16
173 and esp, ~15
174 mov [esp + 16], ecx
205 movdqa [esp], xmm5 // backup xmm5
226 movdqa xmm5, [esp] // restore xmm5
227 movdqa [esp], xmm6 // backup xmm6
238 movdqa xmm6, [esp] // restore xmm6
275 mov esp, [esp + 16]
384 "mov 0x14(%esp),%eax \n"
385 "mov 0x18(%esp),%edi \n"
386 "mov 0x1c(%esp),%edx \n"
387 "mov 0x20(%esp),%esi \n"
388 "mov 0x24(%esp),%ebx \n"
389 "mov 0x28(%esp),%ebp \n"
390 "mov %esp,%ecx \n"
391 "sub $0x14,%esp \n"
392 "and $0xfffffff0,%esp \n"
393 "mov %ecx,0x10(%esp) \n"
421 "movdqa %xmm5,(%esp) \n"
441 "movdqa (%esp),%xmm5 \n"
442 "movdqa %xmm6,(%esp) \n"
451 "movdqa (%esp),%xmm6 \n"
487 "mov 0x10(%esp),%esp \n"