Files
llvm-6502/test/CodeGen/ARM
Evan Cheng 78fe9ababe Optimizing (zext A + zext B) * C, to (VMULL A, C) + (VMULL B, C) during
isel lowering to fold the zero-extend's and take advantage of no-stall
back to back vmul + vmla:
 vmull q0, d4, d6
 vmlal q0, d5, d6
is faster than
 vaddl q0, d4, d5
 vmovl q1, d6                                                                                                                                                                             
 vmul  q0, q0, q1

This allows us to vmull + vmlal for:
    f = vmull_u8(   vget_high_u8(s), c);
    f = vmlal_u8(f, vget_low_u8(s),  c);

rdar://9197392


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@128444 91177308-0d34-0410-b5e6-96231b3b80d8
2011-03-29 01:56:09 +00:00
..
2010-11-03 18:21:33 +00:00
2010-08-04 21:56:30 +00:00
2010-09-17 21:46:16 +00:00
2010-10-22 21:55:03 +00:00
2011-03-11 21:52:04 +00:00
2010-01-19 00:35:20 +00:00
2011-03-14 21:05:21 +00:00
2010-09-02 20:19:24 +00:00
2010-03-14 18:42:52 +00:00
2011-03-23 23:34:19 +00:00
2010-04-17 16:29:15 +00:00
2010-04-17 16:29:15 +00:00
2010-02-08 23:47:34 +00:00
2010-06-04 01:01:24 +00:00
2010-06-04 23:28:13 +00:00
2009-10-21 08:15:52 +00:00
2010-10-22 22:04:28 +00:00
2010-01-05 17:55:26 +00:00
2011-03-15 01:49:08 +00:00
2011-01-01 20:57:26 +00:00
2010-08-30 23:41:49 +00:00
2009-09-24 20:23:02 +00:00
2010-11-29 22:39:38 +00:00
2011-03-15 20:47:26 +00:00
2009-09-15 20:58:02 +00:00
2009-09-15 20:58:02 +00:00
2009-10-07 20:51:42 +00:00