forked from OSchip/llvm-project
				
			
		
			
				
	
	
		
			2597 lines
		
	
	
		
			78 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			2597 lines
		
	
	
		
			78 KiB
		
	
	
	
		
			C++
		
	
	
	
| //===- NeonEmitter.cpp - Generate arm_neon.h for use with clang -*- C++ -*-===//
 | |
| //
 | |
| //                     The LLVM Compiler Infrastructure
 | |
| //
 | |
| // This file is distributed under the University of Illinois Open Source
 | |
| // License. See LICENSE.TXT for details.
 | |
| //
 | |
| //===----------------------------------------------------------------------===//
 | |
| //
 | |
| // This tablegen backend is responsible for emitting arm_neon.h, which includes
 | |
| // a declaration and definition of each function specified by the ARM NEON
 | |
| // compiler interface.  See ARM document DUI0348B.
 | |
| //
 | |
| // Each NEON instruction is implemented in terms of 1 or more functions which
 | |
| // are suffixed with the element type of the input vectors.  Functions may be
 | |
| // implemented in terms of generic vector operations such as +, *, -, etc. or
 | |
| // by calling a __builtin_-prefixed function which will be handled by clang's
 | |
| // CodeGen library.
 | |
| //
 | |
| // Additional validation code can be generated by this file when runHeader() is
 | |
| // called, rather than the normal run() entry point.  A complete set of tests
 | |
| // for Neon intrinsics can be generated by calling the runTests() entry point.
 | |
| //
 | |
| //===----------------------------------------------------------------------===//
 | |
| 
 | |
| #include "llvm/ADT/DenseMap.h"
 | |
| #include "llvm/ADT/SmallString.h"
 | |
| #include "llvm/ADT/SmallVector.h"
 | |
| #include "llvm/ADT/StringExtras.h"
 | |
| #include "llvm/ADT/StringMap.h"
 | |
| #include "llvm/Support/ErrorHandling.h"
 | |
| #include "llvm/TableGen/Error.h"
 | |
| #include "llvm/TableGen/Record.h"
 | |
| #include "llvm/TableGen/TableGenBackend.h"
 | |
| #include <string>
 | |
| using namespace llvm;
 | |
| 
 | |
| enum OpKind {
 | |
|   OpNone,
 | |
|   OpUnavailable,
 | |
|   OpAdd,
 | |
|   OpAddl,
 | |
|   OpAddw,
 | |
|   OpSub,
 | |
|   OpSubl,
 | |
|   OpSubw,
 | |
|   OpMul,
 | |
|   OpMla,
 | |
|   OpMlal,
 | |
|   OpMls,
 | |
|   OpMlsl,
 | |
|   OpMulN,
 | |
|   OpMlaN,
 | |
|   OpMlsN,
 | |
|   OpMlalN,
 | |
|   OpMlslN,
 | |
|   OpMulLane,
 | |
|   OpMullLane,
 | |
|   OpMlaLane,
 | |
|   OpMlsLane,
 | |
|   OpMlalLane,
 | |
|   OpMlslLane,
 | |
|   OpQDMullLane,
 | |
|   OpQDMlalLane,
 | |
|   OpQDMlslLane,
 | |
|   OpQDMulhLane,
 | |
|   OpQRDMulhLane,
 | |
|   OpEq,
 | |
|   OpGe,
 | |
|   OpLe,
 | |
|   OpGt,
 | |
|   OpLt,
 | |
|   OpNeg,
 | |
|   OpNot,
 | |
|   OpAnd,
 | |
|   OpOr,
 | |
|   OpXor,
 | |
|   OpAndNot,
 | |
|   OpOrNot,
 | |
|   OpCast,
 | |
|   OpConcat,
 | |
|   OpDup,
 | |
|   OpDupLane,
 | |
|   OpHi,
 | |
|   OpLo,
 | |
|   OpSelect,
 | |
|   OpRev16,
 | |
|   OpRev32,
 | |
|   OpRev64,
 | |
|   OpReinterpret,
 | |
|   OpAbdl,
 | |
|   OpAba,
 | |
|   OpAbal,
 | |
|   OpDiv
 | |
| };
 | |
| 
 | |
| enum ClassKind {
 | |
|   ClassNone,
 | |
|   ClassI,           // generic integer instruction, e.g., "i8" suffix
 | |
|   ClassS,           // signed/unsigned/poly, e.g., "s8", "u8" or "p8" suffix
 | |
|   ClassW,           // width-specific instruction, e.g., "8" suffix
 | |
|   ClassB,           // bitcast arguments with enum argument to specify type
 | |
|   ClassL,           // Logical instructions which are op instructions
 | |
|                     // but we need to not emit any suffix for in our
 | |
|                     // tests.
 | |
|   ClassNoTest       // Instructions which we do not test since they are
 | |
|                     // not TRUE instructions.
 | |
| };
 | |
| 
 | |
| /// NeonTypeFlags - Flags to identify the types for overloaded Neon
 | |
| /// builtins.  These must be kept in sync with the flags in
 | |
| /// include/clang/Basic/TargetBuiltins.h.
 | |
| namespace {
 | |
| class NeonTypeFlags {
 | |
|   enum {
 | |
|     EltTypeMask = 0xf,
 | |
|     UnsignedFlag = 0x10,
 | |
|     QuadFlag = 0x20
 | |
|   };
 | |
|   uint32_t Flags;
 | |
| 
 | |
| public:
 | |
|   enum EltType {
 | |
|     Int8,
 | |
|     Int16,
 | |
|     Int32,
 | |
|     Int64,
 | |
|     Poly8,
 | |
|     Poly16,
 | |
|     Float16,
 | |
|     Float32,
 | |
|     Float64
 | |
|   };
 | |
| 
 | |
|   NeonTypeFlags(unsigned F) : Flags(F) {}
 | |
|   NeonTypeFlags(EltType ET, bool IsUnsigned, bool IsQuad) : Flags(ET) {
 | |
|     if (IsUnsigned)
 | |
|       Flags |= UnsignedFlag;
 | |
|     if (IsQuad)
 | |
|       Flags |= QuadFlag;
 | |
|   }
 | |
| 
 | |
|   uint32_t getFlags() const { return Flags; }
 | |
| };
 | |
| } // end anonymous namespace
 | |
| 
 | |
| namespace {
 | |
| class NeonEmitter {
 | |
|   RecordKeeper &Records;
 | |
|   StringMap<OpKind> OpMap;
 | |
|   DenseMap<Record*, ClassKind> ClassMap;
 | |
| 
 | |
| public:
 | |
|   NeonEmitter(RecordKeeper &R) : Records(R) {
 | |
|     OpMap["OP_NONE"]  = OpNone;
 | |
|     OpMap["OP_UNAVAILABLE"] = OpUnavailable;
 | |
|     OpMap["OP_ADD"]   = OpAdd;
 | |
|     OpMap["OP_ADDL"]  = OpAddl;
 | |
|     OpMap["OP_ADDW"]  = OpAddw;
 | |
|     OpMap["OP_SUB"]   = OpSub;
 | |
|     OpMap["OP_SUBL"]  = OpSubl;
 | |
|     OpMap["OP_SUBW"]  = OpSubw;
 | |
|     OpMap["OP_MUL"]   = OpMul;
 | |
|     OpMap["OP_MLA"]   = OpMla;
 | |
|     OpMap["OP_MLAL"]  = OpMlal;
 | |
|     OpMap["OP_MLS"]   = OpMls;
 | |
|     OpMap["OP_MLSL"]  = OpMlsl;
 | |
|     OpMap["OP_MUL_N"] = OpMulN;
 | |
|     OpMap["OP_MLA_N"] = OpMlaN;
 | |
|     OpMap["OP_MLS_N"] = OpMlsN;
 | |
|     OpMap["OP_MLAL_N"] = OpMlalN;
 | |
|     OpMap["OP_MLSL_N"] = OpMlslN;
 | |
|     OpMap["OP_MUL_LN"]= OpMulLane;
 | |
|     OpMap["OP_MULL_LN"] = OpMullLane;
 | |
|     OpMap["OP_MLA_LN"]= OpMlaLane;
 | |
|     OpMap["OP_MLS_LN"]= OpMlsLane;
 | |
|     OpMap["OP_MLAL_LN"] = OpMlalLane;
 | |
|     OpMap["OP_MLSL_LN"] = OpMlslLane;
 | |
|     OpMap["OP_QDMULL_LN"] = OpQDMullLane;
 | |
|     OpMap["OP_QDMLAL_LN"] = OpQDMlalLane;
 | |
|     OpMap["OP_QDMLSL_LN"] = OpQDMlslLane;
 | |
|     OpMap["OP_QDMULH_LN"] = OpQDMulhLane;
 | |
|     OpMap["OP_QRDMULH_LN"] = OpQRDMulhLane;
 | |
|     OpMap["OP_EQ"]    = OpEq;
 | |
|     OpMap["OP_GE"]    = OpGe;
 | |
|     OpMap["OP_LE"]    = OpLe;
 | |
|     OpMap["OP_GT"]    = OpGt;
 | |
|     OpMap["OP_LT"]    = OpLt;
 | |
|     OpMap["OP_NEG"]   = OpNeg;
 | |
|     OpMap["OP_NOT"]   = OpNot;
 | |
|     OpMap["OP_AND"]   = OpAnd;
 | |
|     OpMap["OP_OR"]    = OpOr;
 | |
|     OpMap["OP_XOR"]   = OpXor;
 | |
|     OpMap["OP_ANDN"]  = OpAndNot;
 | |
|     OpMap["OP_ORN"]   = OpOrNot;
 | |
|     OpMap["OP_CAST"]  = OpCast;
 | |
|     OpMap["OP_CONC"]  = OpConcat;
 | |
|     OpMap["OP_HI"]    = OpHi;
 | |
|     OpMap["OP_LO"]    = OpLo;
 | |
|     OpMap["OP_DUP"]   = OpDup;
 | |
|     OpMap["OP_DUP_LN"] = OpDupLane;
 | |
|     OpMap["OP_SEL"]   = OpSelect;
 | |
|     OpMap["OP_REV16"] = OpRev16;
 | |
|     OpMap["OP_REV32"] = OpRev32;
 | |
|     OpMap["OP_REV64"] = OpRev64;
 | |
|     OpMap["OP_REINT"] = OpReinterpret;
 | |
|     OpMap["OP_ABDL"]  = OpAbdl;
 | |
|     OpMap["OP_ABA"]   = OpAba;
 | |
|     OpMap["OP_ABAL"]  = OpAbal;
 | |
|     OpMap["OP_DIV"] = OpDiv;
 | |
| 
 | |
|     Record *SI = R.getClass("SInst");
 | |
|     Record *II = R.getClass("IInst");
 | |
|     Record *WI = R.getClass("WInst");
 | |
|     Record *SOpI = R.getClass("SOpInst");
 | |
|     Record *IOpI = R.getClass("IOpInst");
 | |
|     Record *WOpI = R.getClass("WOpInst");
 | |
|     Record *LOpI = R.getClass("LOpInst");
 | |
|     Record *NoTestOpI = R.getClass("NoTestOpInst");
 | |
| 
 | |
|     ClassMap[SI] = ClassS;
 | |
|     ClassMap[II] = ClassI;
 | |
|     ClassMap[WI] = ClassW;
 | |
|     ClassMap[SOpI] = ClassS;
 | |
|     ClassMap[IOpI] = ClassI;
 | |
|     ClassMap[WOpI] = ClassW;
 | |
|     ClassMap[LOpI] = ClassL;
 | |
|     ClassMap[NoTestOpI] = ClassNoTest;
 | |
|   }
 | |
| 
 | |
|   // run - Emit arm_neon.h.inc
 | |
|   void run(raw_ostream &o);
 | |
| 
 | |
|   // runHeader - Emit all the __builtin prototypes used in arm_neon.h
 | |
|   void runHeader(raw_ostream &o);
 | |
| 
 | |
|   // runTests - Emit tests for all the Neon intrinsics.
 | |
|   void runTests(raw_ostream &o);
 | |
| 
 | |
| private:
 | |
|   void emitIntrinsic(raw_ostream &OS, Record *R,
 | |
|                      StringMap<ClassKind> &EmittedMap);
 | |
|   void genBuiltinsDef(raw_ostream &OS, StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                       bool isA64GenBuiltinDef);
 | |
|   void genOverloadTypeCheckCode(raw_ostream &OS,
 | |
|                                 StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                                 bool isA64TypeCheck);
 | |
|   void genIntrinsicRangeCheckCode(raw_ostream &OS,
 | |
|                                   StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                                   bool isA64RangeCheck);
 | |
|   void genTargetTest(raw_ostream &OS, StringMap<OpKind> &EmittedMap,
 | |
|                      bool isA64TestGen);
 | |
| };
 | |
| } // end anonymous namespace
 | |
| 
 | |
| /// ParseTypes - break down a string such as "fQf" into a vector of StringRefs,
 | |
| /// which each StringRef representing a single type declared in the string.
 | |
| /// for "fQf" we would end up with 2 StringRefs, "f", and "Qf", representing
 | |
| /// 2xfloat and 4xfloat respectively.
 | |
| static void ParseTypes(Record *r, std::string &s,
 | |
|                        SmallVectorImpl<StringRef> &TV) {
 | |
|   const char *data = s.data();
 | |
|   int len = 0;
 | |
| 
 | |
|   for (unsigned i = 0, e = s.size(); i != e; ++i, ++len) {
 | |
|     if (data[len] == 'P' || data[len] == 'Q' || data[len] == 'U')
 | |
|       continue;
 | |
| 
 | |
|     switch (data[len]) {
 | |
|       case 'c':
 | |
|       case 's':
 | |
|       case 'i':
 | |
|       case 'l':
 | |
|       case 'h':
 | |
|       case 'f':
 | |
|       case 'd':
 | |
|         break;
 | |
|       default:
 | |
|         PrintFatalError(r->getLoc(),
 | |
|                       "Unexpected letter: " + std::string(data + len, 1));
 | |
|     }
 | |
|     TV.push_back(StringRef(data, len + 1));
 | |
|     data += len + 1;
 | |
|     len = -1;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// Widen - Convert a type code into the next wider type.  char -> short,
 | |
| /// short -> int, etc.
 | |
| static char Widen(const char t) {
 | |
|   switch (t) {
 | |
|     case 'c':
 | |
|       return 's';
 | |
|     case 's':
 | |
|       return 'i';
 | |
|     case 'i':
 | |
|       return 'l';
 | |
|     case 'h':
 | |
|       return 'f';
 | |
|     default:
 | |
|       PrintFatalError("unhandled type in widen!");
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// Narrow - Convert a type code into the next smaller type.  short -> char,
 | |
| /// float -> half float, etc.
 | |
| static char Narrow(const char t) {
 | |
|   switch (t) {
 | |
|     case 's':
 | |
|       return 'c';
 | |
|     case 'i':
 | |
|       return 's';
 | |
|     case 'l':
 | |
|       return 'i';
 | |
|     case 'f':
 | |
|       return 'h';
 | |
|     default:
 | |
|       PrintFatalError("unhandled type in narrow!");
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// For a particular StringRef, return the base type code, and whether it has
 | |
| /// the quad-vector, polynomial, or unsigned modifiers set.
 | |
| static char ClassifyType(StringRef ty, bool &quad, bool &poly, bool &usgn) {
 | |
|   unsigned off = 0;
 | |
| 
 | |
|   // remember quad.
 | |
|   if (ty[off] == 'Q') {
 | |
|     quad = true;
 | |
|     ++off;
 | |
|   }
 | |
| 
 | |
|   // remember poly.
 | |
|   if (ty[off] == 'P') {
 | |
|     poly = true;
 | |
|     ++off;
 | |
|   }
 | |
| 
 | |
|   // remember unsigned.
 | |
|   if (ty[off] == 'U') {
 | |
|     usgn = true;
 | |
|     ++off;
 | |
|   }
 | |
| 
 | |
|   // base type to get the type string for.
 | |
|   return ty[off];
 | |
| }
 | |
| 
 | |
| /// ModType - Transform a type code and its modifiers based on a mod code. The
 | |
| /// mod code definitions may be found at the top of arm_neon.td.
 | |
| static char ModType(const char mod, char type, bool &quad, bool &poly,
 | |
|                     bool &usgn, bool &scal, bool &cnst, bool &pntr) {
 | |
|   switch (mod) {
 | |
|     case 't':
 | |
|       if (poly) {
 | |
|         poly = false;
 | |
|         usgn = true;
 | |
|       }
 | |
|       break;
 | |
|     case 'u':
 | |
|       usgn = true;
 | |
|       poly = false;
 | |
|       if (type == 'f')
 | |
|         type = 'i';
 | |
|       if (type == 'd')
 | |
|         type = 'l';
 | |
|       break;
 | |
|     case 'x':
 | |
|       usgn = false;
 | |
|       poly = false;
 | |
|       if (type == 'f')
 | |
|         type = 'i';
 | |
|       break;
 | |
|     case 'f':
 | |
|       if (type == 'h')
 | |
|         quad = true;
 | |
|       type = 'f';
 | |
|       usgn = false;
 | |
|       break;
 | |
|     case 'g':
 | |
|       quad = false;
 | |
|       break;
 | |
|     case 'w':
 | |
|       type = Widen(type);
 | |
|       quad = true;
 | |
|       break;
 | |
|     case 'n':
 | |
|       type = Widen(type);
 | |
|       break;
 | |
|     case 'i':
 | |
|       type = 'i';
 | |
|       scal = true;
 | |
|       break;
 | |
|     case 'l':
 | |
|       type = 'l';
 | |
|       scal = true;
 | |
|       usgn = true;
 | |
|       break;
 | |
|     case 's':
 | |
|     case 'a':
 | |
|       scal = true;
 | |
|       break;
 | |
|     case 'k':
 | |
|       quad = true;
 | |
|       break;
 | |
|     case 'c':
 | |
|       cnst = true;
 | |
|     case 'p':
 | |
|       pntr = true;
 | |
|       scal = true;
 | |
|       break;
 | |
|     case 'h':
 | |
|       type = Narrow(type);
 | |
|       if (type == 'h')
 | |
|         quad = false;
 | |
|       break;
 | |
|     case 'e':
 | |
|       type = Narrow(type);
 | |
|       usgn = true;
 | |
|       break;
 | |
|     default:
 | |
|       break;
 | |
|   }
 | |
|   return type;
 | |
| }
 | |
| 
 | |
| /// TypeString - for a modifier and type, generate the name of the typedef for
 | |
| /// that type.  QUc -> uint8x8_t.
 | |
| static std::string TypeString(const char mod, StringRef typestr) {
 | |
|   bool quad = false;
 | |
|   bool poly = false;
 | |
|   bool usgn = false;
 | |
|   bool scal = false;
 | |
|   bool cnst = false;
 | |
|   bool pntr = false;
 | |
| 
 | |
|   if (mod == 'v')
 | |
|     return "void";
 | |
|   if (mod == 'i')
 | |
|     return "int";
 | |
| 
 | |
|   // base type to get the type string for.
 | |
|   char type = ClassifyType(typestr, quad, poly, usgn);
 | |
| 
 | |
|   // Based on the modifying character, change the type and width if necessary.
 | |
|   type = ModType(mod, type, quad, poly, usgn, scal, cnst, pntr);
 | |
| 
 | |
|   SmallString<128> s;
 | |
| 
 | |
|   if (usgn)
 | |
|     s.push_back('u');
 | |
| 
 | |
|   switch (type) {
 | |
|     case 'c':
 | |
|       s += poly ? "poly8" : "int8";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x16" : "x8";
 | |
|       break;
 | |
|     case 's':
 | |
|       s += poly ? "poly16" : "int16";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x8" : "x4";
 | |
|       break;
 | |
|     case 'i':
 | |
|       s += "int32";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x4" : "x2";
 | |
|       break;
 | |
|     case 'l':
 | |
|       s += "int64";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x2" : "x1";
 | |
|       break;
 | |
|     case 'h':
 | |
|       s += "float16";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x8" : "x4";
 | |
|       break;
 | |
|     case 'f':
 | |
|       s += "float32";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x4" : "x2";
 | |
|       break;
 | |
|     case 'd':
 | |
|       s += "float64";
 | |
|       if (scal)
 | |
|         break;
 | |
|       s += quad ? "x2" : "x1";
 | |
|       break;
 | |
| 
 | |
|     default:
 | |
|       PrintFatalError("unhandled type!");
 | |
|   }
 | |
| 
 | |
|   if (mod == '2')
 | |
|     s += "x2";
 | |
|   if (mod == '3')
 | |
|     s += "x3";
 | |
|   if (mod == '4')
 | |
|     s += "x4";
 | |
| 
 | |
|   // Append _t, finishing the type string typedef type.
 | |
|   s += "_t";
 | |
| 
 | |
|   if (cnst)
 | |
|     s += " const";
 | |
| 
 | |
|   if (pntr)
 | |
|     s += " *";
 | |
| 
 | |
|   return s.str();
 | |
| }
 | |
| 
 | |
| /// BuiltinTypeString - for a modifier and type, generate the clang
 | |
| /// BuiltinsARM.def prototype code for the function.  See the top of clang's
 | |
| /// Builtins.def for a description of the type strings.
 | |
| static std::string BuiltinTypeString(const char mod, StringRef typestr,
 | |
|                                      ClassKind ck, bool ret) {
 | |
|   bool quad = false;
 | |
|   bool poly = false;
 | |
|   bool usgn = false;
 | |
|   bool scal = false;
 | |
|   bool cnst = false;
 | |
|   bool pntr = false;
 | |
| 
 | |
|   if (mod == 'v')
 | |
|     return "v"; // void
 | |
|   if (mod == 'i')
 | |
|     return "i"; // int
 | |
| 
 | |
|   // base type to get the type string for.
 | |
|   char type = ClassifyType(typestr, quad, poly, usgn);
 | |
| 
 | |
|   // Based on the modifying character, change the type and width if necessary.
 | |
|   type = ModType(mod, type, quad, poly, usgn, scal, cnst, pntr);
 | |
| 
 | |
|   // All pointers are void* pointers.  Change type to 'v' now.
 | |
|   if (pntr) {
 | |
|     usgn = false;
 | |
|     poly = false;
 | |
|     type = 'v';
 | |
|   }
 | |
|   // Treat half-float ('h') types as unsigned short ('s') types.
 | |
|   if (type == 'h') {
 | |
|     type = 's';
 | |
|     usgn = true;
 | |
|   }
 | |
|   usgn = usgn | poly | ((ck == ClassI || ck == ClassW) && scal && type != 'f');
 | |
| 
 | |
|   if (scal) {
 | |
|     SmallString<128> s;
 | |
| 
 | |
|     if (usgn)
 | |
|       s.push_back('U');
 | |
|     else if (type == 'c')
 | |
|       s.push_back('S'); // make chars explicitly signed
 | |
| 
 | |
|     if (type == 'l') // 64-bit long
 | |
|       s += "LLi";
 | |
|     else
 | |
|       s.push_back(type);
 | |
| 
 | |
|     if (cnst)
 | |
|       s.push_back('C');
 | |
|     if (pntr)
 | |
|       s.push_back('*');
 | |
|     return s.str();
 | |
|   }
 | |
| 
 | |
|   // Since the return value must be one type, return a vector type of the
 | |
|   // appropriate width which we will bitcast.  An exception is made for
 | |
|   // returning structs of 2, 3, or 4 vectors which are returned in a sret-like
 | |
|   // fashion, storing them to a pointer arg.
 | |
|   if (ret) {
 | |
|     if (mod >= '2' && mod <= '4')
 | |
|       return "vv*"; // void result with void* first argument
 | |
|     if (mod == 'f' || (ck != ClassB && type == 'f'))
 | |
|       return quad ? "V4f" : "V2f";
 | |
|     if (ck != ClassB && type == 's')
 | |
|       return quad ? "V8s" : "V4s";
 | |
|     if (ck != ClassB && type == 'i')
 | |
|       return quad ? "V4i" : "V2i";
 | |
|     if (ck != ClassB && type == 'l')
 | |
|       return quad ? "V2LLi" : "V1LLi";
 | |
| 
 | |
|     return quad ? "V16Sc" : "V8Sc";
 | |
|   }
 | |
| 
 | |
|   // Non-return array types are passed as individual vectors.
 | |
|   if (mod == '2')
 | |
|     return quad ? "V16ScV16Sc" : "V8ScV8Sc";
 | |
|   if (mod == '3')
 | |
|     return quad ? "V16ScV16ScV16Sc" : "V8ScV8ScV8Sc";
 | |
|   if (mod == '4')
 | |
|     return quad ? "V16ScV16ScV16ScV16Sc" : "V8ScV8ScV8ScV8Sc";
 | |
| 
 | |
|   if (mod == 'f' || (ck != ClassB && type == 'f'))
 | |
|     return quad ? "V4f" : "V2f";
 | |
|   if (ck != ClassB && type == 's')
 | |
|     return quad ? "V8s" : "V4s";
 | |
|   if (ck != ClassB && type == 'i')
 | |
|     return quad ? "V4i" : "V2i";
 | |
|   if (ck != ClassB && type == 'l')
 | |
|     return quad ? "V2LLi" : "V1LLi";
 | |
| 
 | |
|   return quad ? "V16Sc" : "V8Sc";
 | |
| }
 | |
| 
 | |
| /// InstructionTypeCode - Computes the ARM argument character code and
 | |
| /// quad status for a specific type string and ClassKind.
 | |
| static void InstructionTypeCode(const StringRef &typeStr,
 | |
|                                 const ClassKind ck,
 | |
|                                 bool &quad,
 | |
|                                 std::string &typeCode) {
 | |
|   bool poly = false;
 | |
|   bool usgn = false;
 | |
|   char type = ClassifyType(typeStr, quad, poly, usgn);
 | |
| 
 | |
|   switch (type) {
 | |
|   case 'c':
 | |
|     switch (ck) {
 | |
|     case ClassS: typeCode = poly ? "p8" : usgn ? "u8" : "s8"; break;
 | |
|     case ClassI: typeCode = "i8"; break;
 | |
|     case ClassW: typeCode = "8"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 's':
 | |
|     switch (ck) {
 | |
|     case ClassS: typeCode = poly ? "p16" : usgn ? "u16" : "s16"; break;
 | |
|     case ClassI: typeCode = "i16"; break;
 | |
|     case ClassW: typeCode = "16"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 'i':
 | |
|     switch (ck) {
 | |
|     case ClassS: typeCode = usgn ? "u32" : "s32"; break;
 | |
|     case ClassI: typeCode = "i32"; break;
 | |
|     case ClassW: typeCode = "32"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 'l':
 | |
|     switch (ck) {
 | |
|     case ClassS: typeCode = usgn ? "u64" : "s64"; break;
 | |
|     case ClassI: typeCode = "i64"; break;
 | |
|     case ClassW: typeCode = "64"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 'h':
 | |
|     switch (ck) {
 | |
|     case ClassS:
 | |
|     case ClassI: typeCode = "f16"; break;
 | |
|     case ClassW: typeCode = "16"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 'f':
 | |
|     switch (ck) {
 | |
|     case ClassS:
 | |
|     case ClassI: typeCode = "f32"; break;
 | |
|     case ClassW: typeCode = "32"; break;
 | |
|     default: break;
 | |
|     }
 | |
|     break;
 | |
|   case 'd':
 | |
|     switch (ck) {
 | |
|     case ClassS:
 | |
|     case ClassI:
 | |
|       typeCode += "f64";
 | |
|       break;
 | |
|     case ClassW:
 | |
|       PrintFatalError("unhandled type!");
 | |
|     default:
 | |
|       break;
 | |
|     }
 | |
|     break;
 | |
|   default:
 | |
|     PrintFatalError("unhandled type!");
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// MangleName - Append a type or width suffix to a base neon function name,
 | |
| /// and insert a 'q' in the appropriate location if the operation works on
 | |
| /// 128b rather than 64b.   E.g. turn "vst2_lane" into "vst2q_lane_f32", etc.
 | |
| static std::string MangleName(const std::string &name, StringRef typestr,
 | |
|                               ClassKind ck) {
 | |
|   if (name == "vcvt_f32_f16")
 | |
|     return name;
 | |
| 
 | |
|   bool quad = false;
 | |
|   std::string typeCode = "";
 | |
| 
 | |
|   InstructionTypeCode(typestr, ck, quad, typeCode);
 | |
| 
 | |
|   std::string s = name;
 | |
| 
 | |
|   if (typeCode.size() > 0) {
 | |
|     s += "_" + typeCode;
 | |
|   }
 | |
| 
 | |
|   if (ck == ClassB)
 | |
|     s += "_v";
 | |
| 
 | |
|   // Insert a 'q' before the first '_' character so that it ends up before
 | |
|   // _lane or _n on vector-scalar operations.
 | |
|   if (quad) {
 | |
|     size_t pos = s.find('_');
 | |
|     s = s.insert(pos, "q");
 | |
|   }
 | |
| 
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static void PreprocessInstruction(const StringRef &Name,
 | |
|                                   const std::string &InstName,
 | |
|                                   std::string &Prefix,
 | |
|                                   bool &HasNPostfix,
 | |
|                                   bool &HasLanePostfix,
 | |
|                                   bool &HasDupPostfix,
 | |
|                                   bool &IsSpecialVCvt,
 | |
|                                   size_t &TBNumber) {
 | |
|   // All of our instruction name fields from arm_neon.td are of the form
 | |
|   //   <instructionname>_...
 | |
|   // Thus we grab our instruction name via computation of said Prefix.
 | |
|   const size_t PrefixEnd = Name.find_first_of('_');
 | |
|   // If InstName is passed in, we use that instead of our name Prefix.
 | |
|   Prefix = InstName.size() == 0? Name.slice(0, PrefixEnd).str() : InstName;
 | |
| 
 | |
|   const StringRef Postfix = Name.slice(PrefixEnd, Name.size());
 | |
| 
 | |
|   HasNPostfix = Postfix.count("_n");
 | |
|   HasLanePostfix = Postfix.count("_lane");
 | |
|   HasDupPostfix = Postfix.count("_dup");
 | |
|   IsSpecialVCvt = Postfix.size() != 0 && Name.count("vcvt");
 | |
| 
 | |
|   if (InstName.compare("vtbl") == 0 ||
 | |
|       InstName.compare("vtbx") == 0) {
 | |
|     // If we have a vtblN/vtbxN instruction, use the instruction's ASCII
 | |
|     // encoding to get its true value.
 | |
|     TBNumber = Name[Name.size()-1] - 48;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// GenerateRegisterCheckPatternsForLoadStores - Given a bunch of data we have
 | |
| /// extracted, generate a FileCheck pattern for a Load Or Store
 | |
| static void
 | |
| GenerateRegisterCheckPatternForLoadStores(const StringRef &NameRef,
 | |
|                                           const std::string& OutTypeCode,
 | |
|                                           const bool &IsQuad,
 | |
|                                           const bool &HasDupPostfix,
 | |
|                                           const bool &HasLanePostfix,
 | |
|                                           const size_t Count,
 | |
|                                           std::string &RegisterSuffix) {
 | |
|   const bool IsLDSTOne = NameRef.count("vld1") || NameRef.count("vst1");
 | |
|   // If N == 3 || N == 4 and we are dealing with a quad instruction, Clang
 | |
|   // will output a series of v{ld,st}1s, so we have to handle it specially.
 | |
|   if ((Count == 3 || Count == 4) && IsQuad) {
 | |
|     RegisterSuffix += "{";
 | |
|     for (size_t i = 0; i < Count; i++) {
 | |
|       RegisterSuffix += "d{{[0-9]+}}";
 | |
|       if (HasDupPostfix) {
 | |
|         RegisterSuffix += "[]";
 | |
|       }
 | |
|       if (HasLanePostfix) {
 | |
|         RegisterSuffix += "[{{[0-9]+}}]";
 | |
|       }
 | |
|       if (i < Count-1) {
 | |
|         RegisterSuffix += ", ";
 | |
|       }
 | |
|     }
 | |
|     RegisterSuffix += "}";
 | |
|   } else {
 | |
| 
 | |
|     // Handle normal loads and stores.
 | |
|     RegisterSuffix += "{";
 | |
|     for (size_t i = 0; i < Count; i++) {
 | |
|       RegisterSuffix += "d{{[0-9]+}}";
 | |
|       if (HasDupPostfix) {
 | |
|         RegisterSuffix += "[]";
 | |
|       }
 | |
|       if (HasLanePostfix) {
 | |
|         RegisterSuffix += "[{{[0-9]+}}]";
 | |
|       }
 | |
|       if (IsQuad && !HasLanePostfix) {
 | |
|         RegisterSuffix += ", d{{[0-9]+}}";
 | |
|         if (HasDupPostfix) {
 | |
|           RegisterSuffix += "[]";
 | |
|         }
 | |
|       }
 | |
|       if (i < Count-1) {
 | |
|         RegisterSuffix += ", ";
 | |
|       }
 | |
|     }
 | |
|     RegisterSuffix += "}, [r{{[0-9]+}}";
 | |
| 
 | |
|     // We only include the alignment hint if we have a vld1.*64 or
 | |
|     // a dup/lane instruction.
 | |
|     if (IsLDSTOne) {
 | |
|       if ((HasLanePostfix || HasDupPostfix) && OutTypeCode != "8") {
 | |
|         RegisterSuffix += ":" + OutTypeCode;
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     RegisterSuffix += "]";
 | |
|   }
 | |
| }
 | |
| 
 | |
| static bool HasNPostfixAndScalarArgs(const StringRef &NameRef,
 | |
|                                      const bool &HasNPostfix) {
 | |
|   return (NameRef.count("vmla") ||
 | |
|           NameRef.count("vmlal") ||
 | |
|           NameRef.count("vmlsl") ||
 | |
|           NameRef.count("vmull") ||
 | |
|           NameRef.count("vqdmlal") ||
 | |
|           NameRef.count("vqdmlsl") ||
 | |
|           NameRef.count("vqdmulh") ||
 | |
|           NameRef.count("vqdmull") ||
 | |
|           NameRef.count("vqrdmulh")) && HasNPostfix;
 | |
| }
 | |
| 
 | |
| static bool IsFiveOperandLaneAccumulator(const StringRef &NameRef,
 | |
|                                          const bool &HasLanePostfix) {
 | |
|   return (NameRef.count("vmla") ||
 | |
|           NameRef.count("vmls") ||
 | |
|           NameRef.count("vmlal") ||
 | |
|           NameRef.count("vmlsl") ||
 | |
|           (NameRef.count("vmul") && NameRef.size() == 3)||
 | |
|           NameRef.count("vqdmlal") ||
 | |
|           NameRef.count("vqdmlsl") ||
 | |
|           NameRef.count("vqdmulh") ||
 | |
|           NameRef.count("vqrdmulh")) && HasLanePostfix;
 | |
| }
 | |
| 
 | |
| static bool IsSpecialLaneMultiply(const StringRef &NameRef,
 | |
|                                   const bool &HasLanePostfix,
 | |
|                                   const bool &IsQuad) {
 | |
|   const bool IsVMulOrMulh = (NameRef.count("vmul") || NameRef.count("mulh"))
 | |
|                                && IsQuad;
 | |
|   const bool IsVMull = NameRef.count("mull") && !IsQuad;
 | |
|   return (IsVMulOrMulh || IsVMull) && HasLanePostfix;
 | |
| }
 | |
| 
 | |
| static void NormalizeProtoForRegisterPatternCreation(const std::string &Name,
 | |
|                                                      const std::string &Proto,
 | |
|                                                      const bool &HasNPostfix,
 | |
|                                                      const bool &IsQuad,
 | |
|                                                      const bool &HasLanePostfix,
 | |
|                                                      const bool &HasDupPostfix,
 | |
|                                                      std::string &NormedProto) {
 | |
|   // Handle generic case.
 | |
|   const StringRef NameRef(Name);
 | |
|   for (size_t i = 0, end = Proto.size(); i < end; i++) {
 | |
|     switch (Proto[i]) {
 | |
|     case 'u':
 | |
|     case 'f':
 | |
|     case 'd':
 | |
|     case 's':
 | |
|     case 'x':
 | |
|     case 't':
 | |
|     case 'n':
 | |
|       NormedProto += IsQuad? 'q' : 'd';
 | |
|       break;
 | |
|     case 'w':
 | |
|     case 'k':
 | |
|       NormedProto += 'q';
 | |
|       break;
 | |
|     case 'g':
 | |
|     case 'h':
 | |
|     case 'e':
 | |
|       NormedProto += 'd';
 | |
|       break;
 | |
|     case 'i':
 | |
|       NormedProto += HasLanePostfix? 'a' : 'i';
 | |
|       break;
 | |
|     case 'a':
 | |
|       if (HasLanePostfix) {
 | |
|         NormedProto += 'a';
 | |
|       } else if (HasNPostfixAndScalarArgs(NameRef, HasNPostfix)) {
 | |
|         NormedProto += IsQuad? 'q' : 'd';
 | |
|       } else {
 | |
|         NormedProto += 'i';
 | |
|       }
 | |
|       break;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   // Handle Special Cases.
 | |
|   const bool IsNotVExt = !NameRef.count("vext");
 | |
|   const bool IsVPADAL = NameRef.count("vpadal");
 | |
|   const bool Is5OpLaneAccum = IsFiveOperandLaneAccumulator(NameRef,
 | |
|                                                            HasLanePostfix);
 | |
|   const bool IsSpecialLaneMul = IsSpecialLaneMultiply(NameRef, HasLanePostfix,
 | |
|                                                       IsQuad);
 | |
| 
 | |
|   if (IsSpecialLaneMul) {
 | |
|     // If
 | |
|     NormedProto[2] = NormedProto[3];
 | |
|     NormedProto.erase(3);
 | |
|   } else if (NormedProto.size() == 4 &&
 | |
|              NormedProto[0] == NormedProto[1] &&
 | |
|              IsNotVExt) {
 | |
|     // If NormedProto.size() == 4 and the first two proto characters are the
 | |
|     // same, ignore the first.
 | |
|     NormedProto = NormedProto.substr(1, 3);
 | |
|   } else if (Is5OpLaneAccum) {
 | |
|     // If we have a 5 op lane accumulator operation, we take characters 1,2,4
 | |
|     std::string tmp = NormedProto.substr(1,2);
 | |
|     tmp += NormedProto[4];
 | |
|     NormedProto = tmp;
 | |
|   } else if (IsVPADAL) {
 | |
|     // If we have VPADAL, ignore the first character.
 | |
|     NormedProto = NormedProto.substr(0, 2);
 | |
|   } else if (NameRef.count("vdup") && NormedProto.size() > 2) {
 | |
|     // If our instruction is a dup instruction, keep only the first and
 | |
|     // last characters.
 | |
|     std::string tmp = "";
 | |
|     tmp += NormedProto[0];
 | |
|     tmp += NormedProto[NormedProto.size()-1];
 | |
|     NormedProto = tmp;
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// GenerateRegisterCheckPatterns - Given a bunch of data we have
 | |
| /// extracted, generate a FileCheck pattern to check that an
 | |
| /// instruction's arguments are correct.
 | |
| static void GenerateRegisterCheckPattern(const std::string &Name,
 | |
|                                          const std::string &Proto,
 | |
|                                          const std::string &OutTypeCode,
 | |
|                                          const bool &HasNPostfix,
 | |
|                                          const bool &IsQuad,
 | |
|                                          const bool &HasLanePostfix,
 | |
|                                          const bool &HasDupPostfix,
 | |
|                                          const size_t &TBNumber,
 | |
|                                          std::string &RegisterSuffix) {
 | |
| 
 | |
|   RegisterSuffix = "";
 | |
| 
 | |
|   const StringRef NameRef(Name);
 | |
|   const StringRef ProtoRef(Proto);
 | |
| 
 | |
|   if ((NameRef.count("vdup") || NameRef.count("vmov")) && HasNPostfix) {
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   const bool IsLoadStore = NameRef.count("vld") || NameRef.count("vst");
 | |
|   const bool IsTBXOrTBL = NameRef.count("vtbl") || NameRef.count("vtbx");
 | |
| 
 | |
|   if (IsLoadStore) {
 | |
|     // Grab N value from  v{ld,st}N using its ascii representation.
 | |
|     const size_t Count = NameRef[3] - 48;
 | |
| 
 | |
|     GenerateRegisterCheckPatternForLoadStores(NameRef, OutTypeCode, IsQuad,
 | |
|                                               HasDupPostfix, HasLanePostfix,
 | |
|                                               Count, RegisterSuffix);
 | |
|   } else if (IsTBXOrTBL) {
 | |
|     RegisterSuffix += "d{{[0-9]+}}, {";
 | |
|     for (size_t i = 0; i < TBNumber-1; i++) {
 | |
|       RegisterSuffix += "d{{[0-9]+}}, ";
 | |
|     }
 | |
|     RegisterSuffix += "d{{[0-9]+}}}, d{{[0-9]+}}";
 | |
|   } else {
 | |
|     // Handle a normal instruction.
 | |
|     if (NameRef.count("vget") || NameRef.count("vset"))
 | |
|       return;
 | |
| 
 | |
|     // We first normalize our proto, since we only need to emit 4
 | |
|     // different types of checks, yet have more than 4 proto types
 | |
|     // that map onto those 4 patterns.
 | |
|     std::string NormalizedProto("");
 | |
|     NormalizeProtoForRegisterPatternCreation(Name, Proto, HasNPostfix, IsQuad,
 | |
|                                              HasLanePostfix, HasDupPostfix,
 | |
|                                              NormalizedProto);
 | |
| 
 | |
|     for (size_t i = 0, end = NormalizedProto.size(); i < end; i++) {
 | |
|       const char &c = NormalizedProto[i];
 | |
|       switch (c) {
 | |
|       case 'q':
 | |
|         RegisterSuffix += "q{{[0-9]+}}, ";
 | |
|         break;
 | |
| 
 | |
|       case 'd':
 | |
|         RegisterSuffix += "d{{[0-9]+}}, ";
 | |
|         break;
 | |
| 
 | |
|       case 'i':
 | |
|         RegisterSuffix += "#{{[0-9]+}}, ";
 | |
|         break;
 | |
| 
 | |
|       case 'a':
 | |
|         RegisterSuffix += "d{{[0-9]+}}[{{[0-9]}}], ";
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     // Remove extra ", ".
 | |
|     RegisterSuffix = RegisterSuffix.substr(0, RegisterSuffix.size()-2);
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// GenerateChecksForIntrinsic - Given a specific instruction name +
 | |
| /// typestr + class kind, generate the proper set of FileCheck
 | |
| /// Patterns to check for. We could just return a string, but instead
 | |
| /// use a vector since it provides us with the extra flexibility of
 | |
| /// emitting multiple checks, which comes in handy for certain cases
 | |
| /// like mla where we want to check for 2 different instructions.
 | |
| static void GenerateChecksForIntrinsic(const std::string &Name,
 | |
|                                        const std::string &Proto,
 | |
|                                        StringRef &OutTypeStr,
 | |
|                                        StringRef &InTypeStr,
 | |
|                                        ClassKind Ck,
 | |
|                                        const std::string &InstName,
 | |
|                                        bool IsHiddenLOp,
 | |
|                                        std::vector<std::string>& Result) {
 | |
| 
 | |
|   // If Ck is a ClassNoTest instruction, just return so no test is
 | |
|   // emitted.
 | |
|   if(Ck == ClassNoTest)
 | |
|     return;
 | |
| 
 | |
|   if (Name == "vcvt_f32_f16") {
 | |
|     Result.push_back("vcvt.f32.f16");
 | |
|     return;
 | |
|   }
 | |
| 
 | |
| 
 | |
|   // Now we preprocess our instruction given the data we have to get the
 | |
|   // data that we need.
 | |
|   // Create a StringRef for String Manipulation of our Name.
 | |
|   const StringRef NameRef(Name);
 | |
|   // Instruction Prefix.
 | |
|   std::string Prefix;
 | |
|   // The type code for our out type string.
 | |
|   std::string OutTypeCode;
 | |
|   // To handle our different cases, we need to check for different postfixes.
 | |
|   // Is our instruction a quad instruction.
 | |
|   bool IsQuad = false;
 | |
|   // Our instruction is of the form <instructionname>_n.
 | |
|   bool HasNPostfix = false;
 | |
|   // Our instruction is of the form <instructionname>_lane.
 | |
|   bool HasLanePostfix = false;
 | |
|   // Our instruction is of the form <instructionname>_dup.
 | |
|   bool HasDupPostfix  = false;
 | |
|   // Our instruction is a vcvt instruction which requires special handling.
 | |
|   bool IsSpecialVCvt = false;
 | |
|   // If we have a vtbxN or vtblN instruction, this is set to N.
 | |
|   size_t TBNumber = -1;
 | |
|   // Register Suffix
 | |
|   std::string RegisterSuffix;
 | |
| 
 | |
|   PreprocessInstruction(NameRef, InstName, Prefix,
 | |
|                         HasNPostfix, HasLanePostfix, HasDupPostfix,
 | |
|                         IsSpecialVCvt, TBNumber);
 | |
| 
 | |
|   InstructionTypeCode(OutTypeStr, Ck, IsQuad, OutTypeCode);
 | |
|   GenerateRegisterCheckPattern(Name, Proto, OutTypeCode, HasNPostfix, IsQuad,
 | |
|                                HasLanePostfix, HasDupPostfix, TBNumber,
 | |
|                                RegisterSuffix);
 | |
| 
 | |
|   // In the following section, we handle a bunch of special cases. You can tell
 | |
|   // a special case by the fact we are returning early.
 | |
| 
 | |
|   // If our instruction is a logical instruction without postfix or a
 | |
|   // hidden LOp just return the current Prefix.
 | |
|   if (Ck == ClassL || IsHiddenLOp) {
 | |
|     Result.push_back(Prefix + " " + RegisterSuffix);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   // If we have a vmov, due to the many different cases, some of which
 | |
|   // vary within the different intrinsics generated for a single
 | |
|   // instruction type, just output a vmov. (e.g. given an instruction
 | |
|   // A, A.u32 might be vmov and A.u8 might be vmov.8).
 | |
|   //
 | |
|   // FIXME: Maybe something can be done about this. The two cases that we care
 | |
|   // about are vmov as an LType and vmov as a WType.
 | |
|   if (Prefix == "vmov") {
 | |
|     Result.push_back(Prefix + " " + RegisterSuffix);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   // In the following section, we handle special cases.
 | |
| 
 | |
|   if (OutTypeCode == "64") {
 | |
|     // If we have a 64 bit vdup/vext and are handling an uint64x1_t
 | |
|     // type, the intrinsic will be optimized away, so just return
 | |
|     // nothing.  On the other hand if we are handling an uint64x2_t
 | |
|     // (i.e. quad instruction), vdup/vmov instructions should be
 | |
|     // emitted.
 | |
|     if (Prefix == "vdup" || Prefix == "vext") {
 | |
|       if (IsQuad) {
 | |
|         Result.push_back("{{vmov|vdup}}");
 | |
|       }
 | |
|       return;
 | |
|     }
 | |
| 
 | |
|     // v{st,ld}{2,3,4}_{u,s}64 emit v{st,ld}1.64 instructions with
 | |
|     // multiple register operands.
 | |
|     bool MultiLoadPrefix = Prefix == "vld2" || Prefix == "vld3"
 | |
|                             || Prefix == "vld4";
 | |
|     bool MultiStorePrefix = Prefix == "vst2" || Prefix == "vst3"
 | |
|                             || Prefix == "vst4";
 | |
|     if (MultiLoadPrefix || MultiStorePrefix) {
 | |
|       Result.push_back(NameRef.slice(0, 3).str() + "1.64");
 | |
|       return;
 | |
|     }
 | |
| 
 | |
|     // v{st,ld}1_{lane,dup}_{u64,s64} use vldr/vstr/vmov/str instead of
 | |
|     // emitting said instructions. So return a check for
 | |
|     // vldr/vstr/vmov/str instead.
 | |
|     if (HasLanePostfix || HasDupPostfix) {
 | |
|       if (Prefix == "vst1") {
 | |
|         Result.push_back("{{str|vstr|vmov}}");
 | |
|         return;
 | |
|       } else if (Prefix == "vld1") {
 | |
|         Result.push_back("{{ldr|vldr|vmov}}");
 | |
|         return;
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   // vzip.32/vuzp.32 are the same instruction as vtrn.32 and are
 | |
|   // sometimes disassembled as vtrn.32. We use a regex to handle both
 | |
|   // cases.
 | |
|   if ((Prefix == "vzip" || Prefix == "vuzp") && OutTypeCode == "32") {
 | |
|     Result.push_back("{{vtrn|" + Prefix + "}}.32 " + RegisterSuffix);
 | |
|     return;
 | |
|   }
 | |
| 
 | |
|   // Currently on most ARM processors, we do not use vmla/vmls for
 | |
|   // quad floating point operations. Instead we output vmul + vadd. So
 | |
|   // check if we have one of those instructions and just output a
 | |
|   // check for vmul.
 | |
|   if (OutTypeCode == "f32") {
 | |
|     if (Prefix == "vmls") {
 | |
|       Result.push_back("vmul." + OutTypeCode + " " + RegisterSuffix);
 | |
|       Result.push_back("vsub." + OutTypeCode);
 | |
|       return;
 | |
|     } else if (Prefix == "vmla") {
 | |
|       Result.push_back("vmul." + OutTypeCode + " " + RegisterSuffix);
 | |
|       Result.push_back("vadd." + OutTypeCode);
 | |
|       return;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   // If we have vcvt, get the input type from the instruction name
 | |
|   // (which should be of the form instname_inputtype) and append it
 | |
|   // before the output type.
 | |
|   if (Prefix == "vcvt") {
 | |
|     const std::string inTypeCode = NameRef.substr(NameRef.find_last_of("_")+1);
 | |
|     Prefix += "." + inTypeCode;
 | |
|   }
 | |
| 
 | |
|   // Append output type code to get our final mangled instruction.
 | |
|   Prefix += "." + OutTypeCode;
 | |
| 
 | |
|   Result.push_back(Prefix + " " + RegisterSuffix);
 | |
| }
 | |
| 
 | |
| /// UseMacro - Examine the prototype string to determine if the intrinsic
 | |
| /// should be defined as a preprocessor macro instead of an inline function.
 | |
| static bool UseMacro(const std::string &proto) {
 | |
|   // If this builtin takes an immediate argument, we need to #define it rather
 | |
|   // than use a standard declaration, so that SemaChecking can range check
 | |
|   // the immediate passed by the user.
 | |
|   if (proto.find('i') != std::string::npos)
 | |
|     return true;
 | |
| 
 | |
|   // Pointer arguments need to use macros to avoid hiding aligned attributes
 | |
|   // from the pointer type.
 | |
|   if (proto.find('p') != std::string::npos ||
 | |
|       proto.find('c') != std::string::npos)
 | |
|     return true;
 | |
| 
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| /// MacroArgUsedDirectly - Return true if argument i for an intrinsic that is
 | |
| /// defined as a macro should be accessed directly instead of being first
 | |
| /// assigned to a local temporary.
 | |
| static bool MacroArgUsedDirectly(const std::string &proto, unsigned i) {
 | |
|   // True for constant ints (i), pointers (p) and const pointers (c).
 | |
|   return (proto[i] == 'i' || proto[i] == 'p' || proto[i] == 'c');
 | |
| }
 | |
| 
 | |
| // Generate the string "(argtype a, argtype b, ...)"
 | |
| static std::string GenArgs(const std::string &proto, StringRef typestr) {
 | |
|   bool define = UseMacro(proto);
 | |
|   char arg = 'a';
 | |
| 
 | |
|   std::string s;
 | |
|   s += "(";
 | |
| 
 | |
|   for (unsigned i = 1, e = proto.size(); i != e; ++i, ++arg) {
 | |
|     if (define) {
 | |
|       // Some macro arguments are used directly instead of being assigned
 | |
|       // to local temporaries; prepend an underscore prefix to make their
 | |
|       // names consistent with the local temporaries.
 | |
|       if (MacroArgUsedDirectly(proto, i))
 | |
|         s += "__";
 | |
|     } else {
 | |
|       s += TypeString(proto[i], typestr) + " __";
 | |
|     }
 | |
|     s.push_back(arg);
 | |
|     if ((i + 1) < e)
 | |
|       s += ", ";
 | |
|   }
 | |
| 
 | |
|   s += ")";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| // Macro arguments are not type-checked like inline function arguments, so
 | |
| // assign them to local temporaries to get the right type checking.
 | |
| static std::string GenMacroLocals(const std::string &proto, StringRef typestr) {
 | |
|   char arg = 'a';
 | |
|   std::string s;
 | |
|   bool generatedLocal = false;
 | |
| 
 | |
|   for (unsigned i = 1, e = proto.size(); i != e; ++i, ++arg) {
 | |
|     // Do not create a temporary for an immediate argument.
 | |
|     // That would defeat the whole point of using a macro!
 | |
|     if (MacroArgUsedDirectly(proto, i))
 | |
|       continue;
 | |
|     generatedLocal = true;
 | |
| 
 | |
|     s += TypeString(proto[i], typestr) + " __";
 | |
|     s.push_back(arg);
 | |
|     s += " = (";
 | |
|     s.push_back(arg);
 | |
|     s += "); ";
 | |
|   }
 | |
| 
 | |
|   if (generatedLocal)
 | |
|     s += "\\\n  ";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| // Use the vmovl builtin to sign-extend or zero-extend a vector.
 | |
| static std::string Extend(StringRef typestr, const std::string &a) {
 | |
|   std::string s;
 | |
|   s = MangleName("vmovl", typestr, ClassS);
 | |
|   s += "(" + a + ")";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static std::string Duplicate(unsigned nElts, StringRef typestr,
 | |
|                              const std::string &a) {
 | |
|   std::string s;
 | |
| 
 | |
|   s = "(" + TypeString('d', typestr) + "){ ";
 | |
|   for (unsigned i = 0; i != nElts; ++i) {
 | |
|     s += a;
 | |
|     if ((i + 1) < nElts)
 | |
|       s += ", ";
 | |
|   }
 | |
|   s += " }";
 | |
| 
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static std::string SplatLane(unsigned nElts, const std::string &vec,
 | |
|                              const std::string &lane) {
 | |
|   std::string s = "__builtin_shufflevector(" + vec + ", " + vec;
 | |
|   for (unsigned i = 0; i < nElts; ++i)
 | |
|     s += ", " + lane;
 | |
|   s += ")";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static unsigned GetNumElements(StringRef typestr, bool &quad) {
 | |
|   quad = false;
 | |
|   bool dummy = false;
 | |
|   char type = ClassifyType(typestr, quad, dummy, dummy);
 | |
|   unsigned nElts = 0;
 | |
|   switch (type) {
 | |
|   case 'c': nElts = 8; break;
 | |
|   case 's': nElts = 4; break;
 | |
|   case 'i': nElts = 2; break;
 | |
|   case 'l': nElts = 1; break;
 | |
|   case 'h': nElts = 4; break;
 | |
|   case 'f': nElts = 2; break;
 | |
|   case 'd':
 | |
|     nElts = 1;
 | |
|     break;
 | |
|   default:
 | |
|     PrintFatalError("unhandled type!");
 | |
|   }
 | |
|   if (quad) nElts <<= 1;
 | |
|   return nElts;
 | |
| }
 | |
| 
 | |
| // Generate the definition for this intrinsic, e.g. "a + b" for OpAdd.
 | |
| static std::string GenOpString(OpKind op, const std::string &proto,
 | |
|                                StringRef typestr) {
 | |
|   bool quad;
 | |
|   unsigned nElts = GetNumElements(typestr, quad);
 | |
|   bool define = UseMacro(proto);
 | |
| 
 | |
|   std::string ts = TypeString(proto[0], typestr);
 | |
|   std::string s;
 | |
|   if (!define) {
 | |
|     s = "return ";
 | |
|   }
 | |
| 
 | |
|   switch(op) {
 | |
|   case OpAdd:
 | |
|     s += "__a + __b;";
 | |
|     break;
 | |
|   case OpAddl:
 | |
|     s += Extend(typestr, "__a") + " + " + Extend(typestr, "__b") + ";";
 | |
|     break;
 | |
|   case OpAddw:
 | |
|     s += "__a + " + Extend(typestr, "__b") + ";";
 | |
|     break;
 | |
|   case OpSub:
 | |
|     s += "__a - __b;";
 | |
|     break;
 | |
|   case OpSubl:
 | |
|     s += Extend(typestr, "__a") + " - " + Extend(typestr, "__b") + ";";
 | |
|     break;
 | |
|   case OpSubw:
 | |
|     s += "__a - " + Extend(typestr, "__b") + ";";
 | |
|     break;
 | |
|   case OpMulN:
 | |
|     s += "__a * " + Duplicate(nElts, typestr, "__b") + ";";
 | |
|     break;
 | |
|   case OpMulLane:
 | |
|     s += "__a * " + SplatLane(nElts, "__b", "__c") + ";";
 | |
|     break;
 | |
|   case OpMul:
 | |
|     s += "__a * __b;";
 | |
|     break;
 | |
|   case OpMullLane:
 | |
|     s += MangleName("vmull", typestr, ClassS) + "(__a, " +
 | |
|       SplatLane(nElts, "__b", "__c") + ");";
 | |
|     break;
 | |
|   case OpMlaN:
 | |
|     s += "__a + (__b * " + Duplicate(nElts, typestr, "__c") + ");";
 | |
|     break;
 | |
|   case OpMlaLane:
 | |
|     s += "__a + (__b * " + SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpMla:
 | |
|     s += "__a + (__b * __c);";
 | |
|     break;
 | |
|   case OpMlalN:
 | |
|     s += "__a + " + MangleName("vmull", typestr, ClassS) + "(__b, " +
 | |
|       Duplicate(nElts, typestr, "__c") + ");";
 | |
|     break;
 | |
|   case OpMlalLane:
 | |
|     s += "__a + " + MangleName("vmull", typestr, ClassS) + "(__b, " +
 | |
|       SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpMlal:
 | |
|     s += "__a + " + MangleName("vmull", typestr, ClassS) + "(__b, __c);";
 | |
|     break;
 | |
|   case OpMlsN:
 | |
|     s += "__a - (__b * " + Duplicate(nElts, typestr, "__c") + ");";
 | |
|     break;
 | |
|   case OpMlsLane:
 | |
|     s += "__a - (__b * " + SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpMls:
 | |
|     s += "__a - (__b * __c);";
 | |
|     break;
 | |
|   case OpMlslN:
 | |
|     s += "__a - " + MangleName("vmull", typestr, ClassS) + "(__b, " +
 | |
|       Duplicate(nElts, typestr, "__c") + ");";
 | |
|     break;
 | |
|   case OpMlslLane:
 | |
|     s += "__a - " + MangleName("vmull", typestr, ClassS) + "(__b, " +
 | |
|       SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpMlsl:
 | |
|     s += "__a - " + MangleName("vmull", typestr, ClassS) + "(__b, __c);";
 | |
|     break;
 | |
|   case OpQDMullLane:
 | |
|     s += MangleName("vqdmull", typestr, ClassS) + "(__a, " +
 | |
|       SplatLane(nElts, "__b", "__c") + ");";
 | |
|     break;
 | |
|   case OpQDMlalLane:
 | |
|     s += MangleName("vqdmlal", typestr, ClassS) + "(__a, __b, " +
 | |
|       SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpQDMlslLane:
 | |
|     s += MangleName("vqdmlsl", typestr, ClassS) + "(__a, __b, " +
 | |
|       SplatLane(nElts, "__c", "__d") + ");";
 | |
|     break;
 | |
|   case OpQDMulhLane:
 | |
|     s += MangleName("vqdmulh", typestr, ClassS) + "(__a, " +
 | |
|       SplatLane(nElts, "__b", "__c") + ");";
 | |
|     break;
 | |
|   case OpQRDMulhLane:
 | |
|     s += MangleName("vqrdmulh", typestr, ClassS) + "(__a, " +
 | |
|       SplatLane(nElts, "__b", "__c") + ");";
 | |
|     break;
 | |
|   case OpEq:
 | |
|     s += "(" + ts + ")(__a == __b);";
 | |
|     break;
 | |
|   case OpGe:
 | |
|     s += "(" + ts + ")(__a >= __b);";
 | |
|     break;
 | |
|   case OpLe:
 | |
|     s += "(" + ts + ")(__a <= __b);";
 | |
|     break;
 | |
|   case OpGt:
 | |
|     s += "(" + ts + ")(__a > __b);";
 | |
|     break;
 | |
|   case OpLt:
 | |
|     s += "(" + ts + ")(__a < __b);";
 | |
|     break;
 | |
|   case OpNeg:
 | |
|     s += " -__a;";
 | |
|     break;
 | |
|   case OpNot:
 | |
|     s += " ~__a;";
 | |
|     break;
 | |
|   case OpAnd:
 | |
|     s += "__a & __b;";
 | |
|     break;
 | |
|   case OpOr:
 | |
|     s += "__a | __b;";
 | |
|     break;
 | |
|   case OpXor:
 | |
|     s += "__a ^ __b;";
 | |
|     break;
 | |
|   case OpAndNot:
 | |
|     s += "__a & ~__b;";
 | |
|     break;
 | |
|   case OpOrNot:
 | |
|     s += "__a | ~__b;";
 | |
|     break;
 | |
|   case OpCast:
 | |
|     s += "(" + ts + ")__a;";
 | |
|     break;
 | |
|   case OpConcat:
 | |
|     s += "(" + ts + ")__builtin_shufflevector((int64x1_t)__a";
 | |
|     s += ", (int64x1_t)__b, 0, 1);";
 | |
|     break;
 | |
|   case OpHi:
 | |
|     // nElts is for the result vector, so the source is twice that number.
 | |
|     s += "__builtin_shufflevector(__a, __a";
 | |
|     for (unsigned i = nElts; i < nElts * 2; ++i)
 | |
|       s += ", " + utostr(i);
 | |
|     s+= ");";
 | |
|     break;
 | |
|   case OpLo:
 | |
|     s += "__builtin_shufflevector(__a, __a";
 | |
|     for (unsigned i = 0; i < nElts; ++i)
 | |
|       s += ", " + utostr(i);
 | |
|     s+= ");";
 | |
|     break;
 | |
|   case OpDup:
 | |
|     s += Duplicate(nElts, typestr, "__a") + ";";
 | |
|     break;
 | |
|   case OpDupLane:
 | |
|     s += SplatLane(nElts, "__a", "__b") + ";";
 | |
|     break;
 | |
|   case OpSelect:
 | |
|     // ((0 & 1) | (~0 & 2))
 | |
|     s += "(" + ts + ")";
 | |
|     ts = TypeString(proto[1], typestr);
 | |
|     s += "((__a & (" + ts + ")__b) | ";
 | |
|     s += "(~__a & (" + ts + ")__c));";
 | |
|     break;
 | |
|   case OpRev16:
 | |
|     s += "__builtin_shufflevector(__a, __a";
 | |
|     for (unsigned i = 2; i <= nElts; i += 2)
 | |
|       for (unsigned j = 0; j != 2; ++j)
 | |
|         s += ", " + utostr(i - j - 1);
 | |
|     s += ");";
 | |
|     break;
 | |
|   case OpRev32: {
 | |
|     unsigned WordElts = nElts >> (1 + (int)quad);
 | |
|     s += "__builtin_shufflevector(__a, __a";
 | |
|     for (unsigned i = WordElts; i <= nElts; i += WordElts)
 | |
|       for (unsigned j = 0; j != WordElts; ++j)
 | |
|         s += ", " + utostr(i - j - 1);
 | |
|     s += ");";
 | |
|     break;
 | |
|   }
 | |
|   case OpRev64: {
 | |
|     unsigned DblWordElts = nElts >> (int)quad;
 | |
|     s += "__builtin_shufflevector(__a, __a";
 | |
|     for (unsigned i = DblWordElts; i <= nElts; i += DblWordElts)
 | |
|       for (unsigned j = 0; j != DblWordElts; ++j)
 | |
|         s += ", " + utostr(i - j - 1);
 | |
|     s += ");";
 | |
|     break;
 | |
|   }
 | |
|   case OpAbdl: {
 | |
|     std::string abd = MangleName("vabd", typestr, ClassS) + "(__a, __b)";
 | |
|     if (typestr[0] != 'U') {
 | |
|       // vabd results are always unsigned and must be zero-extended.
 | |
|       std::string utype = "U" + typestr.str();
 | |
|       s += "(" + TypeString(proto[0], typestr) + ")";
 | |
|       abd = "(" + TypeString('d', utype) + ")" + abd;
 | |
|       s += Extend(utype, abd) + ";";
 | |
|     } else {
 | |
|       s += Extend(typestr, abd) + ";";
 | |
|     }
 | |
|     break;
 | |
|   }
 | |
|   case OpAba:
 | |
|     s += "__a + " + MangleName("vabd", typestr, ClassS) + "(__b, __c);";
 | |
|     break;
 | |
|   case OpAbal: {
 | |
|     s += "__a + ";
 | |
|     std::string abd = MangleName("vabd", typestr, ClassS) + "(__b, __c)";
 | |
|     if (typestr[0] != 'U') {
 | |
|       // vabd results are always unsigned and must be zero-extended.
 | |
|       std::string utype = "U" + typestr.str();
 | |
|       s += "(" + TypeString(proto[0], typestr) + ")";
 | |
|       abd = "(" + TypeString('d', utype) + ")" + abd;
 | |
|       s += Extend(utype, abd) + ";";
 | |
|     } else {
 | |
|       s += Extend(typestr, abd) + ";";
 | |
|     }
 | |
|     break;
 | |
|   }
 | |
|   case OpDiv:
 | |
|     s += "__a / __b;";
 | |
|     break;
 | |
|   default:
 | |
|     PrintFatalError("unknown OpKind!");
 | |
|   }
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static unsigned GetNeonEnum(const std::string &proto, StringRef typestr) {
 | |
|   unsigned mod = proto[0];
 | |
| 
 | |
|   if (mod == 'v' || mod == 'f')
 | |
|     mod = proto[1];
 | |
| 
 | |
|   bool quad = false;
 | |
|   bool poly = false;
 | |
|   bool usgn = false;
 | |
|   bool scal = false;
 | |
|   bool cnst = false;
 | |
|   bool pntr = false;
 | |
| 
 | |
|   // Base type to get the type string for.
 | |
|   char type = ClassifyType(typestr, quad, poly, usgn);
 | |
| 
 | |
|   // Based on the modifying character, change the type and width if necessary.
 | |
|   type = ModType(mod, type, quad, poly, usgn, scal, cnst, pntr);
 | |
| 
 | |
|   NeonTypeFlags::EltType ET;
 | |
|   switch (type) {
 | |
|     case 'c':
 | |
|       ET = poly ? NeonTypeFlags::Poly8 : NeonTypeFlags::Int8;
 | |
|       break;
 | |
|     case 's':
 | |
|       ET = poly ? NeonTypeFlags::Poly16 : NeonTypeFlags::Int16;
 | |
|       break;
 | |
|     case 'i':
 | |
|       ET = NeonTypeFlags::Int32;
 | |
|       break;
 | |
|     case 'l':
 | |
|       ET = NeonTypeFlags::Int64;
 | |
|       break;
 | |
|     case 'h':
 | |
|       ET = NeonTypeFlags::Float16;
 | |
|       break;
 | |
|     case 'f':
 | |
|       ET = NeonTypeFlags::Float32;
 | |
|       break;
 | |
|     case 'd':
 | |
|       ET = NeonTypeFlags::Float64;
 | |
|       break;
 | |
|     default:
 | |
|       PrintFatalError("unhandled type!");
 | |
|   }
 | |
|   NeonTypeFlags Flags(ET, usgn, quad && proto[1] != 'g');
 | |
|   return Flags.getFlags();
 | |
| }
 | |
| 
 | |
| // Generate the definition for this intrinsic, e.g. __builtin_neon_cls(a)
 | |
| static std::string GenBuiltin(const std::string &name, const std::string &proto,
 | |
|                               StringRef typestr, ClassKind ck) {
 | |
|   std::string s;
 | |
| 
 | |
|   // If this builtin returns a struct 2, 3, or 4 vectors, pass it as an implicit
 | |
|   // sret-like argument.
 | |
|   bool sret = (proto[0] >= '2' && proto[0] <= '4');
 | |
| 
 | |
|   bool define = UseMacro(proto);
 | |
| 
 | |
|   // Check if the prototype has a scalar operand with the type of the vector
 | |
|   // elements.  If not, bitcasting the args will take care of arg checking.
 | |
|   // The actual signedness etc. will be taken care of with special enums.
 | |
|   if (proto.find('s') == std::string::npos)
 | |
|     ck = ClassB;
 | |
| 
 | |
|   if (proto[0] != 'v') {
 | |
|     std::string ts = TypeString(proto[0], typestr);
 | |
| 
 | |
|     if (define) {
 | |
|       if (sret)
 | |
|         s += ts + " r; ";
 | |
|       else
 | |
|         s += "(" + ts + ")";
 | |
|     } else if (sret) {
 | |
|       s += ts + " r; ";
 | |
|     } else {
 | |
|       s += "return (" + ts + ")";
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   bool splat = proto.find('a') != std::string::npos;
 | |
| 
 | |
|   s += "__builtin_neon_";
 | |
|   if (splat) {
 | |
|     // Call the non-splat builtin: chop off the "_n" suffix from the name.
 | |
|     std::string vname(name, 0, name.size()-2);
 | |
|     s += MangleName(vname, typestr, ck);
 | |
|   } else {
 | |
|     s += MangleName(name, typestr, ck);
 | |
|   }
 | |
|   s += "(";
 | |
| 
 | |
|   // Pass the address of the return variable as the first argument to sret-like
 | |
|   // builtins.
 | |
|   if (sret)
 | |
|     s += "&r, ";
 | |
| 
 | |
|   char arg = 'a';
 | |
|   for (unsigned i = 1, e = proto.size(); i != e; ++i, ++arg) {
 | |
|     std::string args = std::string(&arg, 1);
 | |
| 
 | |
|     // Use the local temporaries instead of the macro arguments.
 | |
|     args = "__" + args;
 | |
| 
 | |
|     bool argQuad = false;
 | |
|     bool argPoly = false;
 | |
|     bool argUsgn = false;
 | |
|     bool argScalar = false;
 | |
|     bool dummy = false;
 | |
|     char argType = ClassifyType(typestr, argQuad, argPoly, argUsgn);
 | |
|     argType = ModType(proto[i], argType, argQuad, argPoly, argUsgn, argScalar,
 | |
|                       dummy, dummy);
 | |
| 
 | |
|     // Handle multiple-vector values specially, emitting each subvector as an
 | |
|     // argument to the __builtin.
 | |
|     if (proto[i] >= '2' && proto[i] <= '4') {
 | |
|       // Check if an explicit cast is needed.
 | |
|       if (argType != 'c' || argPoly || argUsgn)
 | |
|         args = (argQuad ? "(int8x16_t)" : "(int8x8_t)") + args;
 | |
| 
 | |
|       for (unsigned vi = 0, ve = proto[i] - '0'; vi != ve; ++vi) {
 | |
|         s += args + ".val[" + utostr(vi) + "]";
 | |
|         if ((vi + 1) < ve)
 | |
|           s += ", ";
 | |
|       }
 | |
|       if ((i + 1) < e)
 | |
|         s += ", ";
 | |
| 
 | |
|       continue;
 | |
|     }
 | |
| 
 | |
|     if (splat && (i + 1) == e)
 | |
|       args = Duplicate(GetNumElements(typestr, argQuad), typestr, args);
 | |
| 
 | |
|     // Check if an explicit cast is needed.
 | |
|     if ((splat || !argScalar) &&
 | |
|         ((ck == ClassB && argType != 'c') || argPoly || argUsgn)) {
 | |
|       std::string argTypeStr = "c";
 | |
|       if (ck != ClassB)
 | |
|         argTypeStr = argType;
 | |
|       if (argQuad)
 | |
|         argTypeStr = "Q" + argTypeStr;
 | |
|       args = "(" + TypeString('d', argTypeStr) + ")" + args;
 | |
|     }
 | |
| 
 | |
|     s += args;
 | |
|     if ((i + 1) < e)
 | |
|       s += ", ";
 | |
|   }
 | |
| 
 | |
|   // Extra constant integer to hold type class enum for this function, e.g. s8
 | |
|   if (ck == ClassB)
 | |
|     s += ", " + utostr(GetNeonEnum(proto, typestr));
 | |
| 
 | |
|   s += ");";
 | |
| 
 | |
|   if (proto[0] != 'v' && sret) {
 | |
|     if (define)
 | |
|       s += " r;";
 | |
|     else
 | |
|       s += " return r;";
 | |
|   }
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static std::string GenBuiltinDef(const std::string &name,
 | |
|                                  const std::string &proto,
 | |
|                                  StringRef typestr, ClassKind ck) {
 | |
|   std::string s("BUILTIN(__builtin_neon_");
 | |
| 
 | |
|   // If all types are the same size, bitcasting the args will take care
 | |
|   // of arg checking.  The actual signedness etc. will be taken care of with
 | |
|   // special enums.
 | |
|   if (proto.find('s') == std::string::npos)
 | |
|     ck = ClassB;
 | |
| 
 | |
|   s += MangleName(name, typestr, ck);
 | |
|   s += ", \"";
 | |
| 
 | |
|   for (unsigned i = 0, e = proto.size(); i != e; ++i)
 | |
|     s += BuiltinTypeString(proto[i], typestr, ck, i == 0);
 | |
| 
 | |
|   // Extra constant integer to hold type class enum for this function, e.g. s8
 | |
|   if (ck == ClassB)
 | |
|     s += "i";
 | |
| 
 | |
|   s += "\", \"n\")";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| static std::string GenIntrinsic(const std::string &name,
 | |
|                                 const std::string &proto,
 | |
|                                 StringRef outTypeStr, StringRef inTypeStr,
 | |
|                                 OpKind kind, ClassKind classKind) {
 | |
|   assert(!proto.empty() && "");
 | |
|   bool define = UseMacro(proto) && kind != OpUnavailable;
 | |
|   std::string s;
 | |
| 
 | |
|   // static always inline + return type
 | |
|   if (define)
 | |
|     s += "#define ";
 | |
|   else
 | |
|     s += "__ai " + TypeString(proto[0], outTypeStr) + " ";
 | |
| 
 | |
|   // Function name with type suffix
 | |
|   std::string mangledName = MangleName(name, outTypeStr, ClassS);
 | |
|   if (outTypeStr != inTypeStr) {
 | |
|     // If the input type is different (e.g., for vreinterpret), append a suffix
 | |
|     // for the input type.  String off a "Q" (quad) prefix so that MangleName
 | |
|     // does not insert another "q" in the name.
 | |
|     unsigned typeStrOff = (inTypeStr[0] == 'Q' ? 1 : 0);
 | |
|     StringRef inTypeNoQuad = inTypeStr.substr(typeStrOff);
 | |
|     mangledName = MangleName(mangledName, inTypeNoQuad, ClassS);
 | |
|   }
 | |
|   s += mangledName;
 | |
| 
 | |
|   // Function arguments
 | |
|   s += GenArgs(proto, inTypeStr);
 | |
| 
 | |
|   // Definition.
 | |
|   if (define) {
 | |
|     s += " __extension__ ({ \\\n  ";
 | |
|     s += GenMacroLocals(proto, inTypeStr);
 | |
|   } else if (kind == OpUnavailable) {
 | |
|     s += " __attribute__((unavailable));\n";
 | |
|     return s;
 | |
|   } else
 | |
|     s += " {\n  ";
 | |
| 
 | |
|   if (kind != OpNone)
 | |
|     s += GenOpString(kind, proto, outTypeStr);
 | |
|   else
 | |
|     s += GenBuiltin(name, proto, outTypeStr, classKind);
 | |
|   if (define)
 | |
|     s += " })";
 | |
|   else
 | |
|     s += " }";
 | |
|   s += "\n";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| /// run - Read the records in arm_neon.td and output arm_neon.h.  arm_neon.h
 | |
| /// is comprised of type definitions and function declarations.
 | |
| void NeonEmitter::run(raw_ostream &OS) {
 | |
|   OS << 
 | |
|     "/*===---- arm_neon.h - ARM Neon intrinsics ------------------------------"
 | |
|     "---===\n"
 | |
|     " *\n"
 | |
|     " * Permission is hereby granted, free of charge, to any person obtaining "
 | |
|     "a copy\n"
 | |
|     " * of this software and associated documentation files (the \"Software\"),"
 | |
|     " to deal\n"
 | |
|     " * in the Software without restriction, including without limitation the "
 | |
|     "rights\n"
 | |
|     " * to use, copy, modify, merge, publish, distribute, sublicense, "
 | |
|     "and/or sell\n"
 | |
|     " * copies of the Software, and to permit persons to whom the Software is\n"
 | |
|     " * furnished to do so, subject to the following conditions:\n"
 | |
|     " *\n"
 | |
|     " * The above copyright notice and this permission notice shall be "
 | |
|     "included in\n"
 | |
|     " * all copies or substantial portions of the Software.\n"
 | |
|     " *\n"
 | |
|     " * THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, "
 | |
|     "EXPRESS OR\n"
 | |
|     " * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF "
 | |
|     "MERCHANTABILITY,\n"
 | |
|     " * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT "
 | |
|     "SHALL THE\n"
 | |
|     " * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR "
 | |
|     "OTHER\n"
 | |
|     " * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, "
 | |
|     "ARISING FROM,\n"
 | |
|     " * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER "
 | |
|     "DEALINGS IN\n"
 | |
|     " * THE SOFTWARE.\n"
 | |
|     " *\n"
 | |
|     " *===--------------------------------------------------------------------"
 | |
|     "---===\n"
 | |
|     " */\n\n";
 | |
| 
 | |
|   OS << "#ifndef __ARM_NEON_H\n";
 | |
|   OS << "#define __ARM_NEON_H\n\n";
 | |
| 
 | |
|   OS << "#if !defined(__ARM_NEON__) && !defined(__AARCH_FEATURE_ADVSIMD)\n";
 | |
|   OS << "#error \"NEON support not enabled\"\n";
 | |
|   OS << "#endif\n\n";
 | |
| 
 | |
|   OS << "#include <stdint.h>\n\n";
 | |
| 
 | |
|   // Emit NEON-specific scalar typedefs.
 | |
|   OS << "typedef float float32_t;\n";
 | |
|   OS << "typedef __fp16 float16_t;\n";
 | |
| 
 | |
|   OS << "#ifdef __aarch64__\n";
 | |
|   OS << "typedef double float64_t;\n";
 | |
|   OS << "#endif\n\n";
 | |
| 
 | |
|   // For now, signedness of polynomial types depends on target
 | |
|   OS << "#ifdef __aarch64__\n";
 | |
|   OS << "typedef uint8_t poly8_t;\n";
 | |
|   OS << "typedef uint16_t poly16_t;\n";
 | |
|   OS << "#else\n";
 | |
|   OS << "typedef int8_t poly8_t;\n";
 | |
|   OS << "typedef int16_t poly16_t;\n";
 | |
|   OS << "#endif\n";
 | |
| 
 | |
|   // Emit Neon vector typedefs.
 | |
|   std::string TypedefTypes(
 | |
|       "cQcsQsiQilQlUcQUcUsQUsUiQUiUlQUlhQhfQfQdPcQPcPsQPs");
 | |
|   SmallVector<StringRef, 24> TDTypeVec;
 | |
|   ParseTypes(0, TypedefTypes, TDTypeVec);
 | |
| 
 | |
|   // Emit vector typedefs.
 | |
|   for (unsigned i = 0, e = TDTypeVec.size(); i != e; ++i) {
 | |
|     bool dummy, quad = false, poly = false;
 | |
|     char type = ClassifyType(TDTypeVec[i], quad, poly, dummy);
 | |
|     bool isA64 = false;
 | |
| 
 | |
|     if (type == 'd' && quad)
 | |
|       isA64 = true;
 | |
| 
 | |
|     if (isA64)
 | |
|       OS << "#ifdef __aarch64__\n";
 | |
| 
 | |
|     if (poly)
 | |
|       OS << "typedef __attribute__((neon_polyvector_type(";
 | |
|     else
 | |
|       OS << "typedef __attribute__((neon_vector_type(";
 | |
| 
 | |
|     unsigned nElts = GetNumElements(TDTypeVec[i], quad);
 | |
|     OS << utostr(nElts) << "))) ";
 | |
|     if (nElts < 10)
 | |
|       OS << " ";
 | |
| 
 | |
|     OS << TypeString('s', TDTypeVec[i]);
 | |
|     OS << " " << TypeString('d', TDTypeVec[i]) << ";\n";
 | |
| 
 | |
|     if (isA64)
 | |
|       OS << "#endif\n";
 | |
|   }
 | |
|   OS << "\n";
 | |
| 
 | |
|   // Emit struct typedefs.
 | |
|   for (unsigned vi = 2; vi != 5; ++vi) {
 | |
|     for (unsigned i = 0, e = TDTypeVec.size(); i != e; ++i) {
 | |
|       bool dummy, quad = false, poly = false;
 | |
|       char type = ClassifyType(TDTypeVec[i], quad, poly, dummy);
 | |
|       bool isA64 = false;
 | |
| 
 | |
|       if (type == 'd' && quad)
 | |
|         isA64 = true;
 | |
| 
 | |
|       if (isA64)
 | |
|         OS << "#ifdef __aarch64__\n";
 | |
| 
 | |
|       std::string ts = TypeString('d', TDTypeVec[i]);
 | |
|       std::string vs = TypeString('0' + vi, TDTypeVec[i]);
 | |
|       OS << "typedef struct " << vs << " {\n";
 | |
|       OS << "  " << ts << " val";
 | |
|       OS << "[" << utostr(vi) << "]";
 | |
|       OS << ";\n} ";
 | |
|       OS << vs << ";\n";
 | |
| 
 | |
|       if (isA64)
 | |
|         OS << "#endif\n";
 | |
| 
 | |
|       OS << "\n";
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   OS<<"#define __ai static inline __attribute__((__always_inline__, __nodebug__))\n\n";
 | |
| 
 | |
|   std::vector<Record*> RV = Records.getAllDerivedDefinitions("Inst");
 | |
| 
 | |
|   StringMap<ClassKind> EmittedMap;
 | |
| 
 | |
|   // Emit vmovl, vmull and vabd intrinsics first so they can be used by other
 | |
|   // intrinsics.  (Some of the saturating multiply instructions are also
 | |
|   // used to implement the corresponding "_lane" variants, but tablegen
 | |
|   // sorts the records into alphabetical order so that the "_lane" variants
 | |
|   // come after the intrinsics they use.)
 | |
|   emitIntrinsic(OS, Records.getDef("VMOVL"), EmittedMap);
 | |
|   emitIntrinsic(OS, Records.getDef("VMULL"), EmittedMap);
 | |
|   emitIntrinsic(OS, Records.getDef("VABD"), EmittedMap);
 | |
| 
 | |
|   // ARM intrinsics must be emitted before AArch64 intrinsics to ensure
 | |
|   // common intrinsics appear only once in the output stream.
 | |
|   // The check for uniquiness is done in emitIntrinsic.
 | |
|   // Emit ARM intrinsics.
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
| 
 | |
|     // Skip AArch64 intrinsics; they will be emitted at the end.
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (isA64)
 | |
|       continue;
 | |
| 
 | |
|     if (R->getName() != "VMOVL" && R->getName() != "VMULL" &&
 | |
|         R->getName() != "VABD")
 | |
|       emitIntrinsic(OS, R, EmittedMap);
 | |
|   }
 | |
| 
 | |
|   // Emit AArch64-specific intrinsics.
 | |
|   OS << "#ifdef __aarch64__\n";
 | |
| 
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
| 
 | |
|     // Skip ARM intrinsics already included above.
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (!isA64)
 | |
|       continue;
 | |
| 
 | |
|     emitIntrinsic(OS, R, EmittedMap);
 | |
|   }
 | |
| 
 | |
|   OS << "#endif\n\n";
 | |
| 
 | |
|   OS << "#undef __ai\n\n";
 | |
|   OS << "#endif /* __ARM_NEON_H */\n";
 | |
| }
 | |
| 
 | |
| /// emitIntrinsic - Write out the arm_neon.h header file definitions for the
 | |
| /// intrinsics specified by record R checking for intrinsic uniqueness.
 | |
| void NeonEmitter::emitIntrinsic(raw_ostream &OS, Record *R,
 | |
|                                 StringMap<ClassKind> &EmittedMap) {
 | |
|   std::string name = R->getValueAsString("Name");
 | |
|   std::string Proto = R->getValueAsString("Prototype");
 | |
|   std::string Types = R->getValueAsString("Types");
 | |
| 
 | |
|   SmallVector<StringRef, 16> TypeVec;
 | |
|   ParseTypes(R, Types, TypeVec);
 | |
| 
 | |
|   OpKind kind = OpMap[R->getValueAsDef("Operand")->getName()];
 | |
| 
 | |
|   ClassKind classKind = ClassNone;
 | |
|   if (R->getSuperClasses().size() >= 2)
 | |
|     classKind = ClassMap[R->getSuperClasses()[1]];
 | |
|   if (classKind == ClassNone && kind == OpNone)
 | |
|     PrintFatalError(R->getLoc(), "Builtin has no class kind");
 | |
| 
 | |
|   for (unsigned ti = 0, te = TypeVec.size(); ti != te; ++ti) {
 | |
|     if (kind == OpReinterpret) {
 | |
|       bool outQuad = false;
 | |
|       bool dummy = false;
 | |
|       (void)ClassifyType(TypeVec[ti], outQuad, dummy, dummy);
 | |
|       for (unsigned srcti = 0, srcte = TypeVec.size();
 | |
|            srcti != srcte; ++srcti) {
 | |
|         bool inQuad = false;
 | |
|         (void)ClassifyType(TypeVec[srcti], inQuad, dummy, dummy);
 | |
|         if (srcti == ti || inQuad != outQuad)
 | |
|           continue;
 | |
|         std::string s = GenIntrinsic(name, Proto, TypeVec[ti], TypeVec[srcti],
 | |
|                                      OpCast, ClassS);
 | |
|         if (EmittedMap.count(s))
 | |
|           continue;
 | |
|         EmittedMap[s] = ClassS;
 | |
|         OS << s;
 | |
|       }
 | |
|     } else {
 | |
|       std::string s =
 | |
|           GenIntrinsic(name, Proto, TypeVec[ti], TypeVec[ti], kind, classKind);
 | |
|       if (EmittedMap.count(s))
 | |
|         continue;
 | |
|       EmittedMap[s] = classKind;
 | |
|       OS << s;
 | |
|     }
 | |
|   }
 | |
|   OS << "\n";
 | |
| }
 | |
| 
 | |
| static unsigned RangeFromType(const char mod, StringRef typestr) {
 | |
|   // base type to get the type string for.
 | |
|   bool quad = false, dummy = false;
 | |
|   char type = ClassifyType(typestr, quad, dummy, dummy);
 | |
|   type = ModType(mod, type, quad, dummy, dummy, dummy, dummy, dummy);
 | |
| 
 | |
|   switch (type) {
 | |
|     case 'c':
 | |
|       return (8 << (int)quad) - 1;
 | |
|     case 'h':
 | |
|     case 's':
 | |
|       return (4 << (int)quad) - 1;
 | |
|     case 'f':
 | |
|     case 'i':
 | |
|       return (2 << (int)quad) - 1;
 | |
|     case 'l':
 | |
|       return (1 << (int)quad) - 1;
 | |
|     default:
 | |
|       PrintFatalError("unhandled type!");
 | |
|   }
 | |
| }
 | |
| 
 | |
| /// Generate the ARM and AArch64 intrinsic range checking code for
 | |
| /// shift/lane immediates, checking for unique declarations.
 | |
| void
 | |
| NeonEmitter::genIntrinsicRangeCheckCode(raw_ostream &OS,
 | |
|                                         StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                                         bool isA64RangeCheck) {
 | |
|   std::vector<Record *> RV = Records.getAllDerivedDefinitions("Inst");
 | |
|   StringMap<OpKind> EmittedMap;
 | |
| 
 | |
|   // Generate the intrinsic range checking code for shift/lane immediates.
 | |
|   if (isA64RangeCheck)
 | |
|     OS << "#ifdef GET_NEON_AARCH64_IMMEDIATE_CHECK\n";
 | |
|   else
 | |
|     OS << "#ifdef GET_NEON_IMMEDIATE_CHECK\n";
 | |
| 
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
| 
 | |
|     OpKind k = OpMap[R->getValueAsDef("Operand")->getName()];
 | |
|     if (k != OpNone)
 | |
|       continue;
 | |
| 
 | |
|     std::string name = R->getValueAsString("Name");
 | |
|     std::string Proto = R->getValueAsString("Prototype");
 | |
|     std::string Types = R->getValueAsString("Types");
 | |
| 
 | |
|     // Functions with 'a' (the splat code) in the type prototype should not get
 | |
|     // their own builtin as they use the non-splat variant.
 | |
|     if (Proto.find('a') != std::string::npos)
 | |
|       continue;
 | |
| 
 | |
|     // Functions which do not have an immediate do not need to have range
 | |
|     // checking code emitted.
 | |
|     size_t immPos = Proto.find('i');
 | |
|     if (immPos == std::string::npos)
 | |
|       continue;
 | |
| 
 | |
|     SmallVector<StringRef, 16> TypeVec;
 | |
|     ParseTypes(R, Types, TypeVec);
 | |
| 
 | |
|     if (R->getSuperClasses().size() < 2)
 | |
|       PrintFatalError(R->getLoc(), "Builtin has no class kind");
 | |
| 
 | |
|     ClassKind ck = ClassMap[R->getSuperClasses()[1]];
 | |
| 
 | |
|     // Do not include AArch64 range checks if not generating code for AArch64.
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (!isA64RangeCheck && isA64)
 | |
|       continue;
 | |
| 
 | |
|     // Include ARM range checks in AArch64 but only if ARM intrinsics are not
 | |
|     // redefined by AArch64 to handle new types.
 | |
|     if (isA64RangeCheck && !isA64 && A64IntrinsicMap.count(name)) {
 | |
|       ClassKind &A64CK = A64IntrinsicMap[name];
 | |
|       if (A64CK == ck && ck != ClassNone)
 | |
|         continue;
 | |
|     }
 | |
| 
 | |
|     for (unsigned ti = 0, te = TypeVec.size(); ti != te; ++ti) {
 | |
|       std::string namestr, shiftstr, rangestr;
 | |
| 
 | |
|       if (R->getValueAsBit("isVCVT_N")) {
 | |
|         // VCVT between floating- and fixed-point values takes an immediate
 | |
|         // in the range 1 to 32.
 | |
|         ck = ClassB;
 | |
|         rangestr = "l = 1; u = 31"; // upper bound = l + u
 | |
|       } else if (Proto.find('s') == std::string::npos) {
 | |
|         // Builtins which are overloaded by type will need to have their upper
 | |
|         // bound computed at Sema time based on the type constant.
 | |
|         ck = ClassB;
 | |
|         if (R->getValueAsBit("isShift")) {
 | |
|           shiftstr = ", true";
 | |
| 
 | |
|           // Right shifts have an 'r' in the name, left shifts do not.
 | |
|           if (name.find('r') != std::string::npos)
 | |
|             rangestr = "l = 1; ";
 | |
|         }
 | |
|         rangestr += "u = RFT(TV" + shiftstr + ")";
 | |
|       } else {
 | |
|         // The immediate generally refers to a lane in the preceding argument.
 | |
|         assert(immPos > 0 && "unexpected immediate operand");
 | |
|         rangestr =
 | |
|             "u = " + utostr(RangeFromType(Proto[immPos - 1], TypeVec[ti]));
 | |
|       }
 | |
|       // Make sure cases appear only once by uniquing them in a string map.
 | |
|       namestr = MangleName(name, TypeVec[ti], ck);
 | |
|       if (EmittedMap.count(namestr))
 | |
|         continue;
 | |
|       EmittedMap[namestr] = OpNone;
 | |
| 
 | |
|       // Calculate the index of the immediate that should be range checked.
 | |
|       unsigned immidx = 0;
 | |
| 
 | |
|       // Builtins that return a struct of multiple vectors have an extra
 | |
|       // leading arg for the struct return.
 | |
|       if (Proto[0] >= '2' && Proto[0] <= '4')
 | |
|         ++immidx;
 | |
| 
 | |
|       // Add one to the index for each argument until we reach the immediate
 | |
|       // to be checked.  Structs of vectors are passed as multiple arguments.
 | |
|       for (unsigned ii = 1, ie = Proto.size(); ii != ie; ++ii) {
 | |
|         switch (Proto[ii]) {
 | |
|         default:
 | |
|           immidx += 1;
 | |
|           break;
 | |
|         case '2':
 | |
|           immidx += 2;
 | |
|           break;
 | |
|         case '3':
 | |
|           immidx += 3;
 | |
|           break;
 | |
|         case '4':
 | |
|           immidx += 4;
 | |
|           break;
 | |
|         case 'i':
 | |
|           ie = ii + 1;
 | |
|           break;
 | |
|         }
 | |
|       }
 | |
|       if (isA64RangeCheck)
 | |
|         OS << "case AArch64::BI__builtin_neon_";
 | |
|       else
 | |
|         OS << "case ARM::BI__builtin_neon_";
 | |
|       OS << MangleName(name, TypeVec[ti], ck) << ": i = " << immidx << "; "
 | |
|          << rangestr << "; break;\n";
 | |
|     }
 | |
|   }
 | |
|   OS << "#endif\n\n";
 | |
| }
 | |
| 
 | |
| /// Generate the ARM and AArch64 overloaded type checking code for
 | |
| /// SemaChecking.cpp, checking for unique builtin declarations.
 | |
| void
 | |
| NeonEmitter::genOverloadTypeCheckCode(raw_ostream &OS,
 | |
|                                       StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                                       bool isA64TypeCheck) {
 | |
|   std::vector<Record *> RV = Records.getAllDerivedDefinitions("Inst");
 | |
|   StringMap<OpKind> EmittedMap;
 | |
| 
 | |
|   // Generate the overloaded type checking code for SemaChecking.cpp
 | |
|   if (isA64TypeCheck)
 | |
|     OS << "#ifdef GET_NEON_AARCH64_OVERLOAD_CHECK\n";
 | |
|   else
 | |
|     OS << "#ifdef GET_NEON_OVERLOAD_CHECK\n";
 | |
| 
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
|     OpKind k = OpMap[R->getValueAsDef("Operand")->getName()];
 | |
|     if (k != OpNone)
 | |
|       continue;
 | |
| 
 | |
|     std::string Proto = R->getValueAsString("Prototype");
 | |
|     std::string Types = R->getValueAsString("Types");
 | |
|     std::string name = R->getValueAsString("Name");
 | |
| 
 | |
|     // Functions with 'a' (the splat code) in the type prototype should not get
 | |
|     // their own builtin as they use the non-splat variant.
 | |
|     if (Proto.find('a') != std::string::npos)
 | |
|       continue;
 | |
| 
 | |
|     // Functions which have a scalar argument cannot be overloaded, no need to
 | |
|     // check them if we are emitting the type checking code.
 | |
|     if (Proto.find('s') != std::string::npos)
 | |
|       continue;
 | |
| 
 | |
|     SmallVector<StringRef, 16> TypeVec;
 | |
|     ParseTypes(R, Types, TypeVec);
 | |
| 
 | |
|     if (R->getSuperClasses().size() < 2)
 | |
|       PrintFatalError(R->getLoc(), "Builtin has no class kind");
 | |
| 
 | |
|     // Do not include AArch64 type checks if not generating code for AArch64.
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (!isA64TypeCheck && isA64)
 | |
|       continue;
 | |
| 
 | |
|     // Include ARM  type check in AArch64 but only if ARM intrinsics
 | |
|     // are not redefined in AArch64 to handle new types, e.g. "vabd" is a SIntr
 | |
|     // redefined in AArch64 to handle an additional 2 x f64 type.
 | |
|     ClassKind ck = ClassMap[R->getSuperClasses()[1]];
 | |
|     if (isA64TypeCheck && !isA64 && A64IntrinsicMap.count(name)) {
 | |
|       ClassKind &A64CK = A64IntrinsicMap[name];
 | |
|       if (A64CK == ck && ck != ClassNone)
 | |
|         continue;
 | |
|     }
 | |
| 
 | |
|     int si = -1, qi = -1;
 | |
|     uint64_t mask = 0, qmask = 0;
 | |
|     for (unsigned ti = 0, te = TypeVec.size(); ti != te; ++ti) {
 | |
|       // Generate the switch case(s) for this builtin for the type validation.
 | |
|       bool quad = false, poly = false, usgn = false;
 | |
|       (void) ClassifyType(TypeVec[ti], quad, poly, usgn);
 | |
| 
 | |
|       if (quad) {
 | |
|         qi = ti;
 | |
|         qmask |= 1ULL << GetNeonEnum(Proto, TypeVec[ti]);
 | |
|       } else {
 | |
|         si = ti;
 | |
|         mask |= 1ULL << GetNeonEnum(Proto, TypeVec[ti]);
 | |
|       }
 | |
|     }
 | |
| 
 | |
|     // Check if the builtin function has a pointer or const pointer argument.
 | |
|     int PtrArgNum = -1;
 | |
|     bool HasConstPtr = false;
 | |
|     for (unsigned arg = 1, arge = Proto.size(); arg != arge; ++arg) {
 | |
|       char ArgType = Proto[arg];
 | |
|       if (ArgType == 'c') {
 | |
|         HasConstPtr = true;
 | |
|         PtrArgNum = arg - 1;
 | |
|         break;
 | |
|       }
 | |
|       if (ArgType == 'p') {
 | |
|         PtrArgNum = arg - 1;
 | |
|         break;
 | |
|       }
 | |
|     }
 | |
|     // For sret builtins, adjust the pointer argument index.
 | |
|     if (PtrArgNum >= 0 && (Proto[0] >= '2' && Proto[0] <= '4'))
 | |
|       PtrArgNum += 1;
 | |
| 
 | |
|     // Omit type checking for the pointer arguments of vld1_lane, vld1_dup,
 | |
|     // and vst1_lane intrinsics.  Using a pointer to the vector element
 | |
|     // type with one of those operations causes codegen to select an aligned
 | |
|     // load/store instruction.  If you want an unaligned operation,
 | |
|     // the pointer argument needs to have less alignment than element type,
 | |
|     // so just accept any pointer type.
 | |
|     if (name == "vld1_lane" || name == "vld1_dup" || name == "vst1_lane") {
 | |
|       PtrArgNum = -1;
 | |
|       HasConstPtr = false;
 | |
|     }
 | |
| 
 | |
|     if (mask) {
 | |
|       if (isA64TypeCheck)
 | |
|         OS << "case AArch64::BI__builtin_neon_";
 | |
|       else
 | |
|         OS << "case ARM::BI__builtin_neon_";
 | |
|       OS << MangleName(name, TypeVec[si], ClassB) << ": mask = "
 | |
|          << "0x" << utohexstr(mask) << "ULL";
 | |
|       if (PtrArgNum >= 0)
 | |
|         OS << "; PtrArgNum = " << PtrArgNum;
 | |
|       if (HasConstPtr)
 | |
|         OS << "; HasConstPtr = true";
 | |
|       OS << "; break;\n";
 | |
|     }
 | |
|     if (qmask) {
 | |
|       if (isA64TypeCheck)
 | |
|         OS << "case AArch64::BI__builtin_neon_";
 | |
|       else
 | |
|         OS << "case ARM::BI__builtin_neon_";
 | |
|       OS << MangleName(name, TypeVec[qi], ClassB) << ": mask = "
 | |
|          << "0x" << utohexstr(qmask) << "ULL";
 | |
|       if (PtrArgNum >= 0)
 | |
|         OS << "; PtrArgNum = " << PtrArgNum;
 | |
|       if (HasConstPtr)
 | |
|         OS << "; HasConstPtr = true";
 | |
|       OS << "; break;\n";
 | |
|     }
 | |
|   }
 | |
|   OS << "#endif\n\n";
 | |
| }
 | |
| 
 | |
| /// genBuiltinsDef: Generate the BuiltinsARM.def and  BuiltinsAArch64.def
 | |
| /// declaration of builtins, checking for unique builtin declarations.
 | |
| void NeonEmitter::genBuiltinsDef(raw_ostream &OS,
 | |
|                                  StringMap<ClassKind> &A64IntrinsicMap,
 | |
|                                  bool isA64GenBuiltinDef) {
 | |
|   std::vector<Record *> RV = Records.getAllDerivedDefinitions("Inst");
 | |
|   StringMap<OpKind> EmittedMap;
 | |
| 
 | |
|   // Generate BuiltinsARM.def and BuiltinsAArch64.def
 | |
|   if (isA64GenBuiltinDef)
 | |
|     OS << "#ifdef GET_NEON_AARCH64_BUILTINS\n";
 | |
|   else
 | |
|     OS << "#ifdef GET_NEON_BUILTINS\n";
 | |
| 
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
|     OpKind k = OpMap[R->getValueAsDef("Operand")->getName()];
 | |
|     if (k != OpNone)
 | |
|       continue;
 | |
| 
 | |
|     std::string Proto = R->getValueAsString("Prototype");
 | |
|     std::string name = R->getValueAsString("Name");
 | |
| 
 | |
|     // Functions with 'a' (the splat code) in the type prototype should not get
 | |
|     // their own builtin as they use the non-splat variant.
 | |
|     if (Proto.find('a') != std::string::npos)
 | |
|       continue;
 | |
| 
 | |
|     std::string Types = R->getValueAsString("Types");
 | |
|     SmallVector<StringRef, 16> TypeVec;
 | |
|     ParseTypes(R, Types, TypeVec);
 | |
| 
 | |
|     if (R->getSuperClasses().size() < 2)
 | |
|       PrintFatalError(R->getLoc(), "Builtin has no class kind");
 | |
| 
 | |
|     ClassKind ck = ClassMap[R->getSuperClasses()[1]];
 | |
| 
 | |
|     // Do not include AArch64 BUILTIN() macros if not generating
 | |
|     // code for AArch64
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (!isA64GenBuiltinDef && isA64)
 | |
|       continue;
 | |
| 
 | |
|     // Include ARM  BUILTIN() macros  in AArch64 but only if ARM intrinsics
 | |
|     // are not redefined in AArch64 to handle new types, e.g. "vabd" is a SIntr
 | |
|     // redefined in AArch64 to handle an additional 2 x f64 type.
 | |
|     if (isA64GenBuiltinDef && !isA64 && A64IntrinsicMap.count(name)) {
 | |
|       ClassKind &A64CK = A64IntrinsicMap[name];
 | |
|       if (A64CK == ck && ck != ClassNone)
 | |
|         continue;
 | |
|     }
 | |
| 
 | |
|     for (unsigned ti = 0, te = TypeVec.size(); ti != te; ++ti) {
 | |
|       // Generate the declaration for this builtin, ensuring
 | |
|       // that each unique BUILTIN() macro appears only once in the output
 | |
|       // stream.
 | |
|       std::string bd = GenBuiltinDef(name, Proto, TypeVec[ti], ck);
 | |
|       if (EmittedMap.count(bd))
 | |
|         continue;
 | |
| 
 | |
|       EmittedMap[bd] = OpNone;
 | |
|       OS << bd << "\n";
 | |
|     }
 | |
|   }
 | |
|   OS << "#endif\n\n";
 | |
| }
 | |
| 
 | |
| /// runHeader - Emit a file with sections defining:
 | |
| /// 1. the NEON section of BuiltinsARM.def and BuiltinsAArch64.def.
 | |
| /// 2. the SemaChecking code for the type overload checking.
 | |
| /// 3. the SemaChecking code for validation of intrinsic immediate arguments.
 | |
| void NeonEmitter::runHeader(raw_ostream &OS) {
 | |
|   std::vector<Record *> RV = Records.getAllDerivedDefinitions("Inst");
 | |
| 
 | |
|   // build a map of AArch64 intriniscs to be used in uniqueness checks.
 | |
|   StringMap<ClassKind> A64IntrinsicMap;
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
| 
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
|     if (!isA64)
 | |
|       continue;
 | |
| 
 | |
|     ClassKind CK = ClassNone;
 | |
|     if (R->getSuperClasses().size() >= 2)
 | |
|       CK = ClassMap[R->getSuperClasses()[1]];
 | |
| 
 | |
|     std::string Name = R->getValueAsString("Name");
 | |
|     if (A64IntrinsicMap.count(Name))
 | |
|       continue;
 | |
|     A64IntrinsicMap[Name] = CK;
 | |
|   }
 | |
| 
 | |
|   // Generate BuiltinsARM.def for ARM
 | |
|   genBuiltinsDef(OS, A64IntrinsicMap, false);
 | |
| 
 | |
|   // Generate BuiltinsAArch64.def for AArch64
 | |
|   genBuiltinsDef(OS, A64IntrinsicMap, true);
 | |
| 
 | |
|   // Generate ARM overloaded type checking code for SemaChecking.cpp
 | |
|   genOverloadTypeCheckCode(OS, A64IntrinsicMap, false);
 | |
| 
 | |
|   // Generate AArch64 overloaded type checking code for SemaChecking.cpp
 | |
|   genOverloadTypeCheckCode(OS, A64IntrinsicMap, true);
 | |
| 
 | |
|   // Generate ARM range checking code for shift/lane immediates.
 | |
|   genIntrinsicRangeCheckCode(OS, A64IntrinsicMap, false);
 | |
| 
 | |
|   // Generate the AArch64 range checking code for shift/lane immediates.
 | |
|   genIntrinsicRangeCheckCode(OS, A64IntrinsicMap, true);
 | |
| }
 | |
| 
 | |
| /// GenTest - Write out a test for the intrinsic specified by the name and
 | |
| /// type strings, including the embedded patterns for FileCheck to match.
 | |
| static std::string GenTest(const std::string &name,
 | |
|                            const std::string &proto,
 | |
|                            StringRef outTypeStr, StringRef inTypeStr,
 | |
|                            bool isShift, bool isHiddenLOp,
 | |
|                            ClassKind ck, const std::string &InstName,
 | |
| 						   bool isA64,
 | |
| 						   std::string & testFuncProto) {
 | |
|   assert(!proto.empty() && "");
 | |
|   std::string s;
 | |
| 
 | |
|   // Function name with type suffix
 | |
|   std::string mangledName = MangleName(name, outTypeStr, ClassS);
 | |
|   if (outTypeStr != inTypeStr) {
 | |
|     // If the input type is different (e.g., for vreinterpret), append a suffix
 | |
|     // for the input type.  String off a "Q" (quad) prefix so that MangleName
 | |
|     // does not insert another "q" in the name.
 | |
|     unsigned typeStrOff = (inTypeStr[0] == 'Q' ? 1 : 0);
 | |
|     StringRef inTypeNoQuad = inTypeStr.substr(typeStrOff);
 | |
|     mangledName = MangleName(mangledName, inTypeNoQuad, ClassS);
 | |
|   }
 | |
| 
 | |
|   // todo: GenerateChecksForIntrinsic does not generate CHECK
 | |
|   // for aarch64 instructions yet
 | |
|   std::vector<std::string> FileCheckPatterns;
 | |
|   if (!isA64) {
 | |
| 	GenerateChecksForIntrinsic(name, proto, outTypeStr, inTypeStr, ck, InstName,
 | |
| 							   isHiddenLOp, FileCheckPatterns);
 | |
| 	s+= "// CHECK_ARM: test_" + mangledName + "\n";
 | |
|   }
 | |
|   s += "// CHECK_AARCH64: test_" + mangledName + "\n";
 | |
| 
 | |
|   // Emit the FileCheck patterns.
 | |
|   // If for any reason we do not want to emit a check, mangledInst
 | |
|   // will be the empty string.
 | |
|   if (FileCheckPatterns.size()) {
 | |
|     for (std::vector<std::string>::const_iterator i = FileCheckPatterns.begin(),
 | |
|                                                   e = FileCheckPatterns.end();
 | |
|          i != e;
 | |
|          ++i) {
 | |
|       s += "// CHECK_ARM: " + *i + "\n";
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   // Emit the start of the test function.
 | |
| 
 | |
|   testFuncProto = TypeString(proto[0], outTypeStr) + " test_" + mangledName + "(";
 | |
|   char arg = 'a';
 | |
|   std::string comma;
 | |
|   for (unsigned i = 1, e = proto.size(); i != e; ++i, ++arg) {
 | |
|     // Do not create arguments for values that must be immediate constants.
 | |
|     if (proto[i] == 'i')
 | |
|       continue;
 | |
|     testFuncProto += comma + TypeString(proto[i], inTypeStr) + " ";
 | |
|     testFuncProto.push_back(arg);
 | |
|     comma = ", ";
 | |
|   }
 | |
|   testFuncProto += ")";
 | |
| 
 | |
|   s+= testFuncProto;
 | |
|   s+= " {\n  ";
 | |
| 
 | |
|   if (proto[0] != 'v')
 | |
|     s += "return ";
 | |
|   s += mangledName + "(";
 | |
|   arg = 'a';
 | |
|   for (unsigned i = 1, e = proto.size(); i != e; ++i, ++arg) {
 | |
|     if (proto[i] == 'i') {
 | |
|       // For immediate operands, test the maximum value.
 | |
|       if (isShift)
 | |
|         s += "1"; // FIXME
 | |
|       else
 | |
|         // The immediate generally refers to a lane in the preceding argument.
 | |
|         s += utostr(RangeFromType(proto[i-1], inTypeStr));
 | |
|     } else {
 | |
|       s.push_back(arg);
 | |
|     }
 | |
|     if ((i + 1) < e)
 | |
|       s += ", ";
 | |
|   }
 | |
|   s += ");\n}\n\n";
 | |
|   return s;
 | |
| }
 | |
| 
 | |
| /// Write out all intrinsic tests for the specified target, checking
 | |
| /// for intrinsic test uniqueness.
 | |
| void NeonEmitter::genTargetTest(raw_ostream &OS, StringMap<OpKind> &EmittedMap,
 | |
|                                 bool isA64GenTest) {
 | |
|   if (isA64GenTest)
 | |
| 	OS << "#ifdef __aarch64__\n";
 | |
| 
 | |
|   std::vector<Record *> RV = Records.getAllDerivedDefinitions("Inst");
 | |
|   for (unsigned i = 0, e = RV.size(); i != e; ++i) {
 | |
|     Record *R = RV[i];
 | |
|     std::string name = R->getValueAsString("Name");
 | |
|     std::string Proto = R->getValueAsString("Prototype");
 | |
|     std::string Types = R->getValueAsString("Types");
 | |
|     bool isShift = R->getValueAsBit("isShift");
 | |
|     std::string InstName = R->getValueAsString("InstName");
 | |
|     bool isHiddenLOp = R->getValueAsBit("isHiddenLInst");
 | |
|     bool isA64 = R->getValueAsBit("isA64");
 | |
| 
 | |
|     // do not include AArch64 intrinsic test if not generating
 | |
|     // code for AArch64
 | |
|     if (!isA64GenTest && isA64)
 | |
|       continue;
 | |
| 
 | |
|     SmallVector<StringRef, 16> TypeVec;
 | |
|     ParseTypes(R, Types, TypeVec);
 | |
| 
 | |
|     ClassKind ck = ClassMap[R->getSuperClasses()[1]];
 | |
|     OpKind kind = OpMap[R->getValueAsDef("Operand")->getName()];
 | |
|     if (kind == OpUnavailable)
 | |
|       continue;
 | |
|     for (unsigned ti = 0, te = TypeVec.size(); ti != te; ++ti) {
 | |
|       if (kind == OpReinterpret) {
 | |
|         bool outQuad = false;
 | |
|         bool dummy = false;
 | |
|         (void)ClassifyType(TypeVec[ti], outQuad, dummy, dummy);
 | |
|         for (unsigned srcti = 0, srcte = TypeVec.size();
 | |
|              srcti != srcte; ++srcti) {
 | |
|           bool inQuad = false;
 | |
|           (void)ClassifyType(TypeVec[srcti], inQuad, dummy, dummy);
 | |
|           if (srcti == ti || inQuad != outQuad)
 | |
|             continue;
 | |
| 		  std::string testFuncProto;
 | |
|           std::string s = GenTest(name, Proto, TypeVec[ti], TypeVec[srcti],
 | |
|                                   isShift, isHiddenLOp, ck, InstName, isA64,
 | |
| 								  testFuncProto);
 | |
|           if (EmittedMap.count(testFuncProto))
 | |
|             continue;
 | |
|           EmittedMap[testFuncProto] = kind;
 | |
|           OS << s << "\n";
 | |
|         }
 | |
|       } else {
 | |
| 		std::string testFuncProto;
 | |
|         std::string s = GenTest(name, Proto, TypeVec[ti], TypeVec[ti], isShift,
 | |
|                                 isHiddenLOp, ck, InstName, isA64, testFuncProto);
 | |
|         if (EmittedMap.count(testFuncProto))
 | |
|           continue;
 | |
|         EmittedMap[testFuncProto] = kind;
 | |
|         OS << s << "\n";
 | |
|       }
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (isA64GenTest)
 | |
| 	OS << "#endif\n";
 | |
| }
 | |
| /// runTests - Write out a complete set of tests for all of the Neon
 | |
| /// intrinsics.
 | |
| void NeonEmitter::runTests(raw_ostream &OS) {
 | |
|   OS << "// RUN: %clang_cc1 -triple thumbv7s-apple-darwin -target-abi "
 | |
|         "apcs-gnu\\\n"
 | |
|         "// RUN:  -target-cpu swift -ffreestanding -Os -S -o - %s\\\n"
 | |
|         "// RUN:  | FileCheck %s -check-prefix=CHECK_ARM\n"
 | |
| 		"\n"
 | |
| 	    "// RUN: %clang_cc1 -triple aarch64-none-linux-gnu \\\n"
 | |
| 	    "// RUN -target-feature +neon  -ffreestanding -S -o - %s \\\n"
 | |
| 	    "// RUN:  | FileCheck %s -check-prefix=CHECK_AARCH64\n"
 | |
|         "\n"
 | |
|         "// REQUIRES: long_tests\n"
 | |
|         "\n"
 | |
|         "#include <arm_neon.h>\n"
 | |
|         "\n";
 | |
| 
 | |
|   // ARM tests must be emitted before AArch64 tests to ensure
 | |
|   // tests for intrinsics that are common to ARM and AArch64
 | |
|   // appear only once in the output stream.
 | |
|   // The check for uniqueness is done in genTargetTest.
 | |
|   StringMap<OpKind> EmittedMap;
 | |
| 
 | |
|   genTargetTest(OS, EmittedMap, false);
 | |
| 
 | |
|   genTargetTest(OS, EmittedMap, true);
 | |
| }
 | |
| 
 | |
| namespace clang {
 | |
| void EmitNeon(RecordKeeper &Records, raw_ostream &OS) {
 | |
|   NeonEmitter(Records).run(OS);
 | |
| }
 | |
| void EmitNeonSema(RecordKeeper &Records, raw_ostream &OS) {
 | |
|   NeonEmitter(Records).runHeader(OS);
 | |
| }
 | |
| void EmitNeonTest(RecordKeeper &Records, raw_ostream &OS) {
 | |
|   NeonEmitter(Records).runTests(OS);
 | |
| }
 | |
| } // End namespace clang
 |