123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605 |
- //===- LowerMemIntrinsics.cpp ----------------------------------*- C++ -*--===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- #include "llvm/Transforms/Utils/LowerMemIntrinsics.h"
- #include "llvm/Analysis/ScalarEvolution.h"
- #include "llvm/Analysis/TargetTransformInfo.h"
- #include "llvm/IR/IRBuilder.h"
- #include "llvm/IR/IntrinsicInst.h"
- #include "llvm/IR/MDBuilder.h"
- #include "llvm/Transforms/Utils/BasicBlockUtils.h"
- #include <optional>
- using namespace llvm;
- void llvm::createMemCpyLoopKnownSize(
- Instruction *InsertBefore, Value *SrcAddr, Value *DstAddr,
- ConstantInt *CopyLen, Align SrcAlign, Align DstAlign, bool SrcIsVolatile,
- bool DstIsVolatile, bool CanOverlap, const TargetTransformInfo &TTI,
- std::optional<uint32_t> AtomicElementSize) {
- // No need to expand zero length copies.
- if (CopyLen->isZero())
- return;
- BasicBlock *PreLoopBB = InsertBefore->getParent();
- BasicBlock *PostLoopBB = nullptr;
- Function *ParentFunc = PreLoopBB->getParent();
- LLVMContext &Ctx = PreLoopBB->getContext();
- const DataLayout &DL = ParentFunc->getParent()->getDataLayout();
- MDBuilder MDB(Ctx);
- MDNode *NewDomain = MDB.createAnonymousAliasScopeDomain("MemCopyDomain");
- StringRef Name = "MemCopyAliasScope";
- MDNode *NewScope = MDB.createAnonymousAliasScope(NewDomain, Name);
- unsigned SrcAS = cast<PointerType>(SrcAddr->getType())->getAddressSpace();
- unsigned DstAS = cast<PointerType>(DstAddr->getType())->getAddressSpace();
- Type *TypeOfCopyLen = CopyLen->getType();
- Type *LoopOpType = TTI.getMemcpyLoopLoweringType(
- Ctx, CopyLen, SrcAS, DstAS, SrcAlign.value(), DstAlign.value(),
- AtomicElementSize);
- assert((!AtomicElementSize || !LoopOpType->isVectorTy()) &&
- "Atomic memcpy lowering is not supported for vector operand type");
- unsigned LoopOpSize = DL.getTypeStoreSize(LoopOpType);
- assert((!AtomicElementSize || LoopOpSize % *AtomicElementSize == 0) &&
- "Atomic memcpy lowering is not supported for selected operand size");
- uint64_t LoopEndCount = CopyLen->getZExtValue() / LoopOpSize;
- if (LoopEndCount != 0) {
- // Split
- PostLoopBB = PreLoopBB->splitBasicBlock(InsertBefore, "memcpy-split");
- BasicBlock *LoopBB =
- BasicBlock::Create(Ctx, "load-store-loop", ParentFunc, PostLoopBB);
- PreLoopBB->getTerminator()->setSuccessor(0, LoopBB);
- IRBuilder<> PLBuilder(PreLoopBB->getTerminator());
- // Cast the Src and Dst pointers to pointers to the loop operand type (if
- // needed).
- PointerType *SrcOpType = PointerType::get(LoopOpType, SrcAS);
- PointerType *DstOpType = PointerType::get(LoopOpType, DstAS);
- if (SrcAddr->getType() != SrcOpType) {
- SrcAddr = PLBuilder.CreateBitCast(SrcAddr, SrcOpType);
- }
- if (DstAddr->getType() != DstOpType) {
- DstAddr = PLBuilder.CreateBitCast(DstAddr, DstOpType);
- }
- Align PartDstAlign(commonAlignment(DstAlign, LoopOpSize));
- Align PartSrcAlign(commonAlignment(SrcAlign, LoopOpSize));
- IRBuilder<> LoopBuilder(LoopBB);
- PHINode *LoopIndex = LoopBuilder.CreatePHI(TypeOfCopyLen, 2, "loop-index");
- LoopIndex->addIncoming(ConstantInt::get(TypeOfCopyLen, 0U), PreLoopBB);
- // Loop Body
- Value *SrcGEP =
- LoopBuilder.CreateInBoundsGEP(LoopOpType, SrcAddr, LoopIndex);
- LoadInst *Load = LoopBuilder.CreateAlignedLoad(LoopOpType, SrcGEP,
- PartSrcAlign, SrcIsVolatile);
- if (!CanOverlap) {
- // Set alias scope for loads.
- Load->setMetadata(LLVMContext::MD_alias_scope,
- MDNode::get(Ctx, NewScope));
- }
- Value *DstGEP =
- LoopBuilder.CreateInBoundsGEP(LoopOpType, DstAddr, LoopIndex);
- StoreInst *Store = LoopBuilder.CreateAlignedStore(
- Load, DstGEP, PartDstAlign, DstIsVolatile);
- if (!CanOverlap) {
- // Indicate that stores don't overlap loads.
- Store->setMetadata(LLVMContext::MD_noalias, MDNode::get(Ctx, NewScope));
- }
- if (AtomicElementSize) {
- Load->setAtomic(AtomicOrdering::Unordered);
- Store->setAtomic(AtomicOrdering::Unordered);
- }
- Value *NewIndex =
- LoopBuilder.CreateAdd(LoopIndex, ConstantInt::get(TypeOfCopyLen, 1U));
- LoopIndex->addIncoming(NewIndex, LoopBB);
- // Create the loop branch condition.
- Constant *LoopEndCI = ConstantInt::get(TypeOfCopyLen, LoopEndCount);
- LoopBuilder.CreateCondBr(LoopBuilder.CreateICmpULT(NewIndex, LoopEndCI),
- LoopBB, PostLoopBB);
- }
- uint64_t BytesCopied = LoopEndCount * LoopOpSize;
- uint64_t RemainingBytes = CopyLen->getZExtValue() - BytesCopied;
- if (RemainingBytes) {
- IRBuilder<> RBuilder(PostLoopBB ? PostLoopBB->getFirstNonPHI()
- : InsertBefore);
- SmallVector<Type *, 5> RemainingOps;
- TTI.getMemcpyLoopResidualLoweringType(RemainingOps, Ctx, RemainingBytes,
- SrcAS, DstAS, SrcAlign.value(),
- DstAlign.value(), AtomicElementSize);
- for (auto *OpTy : RemainingOps) {
- Align PartSrcAlign(commonAlignment(SrcAlign, BytesCopied));
- Align PartDstAlign(commonAlignment(DstAlign, BytesCopied));
- // Calculate the new index
- unsigned OperandSize = DL.getTypeStoreSize(OpTy);
- assert(
- (!AtomicElementSize || OperandSize % *AtomicElementSize == 0) &&
- "Atomic memcpy lowering is not supported for selected operand size");
- uint64_t GepIndex = BytesCopied / OperandSize;
- assert(GepIndex * OperandSize == BytesCopied &&
- "Division should have no Remainder!");
- // Cast source to operand type and load
- PointerType *SrcPtrType = PointerType::get(OpTy, SrcAS);
- Value *CastedSrc = SrcAddr->getType() == SrcPtrType
- ? SrcAddr
- : RBuilder.CreateBitCast(SrcAddr, SrcPtrType);
- Value *SrcGEP = RBuilder.CreateInBoundsGEP(
- OpTy, CastedSrc, ConstantInt::get(TypeOfCopyLen, GepIndex));
- LoadInst *Load =
- RBuilder.CreateAlignedLoad(OpTy, SrcGEP, PartSrcAlign, SrcIsVolatile);
- if (!CanOverlap) {
- // Set alias scope for loads.
- Load->setMetadata(LLVMContext::MD_alias_scope,
- MDNode::get(Ctx, NewScope));
- }
- // Cast destination to operand type and store.
- PointerType *DstPtrType = PointerType::get(OpTy, DstAS);
- Value *CastedDst = DstAddr->getType() == DstPtrType
- ? DstAddr
- : RBuilder.CreateBitCast(DstAddr, DstPtrType);
- Value *DstGEP = RBuilder.CreateInBoundsGEP(
- OpTy, CastedDst, ConstantInt::get(TypeOfCopyLen, GepIndex));
- StoreInst *Store = RBuilder.CreateAlignedStore(Load, DstGEP, PartDstAlign,
- DstIsVolatile);
- if (!CanOverlap) {
- // Indicate that stores don't overlap loads.
- Store->setMetadata(LLVMContext::MD_noalias, MDNode::get(Ctx, NewScope));
- }
- if (AtomicElementSize) {
- Load->setAtomic(AtomicOrdering::Unordered);
- Store->setAtomic(AtomicOrdering::Unordered);
- }
- BytesCopied += OperandSize;
- }
- }
- assert(BytesCopied == CopyLen->getZExtValue() &&
- "Bytes copied should match size in the call!");
- }
- void llvm::createMemCpyLoopUnknownSize(
- Instruction *InsertBefore, Value *SrcAddr, Value *DstAddr, Value *CopyLen,
- Align SrcAlign, Align DstAlign, bool SrcIsVolatile, bool DstIsVolatile,
- bool CanOverlap, const TargetTransformInfo &TTI,
- std::optional<uint32_t> AtomicElementSize) {
- BasicBlock *PreLoopBB = InsertBefore->getParent();
- BasicBlock *PostLoopBB =
- PreLoopBB->splitBasicBlock(InsertBefore, "post-loop-memcpy-expansion");
- Function *ParentFunc = PreLoopBB->getParent();
- const DataLayout &DL = ParentFunc->getParent()->getDataLayout();
- LLVMContext &Ctx = PreLoopBB->getContext();
- MDBuilder MDB(Ctx);
- MDNode *NewDomain = MDB.createAnonymousAliasScopeDomain("MemCopyDomain");
- StringRef Name = "MemCopyAliasScope";
- MDNode *NewScope = MDB.createAnonymousAliasScope(NewDomain, Name);
- unsigned SrcAS = cast<PointerType>(SrcAddr->getType())->getAddressSpace();
- unsigned DstAS = cast<PointerType>(DstAddr->getType())->getAddressSpace();
- Type *LoopOpType = TTI.getMemcpyLoopLoweringType(
- Ctx, CopyLen, SrcAS, DstAS, SrcAlign.value(), DstAlign.value(),
- AtomicElementSize);
- assert((!AtomicElementSize || !LoopOpType->isVectorTy()) &&
- "Atomic memcpy lowering is not supported for vector operand type");
- unsigned LoopOpSize = DL.getTypeStoreSize(LoopOpType);
- assert((!AtomicElementSize || LoopOpSize % *AtomicElementSize == 0) &&
- "Atomic memcpy lowering is not supported for selected operand size");
- IRBuilder<> PLBuilder(PreLoopBB->getTerminator());
- PointerType *SrcOpType = PointerType::get(LoopOpType, SrcAS);
- PointerType *DstOpType = PointerType::get(LoopOpType, DstAS);
- if (SrcAddr->getType() != SrcOpType) {
- SrcAddr = PLBuilder.CreateBitCast(SrcAddr, SrcOpType);
- }
- if (DstAddr->getType() != DstOpType) {
- DstAddr = PLBuilder.CreateBitCast(DstAddr, DstOpType);
- }
- // Calculate the loop trip count, and remaining bytes to copy after the loop.
- Type *CopyLenType = CopyLen->getType();
- IntegerType *ILengthType = dyn_cast<IntegerType>(CopyLenType);
- assert(ILengthType &&
- "expected size argument to memcpy to be an integer type!");
- Type *Int8Type = Type::getInt8Ty(Ctx);
- bool LoopOpIsInt8 = LoopOpType == Int8Type;
- ConstantInt *CILoopOpSize = ConstantInt::get(ILengthType, LoopOpSize);
- Value *RuntimeLoopCount = LoopOpIsInt8 ?
- CopyLen :
- PLBuilder.CreateUDiv(CopyLen, CILoopOpSize);
- BasicBlock *LoopBB =
- BasicBlock::Create(Ctx, "loop-memcpy-expansion", ParentFunc, PostLoopBB);
- IRBuilder<> LoopBuilder(LoopBB);
- Align PartSrcAlign(commonAlignment(SrcAlign, LoopOpSize));
- Align PartDstAlign(commonAlignment(DstAlign, LoopOpSize));
- PHINode *LoopIndex = LoopBuilder.CreatePHI(CopyLenType, 2, "loop-index");
- LoopIndex->addIncoming(ConstantInt::get(CopyLenType, 0U), PreLoopBB);
- Value *SrcGEP = LoopBuilder.CreateInBoundsGEP(LoopOpType, SrcAddr, LoopIndex);
- LoadInst *Load = LoopBuilder.CreateAlignedLoad(LoopOpType, SrcGEP,
- PartSrcAlign, SrcIsVolatile);
- if (!CanOverlap) {
- // Set alias scope for loads.
- Load->setMetadata(LLVMContext::MD_alias_scope, MDNode::get(Ctx, NewScope));
- }
- Value *DstGEP = LoopBuilder.CreateInBoundsGEP(LoopOpType, DstAddr, LoopIndex);
- StoreInst *Store =
- LoopBuilder.CreateAlignedStore(Load, DstGEP, PartDstAlign, DstIsVolatile);
- if (!CanOverlap) {
- // Indicate that stores don't overlap loads.
- Store->setMetadata(LLVMContext::MD_noalias, MDNode::get(Ctx, NewScope));
- }
- if (AtomicElementSize) {
- Load->setAtomic(AtomicOrdering::Unordered);
- Store->setAtomic(AtomicOrdering::Unordered);
- }
- Value *NewIndex =
- LoopBuilder.CreateAdd(LoopIndex, ConstantInt::get(CopyLenType, 1U));
- LoopIndex->addIncoming(NewIndex, LoopBB);
- bool requiresResidual =
- !LoopOpIsInt8 && !(AtomicElementSize && LoopOpSize == AtomicElementSize);
- if (requiresResidual) {
- Type *ResLoopOpType = AtomicElementSize
- ? Type::getIntNTy(Ctx, *AtomicElementSize * 8)
- : Int8Type;
- unsigned ResLoopOpSize = DL.getTypeStoreSize(ResLoopOpType);
- assert((ResLoopOpSize == AtomicElementSize ? *AtomicElementSize : 1) &&
- "Store size is expected to match type size");
- // Add in the
- Value *RuntimeResidual = PLBuilder.CreateURem(CopyLen, CILoopOpSize);
- Value *RuntimeBytesCopied = PLBuilder.CreateSub(CopyLen, RuntimeResidual);
- // Loop body for the residual copy.
- BasicBlock *ResLoopBB = BasicBlock::Create(Ctx, "loop-memcpy-residual",
- PreLoopBB->getParent(),
- PostLoopBB);
- // Residual loop header.
- BasicBlock *ResHeaderBB = BasicBlock::Create(
- Ctx, "loop-memcpy-residual-header", PreLoopBB->getParent(), nullptr);
- // Need to update the pre-loop basic block to branch to the correct place.
- // branch to the main loop if the count is non-zero, branch to the residual
- // loop if the copy size is smaller then 1 iteration of the main loop but
- // non-zero and finally branch to after the residual loop if the memcpy
- // size is zero.
- ConstantInt *Zero = ConstantInt::get(ILengthType, 0U);
- PLBuilder.CreateCondBr(PLBuilder.CreateICmpNE(RuntimeLoopCount, Zero),
- LoopBB, ResHeaderBB);
- PreLoopBB->getTerminator()->eraseFromParent();
- LoopBuilder.CreateCondBr(
- LoopBuilder.CreateICmpULT(NewIndex, RuntimeLoopCount), LoopBB,
- ResHeaderBB);
- // Determine if we need to branch to the residual loop or bypass it.
- IRBuilder<> RHBuilder(ResHeaderBB);
- RHBuilder.CreateCondBr(RHBuilder.CreateICmpNE(RuntimeResidual, Zero),
- ResLoopBB, PostLoopBB);
- // Copy the residual with single byte load/store loop.
- IRBuilder<> ResBuilder(ResLoopBB);
- PHINode *ResidualIndex =
- ResBuilder.CreatePHI(CopyLenType, 2, "residual-loop-index");
- ResidualIndex->addIncoming(Zero, ResHeaderBB);
- Value *SrcAsResLoopOpType = ResBuilder.CreateBitCast(
- SrcAddr, PointerType::get(ResLoopOpType, SrcAS));
- Value *DstAsResLoopOpType = ResBuilder.CreateBitCast(
- DstAddr, PointerType::get(ResLoopOpType, DstAS));
- Value *FullOffset = ResBuilder.CreateAdd(RuntimeBytesCopied, ResidualIndex);
- Value *SrcGEP = ResBuilder.CreateInBoundsGEP(
- ResLoopOpType, SrcAsResLoopOpType, FullOffset);
- LoadInst *Load = ResBuilder.CreateAlignedLoad(ResLoopOpType, SrcGEP,
- PartSrcAlign, SrcIsVolatile);
- if (!CanOverlap) {
- // Set alias scope for loads.
- Load->setMetadata(LLVMContext::MD_alias_scope,
- MDNode::get(Ctx, NewScope));
- }
- Value *DstGEP = ResBuilder.CreateInBoundsGEP(
- ResLoopOpType, DstAsResLoopOpType, FullOffset);
- StoreInst *Store = ResBuilder.CreateAlignedStore(Load, DstGEP, PartDstAlign,
- DstIsVolatile);
- if (!CanOverlap) {
- // Indicate that stores don't overlap loads.
- Store->setMetadata(LLVMContext::MD_noalias, MDNode::get(Ctx, NewScope));
- }
- if (AtomicElementSize) {
- Load->setAtomic(AtomicOrdering::Unordered);
- Store->setAtomic(AtomicOrdering::Unordered);
- }
- Value *ResNewIndex = ResBuilder.CreateAdd(
- ResidualIndex, ConstantInt::get(CopyLenType, ResLoopOpSize));
- ResidualIndex->addIncoming(ResNewIndex, ResLoopBB);
- // Create the loop branch condition.
- ResBuilder.CreateCondBr(
- ResBuilder.CreateICmpULT(ResNewIndex, RuntimeResidual), ResLoopBB,
- PostLoopBB);
- } else {
- // In this case the loop operand type was a byte, and there is no need for a
- // residual loop to copy the remaining memory after the main loop.
- // We do however need to patch up the control flow by creating the
- // terminators for the preloop block and the memcpy loop.
- ConstantInt *Zero = ConstantInt::get(ILengthType, 0U);
- PLBuilder.CreateCondBr(PLBuilder.CreateICmpNE(RuntimeLoopCount, Zero),
- LoopBB, PostLoopBB);
- PreLoopBB->getTerminator()->eraseFromParent();
- LoopBuilder.CreateCondBr(
- LoopBuilder.CreateICmpULT(NewIndex, RuntimeLoopCount), LoopBB,
- PostLoopBB);
- }
- }
- // Lower memmove to IR. memmove is required to correctly copy overlapping memory
- // regions; therefore, it has to check the relative positions of the source and
- // destination pointers and choose the copy direction accordingly.
- //
- // The code below is an IR rendition of this C function:
- //
- // void* memmove(void* dst, const void* src, size_t n) {
- // unsigned char* d = dst;
- // const unsigned char* s = src;
- // if (s < d) {
- // // copy backwards
- // while (n--) {
- // d[n] = s[n];
- // }
- // } else {
- // // copy forward
- // for (size_t i = 0; i < n; ++i) {
- // d[i] = s[i];
- // }
- // }
- // return dst;
- // }
- static void createMemMoveLoop(Instruction *InsertBefore, Value *SrcAddr,
- Value *DstAddr, Value *CopyLen, Align SrcAlign,
- Align DstAlign, bool SrcIsVolatile,
- bool DstIsVolatile) {
- Type *TypeOfCopyLen = CopyLen->getType();
- BasicBlock *OrigBB = InsertBefore->getParent();
- Function *F = OrigBB->getParent();
- const DataLayout &DL = F->getParent()->getDataLayout();
- // TODO: Use different element type if possible?
- IRBuilder<> CastBuilder(InsertBefore);
- Type *EltTy = CastBuilder.getInt8Ty();
- Type *PtrTy =
- CastBuilder.getInt8PtrTy(SrcAddr->getType()->getPointerAddressSpace());
- SrcAddr = CastBuilder.CreateBitCast(SrcAddr, PtrTy);
- DstAddr = CastBuilder.CreateBitCast(DstAddr, PtrTy);
- // Create the a comparison of src and dst, based on which we jump to either
- // the forward-copy part of the function (if src >= dst) or the backwards-copy
- // part (if src < dst).
- // SplitBlockAndInsertIfThenElse conveniently creates the basic if-then-else
- // structure. Its block terminators (unconditional branches) are replaced by
- // the appropriate conditional branches when the loop is built.
- ICmpInst *PtrCompare = new ICmpInst(InsertBefore, ICmpInst::ICMP_ULT,
- SrcAddr, DstAddr, "compare_src_dst");
- Instruction *ThenTerm, *ElseTerm;
- SplitBlockAndInsertIfThenElse(PtrCompare, InsertBefore, &ThenTerm,
- &ElseTerm);
- // Each part of the function consists of two blocks:
- // copy_backwards: used to skip the loop when n == 0
- // copy_backwards_loop: the actual backwards loop BB
- // copy_forward: used to skip the loop when n == 0
- // copy_forward_loop: the actual forward loop BB
- BasicBlock *CopyBackwardsBB = ThenTerm->getParent();
- CopyBackwardsBB->setName("copy_backwards");
- BasicBlock *CopyForwardBB = ElseTerm->getParent();
- CopyForwardBB->setName("copy_forward");
- BasicBlock *ExitBB = InsertBefore->getParent();
- ExitBB->setName("memmove_done");
- unsigned PartSize = DL.getTypeStoreSize(EltTy);
- Align PartSrcAlign(commonAlignment(SrcAlign, PartSize));
- Align PartDstAlign(commonAlignment(DstAlign, PartSize));
- // Initial comparison of n == 0 that lets us skip the loops altogether. Shared
- // between both backwards and forward copy clauses.
- ICmpInst *CompareN =
- new ICmpInst(OrigBB->getTerminator(), ICmpInst::ICMP_EQ, CopyLen,
- ConstantInt::get(TypeOfCopyLen, 0), "compare_n_to_0");
- // Copying backwards.
- BasicBlock *LoopBB =
- BasicBlock::Create(F->getContext(), "copy_backwards_loop", F, CopyForwardBB);
- IRBuilder<> LoopBuilder(LoopBB);
- PHINode *LoopPhi = LoopBuilder.CreatePHI(TypeOfCopyLen, 0);
- Value *IndexPtr = LoopBuilder.CreateSub(
- LoopPhi, ConstantInt::get(TypeOfCopyLen, 1), "index_ptr");
- Value *Element = LoopBuilder.CreateAlignedLoad(
- EltTy, LoopBuilder.CreateInBoundsGEP(EltTy, SrcAddr, IndexPtr),
- PartSrcAlign, "element");
- LoopBuilder.CreateAlignedStore(
- Element, LoopBuilder.CreateInBoundsGEP(EltTy, DstAddr, IndexPtr),
- PartDstAlign);
- LoopBuilder.CreateCondBr(
- LoopBuilder.CreateICmpEQ(IndexPtr, ConstantInt::get(TypeOfCopyLen, 0)),
- ExitBB, LoopBB);
- LoopPhi->addIncoming(IndexPtr, LoopBB);
- LoopPhi->addIncoming(CopyLen, CopyBackwardsBB);
- BranchInst::Create(ExitBB, LoopBB, CompareN, ThenTerm);
- ThenTerm->eraseFromParent();
- // Copying forward.
- BasicBlock *FwdLoopBB =
- BasicBlock::Create(F->getContext(), "copy_forward_loop", F, ExitBB);
- IRBuilder<> FwdLoopBuilder(FwdLoopBB);
- PHINode *FwdCopyPhi = FwdLoopBuilder.CreatePHI(TypeOfCopyLen, 0, "index_ptr");
- Value *SrcGEP = FwdLoopBuilder.CreateInBoundsGEP(EltTy, SrcAddr, FwdCopyPhi);
- Value *FwdElement =
- FwdLoopBuilder.CreateAlignedLoad(EltTy, SrcGEP, PartSrcAlign, "element");
- Value *DstGEP = FwdLoopBuilder.CreateInBoundsGEP(EltTy, DstAddr, FwdCopyPhi);
- FwdLoopBuilder.CreateAlignedStore(FwdElement, DstGEP, PartDstAlign);
- Value *FwdIndexPtr = FwdLoopBuilder.CreateAdd(
- FwdCopyPhi, ConstantInt::get(TypeOfCopyLen, 1), "index_increment");
- FwdLoopBuilder.CreateCondBr(FwdLoopBuilder.CreateICmpEQ(FwdIndexPtr, CopyLen),
- ExitBB, FwdLoopBB);
- FwdCopyPhi->addIncoming(FwdIndexPtr, FwdLoopBB);
- FwdCopyPhi->addIncoming(ConstantInt::get(TypeOfCopyLen, 0), CopyForwardBB);
- BranchInst::Create(ExitBB, FwdLoopBB, CompareN, ElseTerm);
- ElseTerm->eraseFromParent();
- }
- static void createMemSetLoop(Instruction *InsertBefore, Value *DstAddr,
- Value *CopyLen, Value *SetValue, Align DstAlign,
- bool IsVolatile) {
- Type *TypeOfCopyLen = CopyLen->getType();
- BasicBlock *OrigBB = InsertBefore->getParent();
- Function *F = OrigBB->getParent();
- const DataLayout &DL = F->getParent()->getDataLayout();
- BasicBlock *NewBB =
- OrigBB->splitBasicBlock(InsertBefore, "split");
- BasicBlock *LoopBB
- = BasicBlock::Create(F->getContext(), "loadstoreloop", F, NewBB);
- IRBuilder<> Builder(OrigBB->getTerminator());
- // Cast pointer to the type of value getting stored
- unsigned dstAS = cast<PointerType>(DstAddr->getType())->getAddressSpace();
- DstAddr = Builder.CreateBitCast(DstAddr,
- PointerType::get(SetValue->getType(), dstAS));
- Builder.CreateCondBr(
- Builder.CreateICmpEQ(ConstantInt::get(TypeOfCopyLen, 0), CopyLen), NewBB,
- LoopBB);
- OrigBB->getTerminator()->eraseFromParent();
- unsigned PartSize = DL.getTypeStoreSize(SetValue->getType());
- Align PartAlign(commonAlignment(DstAlign, PartSize));
- IRBuilder<> LoopBuilder(LoopBB);
- PHINode *LoopIndex = LoopBuilder.CreatePHI(TypeOfCopyLen, 0);
- LoopIndex->addIncoming(ConstantInt::get(TypeOfCopyLen, 0), OrigBB);
- LoopBuilder.CreateAlignedStore(
- SetValue,
- LoopBuilder.CreateInBoundsGEP(SetValue->getType(), DstAddr, LoopIndex),
- PartAlign, IsVolatile);
- Value *NewIndex =
- LoopBuilder.CreateAdd(LoopIndex, ConstantInt::get(TypeOfCopyLen, 1));
- LoopIndex->addIncoming(NewIndex, LoopBB);
- LoopBuilder.CreateCondBr(LoopBuilder.CreateICmpULT(NewIndex, CopyLen), LoopBB,
- NewBB);
- }
- template <typename T>
- static bool canOverlap(MemTransferBase<T> *Memcpy, ScalarEvolution *SE) {
- if (SE) {
- auto *SrcSCEV = SE->getSCEV(Memcpy->getRawSource());
- auto *DestSCEV = SE->getSCEV(Memcpy->getRawDest());
- if (SE->isKnownPredicateAt(CmpInst::ICMP_NE, SrcSCEV, DestSCEV, Memcpy))
- return false;
- }
- return true;
- }
- void llvm::expandMemCpyAsLoop(MemCpyInst *Memcpy,
- const TargetTransformInfo &TTI,
- ScalarEvolution *SE) {
- bool CanOverlap = canOverlap(Memcpy, SE);
- if (ConstantInt *CI = dyn_cast<ConstantInt>(Memcpy->getLength())) {
- createMemCpyLoopKnownSize(
- /* InsertBefore */ Memcpy,
- /* SrcAddr */ Memcpy->getRawSource(),
- /* DstAddr */ Memcpy->getRawDest(),
- /* CopyLen */ CI,
- /* SrcAlign */ Memcpy->getSourceAlign().valueOrOne(),
- /* DestAlign */ Memcpy->getDestAlign().valueOrOne(),
- /* SrcIsVolatile */ Memcpy->isVolatile(),
- /* DstIsVolatile */ Memcpy->isVolatile(),
- /* CanOverlap */ CanOverlap,
- /* TargetTransformInfo */ TTI);
- } else {
- createMemCpyLoopUnknownSize(
- /* InsertBefore */ Memcpy,
- /* SrcAddr */ Memcpy->getRawSource(),
- /* DstAddr */ Memcpy->getRawDest(),
- /* CopyLen */ Memcpy->getLength(),
- /* SrcAlign */ Memcpy->getSourceAlign().valueOrOne(),
- /* DestAlign */ Memcpy->getDestAlign().valueOrOne(),
- /* SrcIsVolatile */ Memcpy->isVolatile(),
- /* DstIsVolatile */ Memcpy->isVolatile(),
- /* CanOverlap */ CanOverlap,
- /* TargetTransformInfo */ TTI);
- }
- }
- void llvm::expandMemMoveAsLoop(MemMoveInst *Memmove) {
- createMemMoveLoop(/* InsertBefore */ Memmove,
- /* SrcAddr */ Memmove->getRawSource(),
- /* DstAddr */ Memmove->getRawDest(),
- /* CopyLen */ Memmove->getLength(),
- /* SrcAlign */ Memmove->getSourceAlign().valueOrOne(),
- /* DestAlign */ Memmove->getDestAlign().valueOrOne(),
- /* SrcIsVolatile */ Memmove->isVolatile(),
- /* DstIsVolatile */ Memmove->isVolatile());
- }
- void llvm::expandMemSetAsLoop(MemSetInst *Memset) {
- createMemSetLoop(/* InsertBefore */ Memset,
- /* DstAddr */ Memset->getRawDest(),
- /* CopyLen */ Memset->getLength(),
- /* SetValue */ Memset->getValue(),
- /* Alignment */ Memset->getDestAlign().valueOrOne(),
- Memset->isVolatile());
- }
- void llvm::expandAtomicMemCpyAsLoop(AtomicMemCpyInst *AtomicMemcpy,
- const TargetTransformInfo &TTI,
- ScalarEvolution *SE) {
- if (ConstantInt *CI = dyn_cast<ConstantInt>(AtomicMemcpy->getLength())) {
- createMemCpyLoopKnownSize(
- /* InsertBefore */ AtomicMemcpy,
- /* SrcAddr */ AtomicMemcpy->getRawSource(),
- /* DstAddr */ AtomicMemcpy->getRawDest(),
- /* CopyLen */ CI,
- /* SrcAlign */ AtomicMemcpy->getSourceAlign().valueOrOne(),
- /* DestAlign */ AtomicMemcpy->getDestAlign().valueOrOne(),
- /* SrcIsVolatile */ AtomicMemcpy->isVolatile(),
- /* DstIsVolatile */ AtomicMemcpy->isVolatile(),
- /* CanOverlap */ false, // SrcAddr & DstAddr may not overlap by spec.
- /* TargetTransformInfo */ TTI,
- /* AtomicCpySize */ AtomicMemcpy->getElementSizeInBytes());
- } else {
- createMemCpyLoopUnknownSize(
- /* InsertBefore */ AtomicMemcpy,
- /* SrcAddr */ AtomicMemcpy->getRawSource(),
- /* DstAddr */ AtomicMemcpy->getRawDest(),
- /* CopyLen */ AtomicMemcpy->getLength(),
- /* SrcAlign */ AtomicMemcpy->getSourceAlign().valueOrOne(),
- /* DestAlign */ AtomicMemcpy->getDestAlign().valueOrOne(),
- /* SrcIsVolatile */ AtomicMemcpy->isVolatile(),
- /* DstIsVolatile */ AtomicMemcpy->isVolatile(),
- /* CanOverlap */ false, // SrcAddr & DstAddr may not overlap by spec.
- /* TargetTransformInfo */ TTI,
- /* AtomicCpySize */ AtomicMemcpy->getElementSizeInBytes());
- }
- }
|