// -*- C++ -*- (c) 2015 Vladimír Štill // DIVINE_RELAX_WARNINGS #include #include #include #include #include #include #include #include #include #include DIVINE_UNRELAX_WARNINGS #include #include #include #include #include #include #include #include #include namespace lart { namespace reduction { /*MD # Const Conditional Jump Elimination */ struct ConstConditionalJumpElimination : lart::Pass { static PassMeta meta() { return passMeta< ConstConditionalJumpElimination >( "ConstConditionalJumpElimination", "Replace constant conditional jumps with unconditional branches." ); } llvm::BasicBlock *constJump( llvm::BasicBlock &bb ) { if ( auto br = llvm::dyn_cast< llvm::BranchInst >( bb.getTerminator() ) ) { if ( br->isConditional() ) { if ( auto cond = llvm::dyn_cast< llvm::ConstantInt >( br->getCondition() ) ) { ++changed; ASSERT_EQ( br->getNumSuccessors(), 2u ); auto unlinked = br->getSuccessor( !cond->isZero() ); unlinked->removePredecessor( &bb ); llvm::ReplaceInstWithInst( br, llvm::BranchInst::Create( br->getSuccessor( cond->isZero() ) ) ); return llvm::pred_begin( unlinked ) == llvm::pred_end( unlinked ) ? unlinked : nullptr; } } } return nullptr; } std::vector< llvm::BasicBlock * > removeUnused( llvm::BasicBlock *bb ) { std::vector< llvm::BasicBlock * > succs; auto term = bb->getTerminator(); ASSERT( term ); int cnt = term->getNumSuccessors(); for ( int i = 0; i < cnt; ++i ) succs.push_back( term->getSuccessor( i ) ); llvm::DeleteDeadBlock( bb ); std::vector< llvm::BasicBlock * > unreachable; for ( auto succ : succs ) { if ( llvm::pred_begin( succ ) == llvm::pred_end( succ ) && succ != bb ) unreachable.push_back( succ ); } ++removed; return unreachable; } using lart::Pass::run; llvm::PreservedAnalyses run( llvm::Module &m ) override { for ( auto &fn : m ) { std::unordered_set< llvm::BasicBlock * > unreachable; for ( auto &bb : fn ) if ( auto unr = constJump( bb ) ) unreachable.insert( unr ); while ( !unreachable.empty() ) { auto todo = *unreachable.begin(); unreachable.erase( todo ); auto unr = removeUnused( todo ); unreachable.insert( unr.begin(), unr.end() ); } } // std::cerr << "INFO: changed " << changed << " branches, removed " << removed << " basic blocks" << std::endl; return llvm::PreservedAnalyses::none(); } private: long changed = 0; long removed = 0; }; /*MD # Merge Basic Blocks */ struct MergeBasicBlocks : lart::Pass { static PassMeta meta() { return passMeta< MergeBasicBlocks >( "MergeBasicBlocks" ); } void mergeBB( llvm::BasicBlock *bb ) { std::unordered_set< llvm::BasicBlock * > seen; mergeBB( 0, bb, seen ); } void mergeBB( int parsuccs, llvm::BasicBlock *bb, std::unordered_set< llvm::BasicBlock * > &seen ) { if ( seen.count( bb ) ) return; seen.insert( bb ); auto term = bb->getTerminator(); int cnt = term->getNumSuccessors(); for ( int i = 0; i < cnt; ++i ) mergeBB( cnt, term->getSuccessor( i ), seen ); if ( parsuccs == 1 ) // this checks if remove is safe if ( llvm::MergeBlockIntoPredecessor( bb ) ) ++merged; } using lart::Pass::run; llvm::PreservedAnalyses run( llvm::Module &m ) override { for ( auto &f : m ) if ( !f.empty() ) mergeBB( &f.getEntryBlock() ); // std::cerr << "INFO: merged " << merged << " basic blocks" << std::endl; return llvm::PreservedAnalyses::none(); } long merged = 0; }; /*MD # Const Alloca Elimination Clang often generates allocas even for variables which are not changed ever, those can be eliminated without changing properties of parallel program. The alloca can be eliminated if: 1. it does not escape function's scope 2. is only accessed by load and store instruction 3. there is only one store 4. the store dominates all loads Not that calls to llvm.dbg.declare do not count as uses. */ struct ConstAllocaElimination : lart::Pass { static auto meta() { return passMeta< ConstAllocaElimination >( "ConstAllocaElimination" ); } void processFunction( llvm::Function &fn ) { std::vector< std::pair< llvm::AllocaInst *, llvm::StoreInst * > > vars; auto dt = brick::types::lazy( [&]() { return llvm::DominatorTreeAnalysis().run( fn ); } ); for ( auto &inst: query::query( fn ).flatten() ) { llvmcase( inst, [&]( llvm::AllocaInst *alloca ) { ++allAllocas; if ( !llvm::PointerMayBeCaptured( alloca, false, true ) ) { auto users = query::query( alloca->users() ).filter( query::isnot< llvm::DbgDeclareInst > ); llvm::StoreInst *store = users.map( query::llvmdyncast< llvm::StoreInst > ) .filter( query::notnull ) .singleton() .getOr( nullptr ); if ( store && store->getPointerOperand() == alloca // single store && users.filter( query::isnot< llvm::StoreInst > ) .all( query::is< llvm::LoadInst > ) // + only loads and dbg info && users.map( query::llvmdyncast< llvm::LoadInst > ) // store dominates all loads .filter( query::notnull ) .all( [&]( auto l ) { return dt->dominates( store, l ); } ) ) vars.emplace_back( alloca, store ); } } ); } for ( auto var : vars ) { auto val = var.second->getValueOperand(); std::vector< llvm::Instruction * > toDelete; for ( auto user : var.first->users() ) llvmcase( user, [val,&toDelete]( llvm::LoadInst *load ) { load->replaceAllUsesWith( val ); toDelete.push_back( load ); }, [&toDelete]( llvm::DbgDeclareInst *dbg ) { toDelete.push_back( dbg ); // TODO: fixme: copy debuginfo somehow }, [var,&toDelete]( llvm::StoreInst *store ) { ASSERT_EQ( store, var.second ); toDelete.push_back( store ); }, [&]( llvm::Value *val ) { std::cerr << "in " << fn.getName().str() << std::endl; fn.dump(); var.first->dump(); var.second->dump(); val->dump(); UNREACHABLE( "unhandled case" ); } ); for ( auto i : toDelete ) i->eraseFromParent(); var.first->eraseFromParent(); ++deletedAllocas; } } using lart::Pass::run; llvm::PreservedAnalyses run( llvm::Module &m ) override { for ( auto &fn : m ) processFunction( fn ); // std::cerr << "INFO: removed " << deletedAllocas << " out of " << allAllocas // << " (" << double( 100 * deletedAllocas ) / allAllocas // << "%) allocas" << std::endl; return llvm::PreservedAnalyses::none(); } long allAllocas = 0; long deletedAllocas = 0; }; PassMeta paroptPass() { return compositePassMeta< ConstConditionalJumpElimination, MergeBasicBlocks , ConstAllocaElimination >( "paropt", "Parallel-safe optimizations" ); } } // namespace reduce } // namespace lart