[klee-dev] Using LLVM cost modeling in KLEE
Matthew Schmitt
mschmit1 at stevens.edu
Wed Jun 18 20:01:50 BST 2025
We are currently attempting to explore methods that would add a provision to KLEE so that a cost metric can be produced for each path that KLEE explores within a program. The instruction cost metrics provided by the LLVM optimizers appear to be sufficient for our purposes, as they provide values that are reflective of the latency of each LLVM bytecode instruction on the particular CPU architecture that the program in question is being targeted for. We would like to replicate the functionality provided by passes like the one located at llvm/Analysis/CostModel.h.
Unfortunately, we have encountered some issues with this when it comes to returning the true latency-aware costs for instructions such as floating-point division (fdiv). It is my understanding that costs such as this are highly architecture specific, though will always be more expensive than multiplication, addition, or subtraction, hence why LLVM must collect information about the current target. In our case, even after this is completed successfully, LLVM encounters some error, and handles it by reverting to a generic cost model that returns costs that are simply unstable for the purposes of this extension, such as returning the same costs for all floating-point operations.
We are currently attempting to implement this feature as presented below:
std::unique_ptr<raw_fd_ostream> InstructionStats = interpreterHandler->openOutputFile("stats.csv");
bindModuleConstants();
InitializeNativeTarget();
InitializeNativeTargetAsmPrinter();
InitializeNativeTargetAsmParser();
Expected<llvm::orc::JITTargetMachineBuilder> JITBE = llvm::orc::JITTargetMachineBuilder::detectHost();
if (!JITBE) klee_error("Failed to detect host target machine");
llvm::orc::JITTargetMachineBuilder JITB = std::move(*JITBE);
Expected<std::unique_ptr<TargetMachine>> TME = JITB.createTargetMachine();
if (!TME) klee_error("Failed to create target machine");
TMown = std::move(*TME);
TargetMachine *TM = TMown.get();
if (InstructionStats) {
*InstructionStats << "Target Triple: " << TM->getTargetTriple().str() << "\n";
*InstructionStats << "CPU: " << TM->getTargetCPU() << "\n";
*InstructionStats << "Features: " << TM->getTargetFeatureString() << "\n";
*InstructionStats << "Instruction/Call, Cost" << "\n";
}
else {
klee_error("Error: could not access file");
}
/* The target machine seems to be constructed correctly. All values appear to reflect the CPU architecture of the current machine. */
Instruction *i = ki->inst;
Function *f = state.stack.back().kf->function;
TargetIRAnalysis TIRA = (TM ? TM->getTargetIRAnalysis() : TargetIRAnalysis());
Module *M = f->getParent();
M->setDataLayout(TM->createDataLayout());
M->setTargetTriple(TM->getTargetTriple().getTriple());
TargetTransformInfo TTI = TM->getTargetTransformInfo(*f);
InstructionCost Cost = TTI.getInstructionCost(i, TargetTransformInfo::TCK_Latency);
/* This cost value is wrong and appears to reflect costs of the generic TTI implementation, not the x86 implementation */
We would appreciate any assistance in this regard, particularly concerning advice as to how getInstructionCost can be effectively utilized outside of an LLVM optimizer pass.
-------------- next part --------------
HTML attachment scrubbed and removed
More information about the klee-dev
mailing list