// Part of the Concrete Compiler Project, under the BSD3 License with Zama // Exceptions. See // https://github.com/zama-ai/concrete-compiler-internal/blob/master/LICENSE.txt // for license information. #include "llvm/Support/Error.h" #include #include #include #include #include #include #include #include #include namespace mlir { namespace concretelang { llvm::Expected> JITLambda::create(llvm::StringRef name, mlir::ModuleOp &module, llvm::function_ref optPipeline, llvm::Optional runtimeLibPath) { // Looking for the function auto rangeOps = module.getOps(); auto funcOp = llvm::find_if(rangeOps, [&](mlir::LLVM::LLVMFuncOp op) { return op.getName() == name; }); if (funcOp == rangeOps.end()) { return llvm::make_error( "cannot find the function to JIT", llvm::inconvertibleErrorCode()); } llvm::InitializeNativeTarget(); llvm::InitializeNativeTargetAsmPrinter(); mlir::registerLLVMDialectTranslation(*module->getContext()); // Create an MLIR execution engine. The execution engine eagerly // JIT-compiles the module. If runtimeLibPath is specified, it's passed as a // shared library to the JIT compiler. std::vector sharedLibPaths; if (runtimeLibPath.hasValue()) sharedLibPaths.push_back(runtimeLibPath.getValue()); auto maybeEngine = mlir::ExecutionEngine::create( module, /*llvmModuleBuilder=*/nullptr, optPipeline, /*jitCodeGenOptLevel=*/llvm::None, sharedLibPaths); if (!maybeEngine) { return StreamStringError("failed to construct the MLIR ExecutionEngine"); } auto &engine = maybeEngine.get(); auto lambda = std::make_unique((*funcOp).getType(), name); lambda->engine = std::move(engine); return std::move(lambda); } llvm::Error JITLambda::invokeRaw(llvm::MutableArrayRef args) { auto found = std::find(args.begin(), args.end(), nullptr); if (found == args.end()) { return this->engine->invokePacked(this->name, args); } int pos = found - args.begin(); return StreamStringError("invoke: argument at pos ") << pos << " is null or missing"; } llvm::Error JITLambda::invoke(Argument &args) { size_t expectedInputs = this->type.getNumParams(); size_t actualInputs = args.inputs.size(); if (expectedInputs == actualInputs) { return invokeRaw(args.rawArg); } return StreamStringError("invokeRaw: received ") << actualInputs << "arguments instead of " << expectedInputs; } JITLambda::Argument::Argument(KeySet &keySet) : keySet(keySet) { // Setting the inputs auto numInputs = 0; { for (size_t i = 0; i < keySet.numInputs(); i++) { auto offset = numInputs; auto gate = keySet.inputGate(i); inputGates.push_back({gate, offset}); if (keySet.inputGate(i).shape.dimensions.empty()) { // scalar gate numInputs = numInputs + 1; continue; } // memref gate, as we follow the standard calling convention numInputs = numInputs + 3; // Offsets and strides are array of size N where N is the number of // dimension of the tensor. numInputs = numInputs + 2 * keySet.inputGate(i).shape.dimensions.size(); } // Reserve for the context argument numInputs = numInputs + 1; inputs = std::vector(numInputs); } // Setting the outputs { auto numOutputs = 0; for (size_t i = 0; i < keySet.numOutputs(); i++) { auto offset = numOutputs; auto gate = keySet.outputGate(i); outputGates.push_back({gate, offset}); if (gate.shape.dimensions.empty()) { // scalar gate numOutputs = numOutputs + 1; continue; } // memref gate, as we follow the standard calling convention numOutputs = numOutputs + 3; // Offsets and strides are array of size N where N is the number of // dimension of the tensor. numOutputs = numOutputs + 2 * keySet.outputGate(i).shape.dimensions.size(); } outputs = std::vector(numOutputs); } // The raw argument contains pointers to inputs and pointers to store the // results rawArg = std::vector(inputs.size() + outputs.size(), nullptr); // Set the pointer on outputs on rawArg for (auto i = inputs.size(); i < rawArg.size(); i++) { rawArg[i] = &outputs[i - inputs.size()]; } // Set the context argument keySet.setRuntimeContext(context); inputs[numInputs - 1] = &context; rawArg[numInputs - 1] = &inputs[numInputs - 1]; } JITLambda::Argument::~Argument() { int err; for (auto ct : allocatedCiphertexts) { free_lwe_ciphertext_u64(&err, ct); } for (auto buffer : ciphertextBuffers) { free(buffer); } } llvm::Expected> JITLambda::Argument::create(KeySet &keySet) { auto args = std::make_unique(keySet); return std::move(args); } llvm::Error JITLambda::Argument::emitErrorIfTooManyArgs(size_t pos) { size_t arity = inputGates.size(); if (pos < arity) { return llvm::Error::success(); } return StreamStringError("The function has arity ") << arity << " but is applied to too many arguments"; } llvm::Error JITLambda::Argument::setArg(size_t pos, uint64_t arg) { if (auto error = emitErrorIfTooManyArgs(pos)) { return error; } auto gate = inputGates[pos]; auto info = std::get<0>(gate); auto offset = std::get<1>(gate); // Check is the argument is a scalar if (!info.shape.dimensions.empty()) { return llvm::make_error( llvm::Twine("argument is not a scalar: pos=").concat(llvm::Twine(pos)), llvm::inconvertibleErrorCode()); } // If argument is not encrypted, just save. if (!info.encryption.hasValue()) { inputs[offset] = (void *)arg; rawArg[offset] = &inputs[offset]; return llvm::Error::success(); } // Else if is encryted, allocate ciphertext and encrypt. LweCiphertext_u64 *ctArg; if (auto err = this->keySet.allocate_lwe(pos, &ctArg)) { return std::move(err); } allocatedCiphertexts.push_back(ctArg); if (auto err = this->keySet.encrypt_lwe(pos, ctArg, arg)) { return std::move(err); } inputs[offset] = ctArg; rawArg[offset] = &inputs[offset]; return llvm::Error::success(); } size_t bitWidthAsWord(size_t exactBitWidth) { size_t sortedWordBitWidths[] = {8, 16, 32, 64}; size_t previousWidth = 0; for (auto currentWidth : sortedWordBitWidths) { if (previousWidth < exactBitWidth && exactBitWidth <= currentWidth) { return currentWidth; } } return exactBitWidth; } llvm::Error JITLambda::Argument::setArg(size_t pos, size_t width, const void *data, llvm::ArrayRef shape) { if (auto error = emitErrorIfTooManyArgs(pos)) { return error; } auto gate = inputGates[pos]; auto info = std::get<0>(gate); auto offset = std::get<1>(gate); // Check if the width is compatible // TODO - I found this rules empirically, they are a spec somewhere? if (info.shape.width > 64) { auto msg = "Bad argument (pos=" + llvm::Twine(pos) + ") : a width of " + llvm::Twine(info.shape.width) + "bits > 64 is not supported: pos=" + llvm::Twine(pos); return llvm::make_error(msg, llvm::inconvertibleErrorCode()); } auto roundedSize = bitWidthAsWord(info.shape.width); if (width != roundedSize) { auto msg = "Bad argument (pos=" + llvm::Twine(pos) + ") : expected " + llvm::Twine(roundedSize) + "bits" + " but received " + llvm::Twine(width) + "bits (rounded from " + llvm::Twine(info.shape.width) + ")"; return llvm::make_error(msg, llvm::inconvertibleErrorCode()); } // Check the size if (info.shape.dimensions.empty()) { return llvm::make_error( llvm::Twine("argument is not a vector: pos=").concat(llvm::Twine(pos)), llvm::inconvertibleErrorCode()); } if (shape.size() != info.shape.dimensions.size()) { return llvm::make_error( llvm::Twine("tensor argument #") .concat(llvm::Twine(pos)) .concat(" has not the expected number of dimension, got ") .concat(llvm::Twine(shape.size())) .concat(" expected ") .concat(llvm::Twine(info.shape.dimensions.size())), llvm::inconvertibleErrorCode()); } for (size_t i = 0; i < shape.size(); i++) { if (shape[i] != info.shape.dimensions[i]) { return llvm::make_error( llvm::Twine("tensor argument #") .concat(llvm::Twine(pos)) .concat(" has not the expected dimension #") .concat(llvm::Twine(i)) .concat(" , got ") .concat(llvm::Twine(shape[i])) .concat(" expected ") .concat(llvm::Twine(info.shape.dimensions[i])), llvm::inconvertibleErrorCode()); } } // If argument is not encrypted, just save with the right calling convention. if (info.encryption.hasValue()) { // Else if is encrypted // For moment we support only 8 bits inputs const uint8_t *data8 = (const uint8_t *)data; if (width != 8) { return llvm::make_error( llvm::Twine( "argument width > 8 for encrypted gates are not supported: pos=") .concat(llvm::Twine(pos)), llvm::inconvertibleErrorCode()); } // Allocate a buffer for ciphertexts. auto ctBuffer = (LweCiphertext_u64 **)malloc(info.shape.size * sizeof(LweCiphertext_u64 *)); ciphertextBuffers.push_back(ctBuffer); // Allocate ciphertexts and encrypt for (size_t i = 0; i < info.shape.size; i++) { if (auto err = this->keySet.allocate_lwe(pos, &ctBuffer[i])) { return std::move(err); } allocatedCiphertexts.push_back(ctBuffer[i]); if (auto err = this->keySet.encrypt_lwe(pos, ctBuffer[i], data8[i])) { return std::move(err); } } // Replace the data by the buffer to ciphertext data = (void *)ctBuffer; } // Set the buffer as the memref calling convention expect. // allocated inputs[offset] = (void *)0; // Indicates that it's not allocated by the MLIR program rawArg[offset] = &inputs[offset]; offset++; // aligned inputs[offset] = data; rawArg[offset] = &inputs[offset]; offset++; // offset inputs[offset] = (void *)0; rawArg[offset] = &inputs[offset]; offset++; // sizes is an array of size equals to numDim for (size_t i = 0; i < shape.size(); i++) { inputs[offset] = (void *)shape[i]; rawArg[offset] = &inputs[offset]; offset++; } // Set the stride for each dimension, equal to the product of the // following dimensions. int64_t stride = 1; for (ssize_t i = shape.size() - 1; i >= 0; i--) { inputs[offset + i] = (void *)stride; rawArg[offset + i] = &inputs[offset + i]; stride *= shape[i]; } offset += shape.size(); return llvm::Error::success(); } llvm::Error JITLambda::Argument::getResult(size_t pos, uint64_t &res) { auto gate = outputGates[pos]; auto info = std::get<0>(gate); auto offset = std::get<1>(gate); // Check is the argument is a scalar if (info.shape.size != 0) { return llvm::make_error( llvm::Twine("output is not a scalar, pos=").concat(llvm::Twine(pos)), llvm::inconvertibleErrorCode()); } // If result is not encrypted, just set the result if (!info.encryption.hasValue()) { res = (uint64_t)(outputs[offset]); return llvm::Error::success(); } // Else if is encryted, decrypt LweCiphertext_u64 *ct = (LweCiphertext_u64 *)(outputs[offset]); if (auto err = this->keySet.decrypt_lwe(pos, ct, res)) { return std::move(err); } return llvm::Error::success(); } // Returns the number of elements of the result vector at position // `pos` or an error if the result is a scalar value llvm::Expected JITLambda::Argument::getResultVectorSize(size_t pos) { auto gate = outputGates[pos]; auto info = std::get<0>(gate); if (info.shape.size == 0) { return llvm::createStringError(llvm::inconvertibleErrorCode(), "Result at pos %zu is not a tensor", pos); } return info.shape.size; } // Returns the dimensions of the result tensor at position `pos` or // an error if the result is a scalar value llvm::Expected> JITLambda::Argument::getResultDimensions(size_t pos) { auto gate = outputGates[pos]; auto info = std::get<0>(gate); if (info.shape.size == 0) { return llvm::createStringError(llvm::inconvertibleErrorCode(), "Result at pos %zu is not a tensor", pos); } return info.shape.dimensions; } llvm::Expected JITLambda::Argument::getResultType(size_t pos) { if (pos >= outputGates.size()) { return llvm::createStringError(llvm::inconvertibleErrorCode(), "Requesting type for result at index %zu, " "but lambda only generates %zu results", pos, outputGates.size()); } auto gate = outputGates[pos]; auto info = std::get<0>(gate); if (info.shape.size == 0) { return ResultType::SCALAR; } else { return ResultType::TENSOR; } } llvm::Expected JITLambda::Argument::getResultWidth(size_t pos) { if (pos >= outputGates.size()) { return llvm::createStringError(llvm::inconvertibleErrorCode(), "Requesting width for result at index %zu, " "but lambda only generates %zu results", pos, outputGates.size()); } auto gate = outputGates[pos]; auto info = std::get<0>(gate); // Encrypted values are always returned as 64-bit values for now if (info.encryption.hasValue()) return 64; else return info.shape.width; } llvm::Error JITLambda::Argument::getResult(size_t pos, void *res, size_t elementSize, size_t numElements) { auto gate = outputGates[pos]; auto info = std::get<0>(gate); auto offset = std::get<1>(gate); // Check is the argument is a scalar if (info.shape.dimensions.empty()) { return llvm::make_error( llvm::Twine("output is not a tensor, pos=").concat(llvm::Twine(pos)), llvm::inconvertibleErrorCode()); } // Check is the argument is a scalar if (info.shape.size != numElements) { return llvm::make_error( llvm::Twine("result #") .concat(llvm::Twine(pos)) .concat(" has not the expected size, got ") .concat(llvm::Twine(numElements)) .concat(" expect ") .concat(llvm::Twine(info.shape.size)), llvm::inconvertibleErrorCode()); } // Get the values as the memref calling convention expect. // aligned uint8_t *alignedBytes = static_cast(outputs[offset + 1]); uint8_t *resBytes = static_cast(res); if (!info.encryption.hasValue()) { // just copy values for (size_t i = 0; i < numElements; i++) { for (size_t j = 0; j < elementSize; j++) { *resBytes = *alignedBytes; resBytes++; alignedBytes++; } } } else { // decrypt and fill the result buffer for (size_t i = 0; i < numElements; i++) { LweCiphertext_u64 *ct = ((LweCiphertext_u64 **)alignedBytes)[i]; if (auto err = this->keySet.decrypt_lwe(pos, ct, ((uint64_t *)res)[i])) { return std::move(err); } } } return llvm::Error::success(); } } // namespace concretelang } // namespace mlir