Files
inji-wallet/components/FaceScanner/FaceScannerHelper.ts
adityankannan-tw 339e08c462 [INJIMOB-1433,528] - Add passive liveness detection with blink detection (#1474)
* [INJIMOB-528] add liveness support for face verification

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-528] add and comment blink detection

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-528] update locales and remove blink detection

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] add blinking and increase threshold if blinking is detected

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] sync package lock json

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] update node version to 18 for android build

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] refactor

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] refactor components

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] use the default version of package resolved file

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] refactor and add new env for liveness in workflow

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] remove new env and unused code

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-1433] add new env for liveness and combine build descriptiona and build name

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

* [INJIMOB-528] update package lock & pbxproj files

Signed-off-by: KiruthikaJeyashankar <81218987+KiruthikaJeyashankar@users.noreply.github.com>

* [INJIMOB-1433] add test id for elements

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>

---------

Signed-off-by: adityankannan-tw <adityan410pm@gmail.com>
Signed-off-by: KiruthikaJeyashankar <81218987+KiruthikaJeyashankar@users.noreply.github.com>
Co-authored-by: adityankannan-tw <adityan410pm@gmail.com>
Co-authored-by: KiruthikaJeyashankar <81218987+KiruthikaJeyashankar@users.noreply.github.com>
2024-06-04 13:59:02 +05:30

264 lines
7.5 KiB
TypeScript

import hexRgb, {RgbaObject} from 'hex-rgb';
import {LIVENESS_THRESHOLD, isAndroid} from '../../shared/constants';
import {closest} from 'color-diff';
import * as FaceDetector from 'expo-face-detector';
import ImageEditor from '@react-native-community/image-editor';
import {ImageType} from 'expo-camera';
import {getColors} from 'react-native-image-colors';
import {faceCompare} from '@iriscan/biometric-sdk-react-native';
let FaceCropPicArray: any[] = new Array();
let EyeCropPicArray: any[] = new Array();
let predictedColorResults: any[] = new Array();
let facePoints;
let calculatedThreshold;
let faceCompareOuptut;
let capturedFaceImage;
let leftEyeWasClosed = false;
let rightEyeWasClosed = false;
let lastBlinkTimestamp = 0;
let blinkCounter = 0;
const offsetX = 200;
const offsetY = 350;
const captureInterval = 650;
const eyeOpenProbability = 0.85;
const blinkConfidenceScore = 0.1;
const blinkThreshold = 0.4;
const blinkTimeInterval = 900;
const eyeCropHeightConst = 50;
const XAndYBoundsMax = 280;
const XAndYBoundsMin = 300;
const rollAngleThreshold = 10;
const yawAngleThreshold = 3;
const colorFiltered = ['background', 'dominant'];
const rxDataURI = /data:(?<mime>[\w/\-.]+);(?<encoding>\w+),(?<data>.*)/;
const colorComparePalette = [
{R: 255, G: 0, B: 0},
{R: 0, G: 255, B: 0},
{R: 0, G: 0, B: 255},
];
export const imageCaptureConfig = {
base64: true,
quality: 1,
imageType: ImageType.jpg,
};
export const faceDetectorConfig : FaceDetectorConfig= {
mode: FaceDetector.FaceDetectorMode.accurate,
detectLandmarks: FaceDetector.FaceDetectorLandmarks.all,
runClassifications: FaceDetector.FaceDetectorClassifications.all,
contourMode: FaceDetector.FaceDetectorClassifications.all,
minDetectionInterval: captureInterval,
tracking: true,
};
export const checkBlink = face => {
const leftEyeOpenProbability = face.leftEyeOpenProbability;
const rightEyeOpenProbability = face.rightEyeOpenProbability;
const currentTime = new Date().getTime();
const leftEyeClosed = leftEyeOpenProbability < blinkThreshold;
const rightEyeClosed = rightEyeOpenProbability < blinkThreshold;
if (leftEyeClosed && rightEyeClosed) {
leftEyeWasClosed = true;
rightEyeWasClosed = true;
}
if (
leftEyeWasClosed &&
rightEyeWasClosed &&
!leftEyeClosed &&
!rightEyeClosed
) {
if (
lastBlinkTimestamp === 0 ||
currentTime - lastBlinkTimestamp > blinkTimeInterval
) {
blinkCounter = blinkCounter + 1;
lastBlinkTimestamp = currentTime;
}
leftEyeWasClosed = false;
rightEyeWasClosed = false;
}
};
export const getFaceBounds = face => {
const {bounds, yawAngle, rollAngle} = face;
const withinXBounds =
bounds.origin.x + bounds.size.width >= XAndYBoundsMax &&
bounds.origin.x <= XAndYBoundsMin;
const withinYBounds =
bounds.origin.y + bounds.size.height >= XAndYBoundsMax &&
bounds.origin.y <= XAndYBoundsMin;
const withinYawAngle =
yawAngle > -yawAngleThreshold && yawAngle < yawAngleThreshold;
const withinRollAngle =
rollAngle > -rollAngleThreshold && rollAngle < rollAngleThreshold;
return [withinXBounds, withinYBounds, withinYawAngle, withinRollAngle];
};
export const getNormalizedFacePoints = (facePoints: any): number[] => {
return isAndroid()
? [
facePoints.LEFT_EYE.x,
facePoints.LEFT_EYE.y,
facePoints.RIGHT_EYE.x,
facePoints.RIGHT_EYE.y,
]
: [
facePoints.leftEyePosition.x,
facePoints.leftEyePosition.y,
facePoints.rightEyePosition.x,
facePoints.rightEyePosition.y,
];
};
export const filterColor = color => {
return (
typeof color === 'string' &&
color.startsWith('#') &&
!colorFiltered.includes(color)
);
};
export const getEyeColorPredictionResult = async (
LeftrgbaColors: RgbaObject[],
color: RgbaObject,
) => {
LeftrgbaColors.forEach(colorRGBA => {
let colorRGB = {};
colorRGB['R'] = colorRGBA.red;
colorRGB['G'] = colorRGBA.green;
colorRGB['B'] = colorRGBA.blue;
const closestColor = closest(colorRGB, colorComparePalette);
const result =
color.red === closestColor.R &&
color.blue === closestColor.B &&
color.green === closestColor.G;
predictedColorResults.push(result);
});
};
export const cropEyeAreaFromFace = async (picArray, vcImage, capturedImage) => {
try {
await Promise.all(
picArray.map(async pic => {
facePoints = (
await FaceDetector.detectFacesAsync(pic.image.uri, faceDetectorConfig)
).faces[0];
if (
facePoints.leftEyeOpenProbability > eyeOpenProbability &&
facePoints.rightEyeOpenProbability > eyeOpenProbability
) {
capturedFaceImage = await ImageEditor.cropImage(pic.image.uri, {
offset: {
x: facePoints.bounds.origin.x,
y: facePoints.bounds.origin.y,
},
size: {
width: facePoints.bounds.size.width,
height: facePoints.bounds.size.height,
},
});
FaceCropPicArray.push({color: pic.color, image: capturedFaceImage});
}
}),
);
await Promise.all(
FaceCropPicArray.map(async pic => {
let [leftEyeX, leftEyeY, rightEyeX, rightEyeY] =
getNormalizedFacePoints(facePoints);
const leftCroppedImage = await ImageEditor.cropImage(pic.image.uri, {
offset: {
x: leftEyeX - offsetX,
y: leftEyeY - offsetY,
},
size: {
width: offsetX * 2,
height: offsetY / 2 - eyeCropHeightConst,
},
});
const rightCroppedImage = await ImageEditor.cropImage(pic.image.uri, {
offset: {
x: rightEyeX - offsetX,
y: rightEyeY - offsetY,
},
size: {
width: offsetX * 2,
height: offsetY / 2 - eyeCropHeightConst,
},
});
EyeCropPicArray.push({
color: pic.color,
leftEye: leftCroppedImage,
rightEye: rightCroppedImage,
});
}),
);
await Promise.all(
EyeCropPicArray.map(async pic => {
const leftEyeColors = await getColors(pic.leftEye.uri);
const rightEyeColors = await getColors(pic.rightEye.uri);
const leftRGBAColors = Object.values(leftEyeColors)
.filter(filterColor)
.map(color => hexRgb(color));
const rightRGBAColors = Object.values(rightEyeColors)
.filter(filterColor)
.map(color => hexRgb(color));
const rgbColor = hexRgb(pic.color);
await getEyeColorPredictionResult(leftRGBAColors, rgbColor);
await getEyeColorPredictionResult(rightRGBAColors, rgbColor);
}),
);
} catch (err) {
console.error('Unable to crop the images::', err);
return false;
}
calculatedThreshold =
predictedColorResults.filter(element => element).length /
predictedColorResults.length;
const matches = rxDataURI.exec(vcImage).groups;
const vcFace = matches.data;
faceCompareOuptut = await faceCompare(vcFace, capturedImage.base64);
if (blinkCounter > 0) {
calculatedThreshold = calculatedThreshold + blinkConfidenceScore;
}
if (calculatedThreshold > LIVENESS_THRESHOLD && faceCompareOuptut) {
return true;
} else {
return false;
}
};
export interface FaceDetectorConfig {
mode: FaceDetector.FaceDetectorMode;
detectLandmarks: FaceDetector.FaceDetectorLandmarks;
runClassifications: FaceDetector.FaceDetectorClassifications;
contourMode: FaceDetector.FaceDetectorClassifications;
minDetectionInterval: number;
tracking: boolean;
};