Larger resolution face masked, weirdly warped, deepfake, requires https://github.com/keras-team/keras-contrib.git and https://github.com/1adrianb/face-alignment.git to be cloned into the root of the repo and A and B's alignments.json to be copied into the correct /data/* folder before training.
Inputs are 64x64 images outputs are a pair of 128x128 images one RGB with the reconstructed face, one B/W to be used as a mask to guide what sections of the image are to be replaced.
For the reconstrcuted face masked DSSIM loss is used that behaves as a standard SSIM difference measure in the central face area and always returns zero loss in the surrounding background area outside of the face so as as not to train irrelevant features.
MSE is used for the mask.
Directory structure for training data should look like (image names for example purposes only):
df │ └───data │ ├───A │ │ alignments.json │ │ trainingimageA1.jpg │ │ trainingimageA2.jpg │ ├───B │ │ alignments.json │ │ trainingimageB1.jpg │ │ trainingimageB2.jpg
or as a file listing:
.\df\data\A\alignments.json .\df\data\A\trainingimageA1.jpg .\df\data\A\trainingimageA2.jpg .\df\data\B\alignments.json .\df\data\B\trainingimageB1.jpg .\df\data\B\trainingimageB2.jpg
Image gaps scanner can be run with this command to scan both the A and B folders:
imageGapsScanner.py data\A data\B
it will pop open a new window showing two rows, the images in A that have the worst matches in B and the images in B that have the worst matches in A: In that example the top row has some clear images that are totally different to the data in B, ideally these would be hunted down, you'd try to find some images of cage shouting at various angles, of the examples in the bottom row however only the first seems to be of a concern, you would consider deleting that image or finding a matching trump image for it, for the remaining images in the bottom row, although they are the top 9 "worst" matches they seem to be relatively nomal front-on shots and can be ignored.