Skip to content

Latest commit

 

History

History
96 lines (52 loc) · 3.65 KB

user_faq.md

File metadata and controls

96 lines (52 loc) · 3.65 KB

How to increase the quality of the picture?

Play with different program settings, try different lighting in the room, a different camera angle.

Why doesn't the replaced face look like a celebrity?

It depends on how much your face fits the shape of the celebrity's face.

Why is the swapped face blurrier than the rest of the image?

It depends on how big the face in the frame, as well as the resolution of the model. The public models have a resolution of 224x224 to work on most PC configurations. For high face resolution, you need to train your own model. Public models are suitable for game streamers that have their face in a small window in the corner of the screen.

How to increase the performance of face replacement?

Play with different program settings. Any module put on the CPU will consume a lot of CPU time, which is not enough to run a game, for example. If the motherboard allows, you can install additional video cards and distribute the load on them.

Will DeepFake detectors be able to detect a fake in my streams?

depends on final quality of your picture. Flickering face, abruptly clipping face mask, irregular color will increase chance to detect the fake.

deepware.ai detects no fakes in the example video of the fake Margot Robbie.

Do I need to train the model from scratch?

No you don't. There are public face models that can swap any face without training.

When will AMD cards be supported?

Depends on microsoft/onnxruntime library developers. AMD cards are supported through DirectML execution provider, which is currently raw and slow.

I want to have more control when changing faces in a video. Will the new functionality be implemented?

No. DeepFaceLive is designed for face swapping in streams. The ability to change faces in the videos - only for test purposes.

I want to swap my face to a particular celebrity. What I need to do?

This is the normal deepfake training process, where src faceset is the celebrity and dst faceset is your faceset. Train the deepfake model as usual with DeepFaceLab, then export the model in .dfm format for use in DeepFaceLive. You can also try ordering a deepfake model from someone in Discord or forum.

I want to train ready-to-use face model to swap any face to celebrity. What I need to do?

If you are familiar with DeepFaceLab, then this tutorial will help you:

Src faceset is celebrity. Must be diverse enough in yaw, light and shadow conditions. Src faceset should be xseg'ed and applied. You can apply Generic XSeg to src faceset.

Dst faceset is RTM WF faceset from the torrent.

Make a backup before every stage !

Using SAEHD model.

res:224, WF, ae_dims:256, e_dims:64, d_dims:64, d_mask_dims 22, eyes_mouth_prio:Y, batch more is better. Others are default.

Assuming 1kk iters with batch 8. If the batch is higher, iters is possible less.

  1. enable pretrain mode. Train to 1kk
  2. disable pretrain mode. Train to 1kk
  3. lrd:N uniform_yaw:True, color_transfer:lct, train +500..800k
  4. lrd:Y train +500..800k iters
  5. disable random warp +500..800k iters
  6. enable gan 0.1 gan_dims:32, train +100..300k iters

You can reuse this model to train new src faceset. In this case you must delete the file inter_AB.npy from the model folder, and train from stage 3.

Using AMP model.

-- coming soon --