ming121419
发表于 2024-1-9 07:38:55
大佬你好,RTM解压后faceset.pak是不是放进_internal里的pretrain_faces里?RTT解压后是不是把model里的文件放进_internal里的model_generic_xseg?希望大佬解答,我是新手。
cic
发表于 2024-1-25 22:34:47
谢谢楼主分享
爱玛
发表于 2024-1-25 23:46:54
谢谢分享,头像pak怎么解压使用?
爱玛
发表于 2024-1-26 10:11:31
我使用了报错是什么问题呢?我换了一个其他的模型就没问题?
Error: 2 root error(s) found
爱玛
发表于 2024-1-26 11:24:10
爱玛 发表于 2024-1-25 23:46
谢谢分享,头像pak怎么解压使用?
我已经学会了这个
luodao11
发表于 2024-1-26 14:26:28
感谢分享。
爱玛
发表于 2024-1-28 14:43:42
================== Model Summary ===================
== ==
== Model name: new_SAEHD ==
== ==
== Current iteration: 1 ==
== ==
==---------------- Model Options -----------------==
== ==
== resolution: 224 ==
== face_type: wf ==
== models_opt_on_gpu: True ==
== archi: liae-udt ==
== ae_dims: 512 ==
== e_dims: 64 ==
== d_dims: 64 ==
== d_mask_dims: 32 ==
== masked_training: True ==
== eyes_mouth_prio: False ==
== uniform_yaw: True ==
== blur_out_mask: True ==
== adabelief: True ==
== lr_dropout: y ==
== random_warp: False ==
== random_hsv_power: 0.0 ==
== true_face_power: 0.0 ==
== face_style_power: 0.0 ==
== bg_style_power: 0.0 ==
== ct_mode: none ==
== clipgrad: False ==
== pretrain: False ==
== autobackup_hour: 0 ==
== write_preview_history: False ==
== target_iter: 0 ==
== random_src_flip: False ==
== random_dst_flip: True ==
== batch_size: 8 ==
== gan_power: 0.1 ==
== gan_patch_size: 28 ==
== gan_dims: 32 ==
== ==
==------------------ Running On ------------------==
== ==
== Device index: 0 ==
== Name: NVIDIA GeForce RTX 4060 ==
== VRAM: 5.33GB ==
== ==
====================================================
==入门教程:https://www.deepfaker.xyz ==
==进阶学习:https://t.zsxq.com/vFuzRNr ==
====================================================
|保存时间 |迭代次数 |单次时间 |源损失 |目标损失
Error: 2 root error(s) found.
(0) Resource exhausted: OOM when allocating tensor with shape and type float on /job:localhost/replica:0/task:0/device:GPU:0 by allocator GPU_0_bfc
[]
Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info. This isn't available when running in Eager mode.
[]
Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info. This isn't available when running in Eager mode.
(1) Resource exhausted: OOM when allocating tensor with shape and type float on /job:localhost/replica:0/task:0/device:GPU:0 by allocator GPU_0_bfc
[]
Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info. This isn't available when running in Eager mode.
0 successful operations.
0 derived errors ignored.
Errors may have originated from an input operation.
Input Source operations connected to node Pad_6:
LeakyRelu_1 (defined at D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\archis\DeepFakeArchi.py:29)
Input Source operations connected to node Pad_6:
LeakyRelu_1 (defined at D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\archis\DeepFakeArchi.py:29)
Original stack trace for 'Pad_6':
File "threading.py", line 884, in _bootstrap
File "threading.py", line 916, in _bootstrap_inner
File "threading.py", line 864, in run
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\models\ModelBase.py", line 193, in __init__
self.on_initialize()
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\models\Model_SAEHD\Model.py", line 415, in on_initialize
gpu_src_code = self.encoder (gpu_warped_src)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\models\ModelBase.py", line 117, in __call__
return self.forward(*args, **kwargs)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\archis\DeepFakeArchi.py", line 112, in forward
x = self.res1(x)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\models\ModelBase.py", line 117, in __call__
return self.forward(*args, **kwargs)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\archis\DeepFakeArchi.py", line 84, in forward
x = self.conv2(x)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\layers\LayerBase.py", line 14, in __call__
return self.forward(*args, **kwargs)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\DeepFaceLab\core\leras\layers\Conv2D.py", line 87, in forward
x = tf.pad (x, padding, mode='CONSTANT')
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\util\dispatch.py", line 206, in wrapper
return target(*args, **kwargs)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\ops\array_ops.py", line 3528, in pad
result = gen_array_ops.pad(tensor, paddings, name=name)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\ops\gen_array_ops.py", line 6487, in pad
"Pad", input=input, paddings=paddings, name=name)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\framework\op_def_library.py", line 750, in _apply_op_helper
attrs=attr_protos, op_def=op_def)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\framework\ops.py", line 3569, in _create_op_internal
op_def=op_def)
File "D:\DeepFaceLab_NV30_221111_CN\DeepFaceLab_NV30_221111_CN\_internal\python-3.6.8\lib\site-packages\tensorflow\python\framework\ops.py", line 2045, in __init__
self._traceback = tf_stack.extract_stack_for_node(self._c_op)
==================================================================================
请教下,我用的这个模型打开就报错,这是什么问题?其他的模型没事,就是这个不行,我把参数和我的其他模型参数设置一样的也不行
boy147
发表于 2024-1-30 20:56:20
感谢大佬分享!
mybestlove
发表于 2024-1-31 15:17:07
感谢楼主分享
leonarrrd
发表于 2024-2-1 06:37:26
谢老板分享 太棒了