Training scripts #15
Replies: 14 comments 9 replies
-
Mobilenet using Adamw + SGDW training on Emore dataset
|
Beta Was this translation helpful? Give feedback.
-
Ghostnet using SGD + L2 regularizer + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
Ghostnet strides=1 float16, using SGD + L2 regularizer + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
Botnet50 using SGD + L2 regularizer + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
Resnet50V2 / Resnet101V2 swish using SGD + L2 regularizer + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
r50 swish using SGD + L2 regularizer + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
se_r50 swish SD (1, 0.8) using SGD + L2 regularizer + cosine lr decay + randaug training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
Converted official r18 / r34 / r50 / r100 model on glint360k trained by partial fc + cosface
|
Beta Was this translation helpful? Give feedback.
-
EffcientNetV2S swish drop_conn 0.2 dropout 0.2 using SGD + L2 regularizer + cosine lr decay + randaug training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
EffcientNetV2S swish drop_conn 0.2 dropout 0.2 using AdamW + cosine lr decay + randaug training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
EffcientNetV2S swish drop_conn 0.2 dropout 0.2 using AdamW + cosine lr decay + randaug + Magface training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
r50/r100 PReLU dropout 0.4 using SGD + l2 regularizer + randaug + AdaFace training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
r100 PReLU dropout 0.4 using SGD + l2 regularizer + randaug + AdaFace training on Glint360K dataset and partial FC
|
Beta Was this translation helpful? Give feedback.
-
SE_MobileFaceNet using SGD + cosine lr decay training on MS1MV3 dataset
|
Beta Was this translation helpful? Give feedback.
-
ResNet101V2 using nadam and finetuning with triplet
nadam
optimizer.BatchHardTripletLoss
only, which may reach better score on these eval datasets, but may harm the margin distance between different classes.bottleneckOnly
may be not a necessary.Beta Was this translation helpful? Give feedback.
All reactions