Giter Club home page Giter Club logo

tensorflow-sphereface-asoftmax's People

Contributors

pppoe avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

tensorflow-sphereface-asoftmax's Issues

Additive Margin Softmax tensorflow

Hi author.
There is a new published paper proposed additive margin softmax(AM-softmax), which seems easier to train than sphereface.
https://arxiv.org/abs/1801.05599
I implemented main part of this paper here by tensorflow:
https://github.com/Joker316701882/Additive-Margin-Softmax
With exact hyper-parameters with author, I can also only achieve 98.x% accu. So I'm wondering is it the problem of tensorflow low-level implementation (like optimizer) different from caffe so that with same parameters, it's hard to reach exact performance. Have you ever tried loss in paper like AM-softmax(cosface), arcface etc.

Will be glad to know your idea about this.

error after changing the Network from TFlearn code to TF slim code

hi, I change the Network from TFlearn code to TF slim code, the tf-slim code as follow
`def prelu(_x, scope=None):
"""parametric ReLU activation"""
with tf.variable_scope(name_or_scope=scope, default_name="prelu"):
_alpha = tf.get_variable("prelu", shape=_x.get_shape()[-1],
dtype=_x.dtype, initializer=tf.constant_initializer(0.1))
return tf.maximum(0.0, _x) + _alpha * tf.minimum(0.0, _x)

def Network(data_input, training = True):
x = slim.conv2d(data_input, 32, [3,3], stride = 1, activation_fn=prelu)
x = slim.conv2d(x, 32, [3, 3], stride = 2, activation_fn=prelu)
x = slim.conv2d(x, 64, [3, 3], activation_fn=prelu)
x = slim.conv2d(x, 64, [3, 3], stride = 2, activation_fn=prelu)
x = slim.conv2d(x, 128, [3, 3], activation_fn=prelu)
x = slim.conv2d(x, 128, [3, 3], stride = 2, activation_fn=prelu)
x = slim.flatten(x)
feat = slim.fully_connected(x, 2)
return feat
`
when I train the network, the loss does not decline, how can i solve this problem?Thanks

0 2.3063145
100 2.3025854
200 2.3025854
300 2.3025854
400 2.3025854
500 2.3025854
600 2.3025854


sloved! just change "feat = slim.fully_connected(x, 2)" to "feat = slim.fully_connected(x, 2,activation_fn=prelu)"

Could you share your training parameters?

I tried to reproduce sphereface via tensorflow on CASIA-WebFace, but I found the loss function is very difficult to convey, and loss function sometimes be very large (10e6). Are there any solutions?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.