Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Face attributes Model] Need visualization of face attrib instead of dumping json #137

Open
sathishram1912 opened this issue Dec 4, 2024 · 3 comments
Labels
question Please ask any questions on Slack. This issue will be closed once responded to.

Comments

@sathishram1912
Copy link

sathishram1912 commented Dec 4, 2024

I tried face attribnet model which saves json file for each face detection.
Then i noticed demo.py which didnt do any post processing rather its dump whatever outputs model has, it simply dumps to json
could you provide solution to process the model output

@mestrona-3
Copy link

We have a collection of output for this model including face recognition related embedding vector(length of 512), liveness related embedding vector(length of 32), and probability of detecting eye closeness, (sun)glasses, mask. Users can check specific result given the key values in the json file.

@sathishram1912
Copy link
Author

sathishram1912 commented Dec 7, 2024

Eye Closeness, Sunglasses, and Mask Probabilities:

  1. Are the probabilities for detecting eye closeness, sunglasses, and masks raw logits or already in the range of [0, 1]?
  2. If they are raw logits, should we apply the sigmoid function to map them to probabilities?
  3. For eye closeness specifically, the output has two values. Does this correspond to the probabilities for the left and right eyes, respectively?

Liveness Embedding Vector:

  1. The liveness embedding vector has a length of 32. How should we interpret or use these embeddings in downstream tasks
  2. In the demo, the liveness result is shown as a boolean (true/false). How is this boolean value derived from the 32-dimensional embedding vector? Are there specific thresholds or logic applied?

@mestrona-3
Copy link

For eye closeness, glasses, and masks, we output raw logits. You need to apply softmax to get the final probabilities.
However, for sunglasses, the output is directly in probabilities.

For question 3: Yes, the two values in eye closeness correspond to the left and right eyes separately.
Regarding the liveness embedding vector, they are subject-specific. This means that to determine the liveness of a specific person, you need to obtain the embedding of the real face photo of that person first. Then, compare the later embedding with this one using L2 or cosine similarity. Additionally, you may train a classifier on your own using these embeddings.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Please ask any questions on Slack. This issue will be closed once responded to.
Projects
None yet
Development

No branches or pull requests

2 participants