Submission declined on 5 January 2024 by
KylieTastic (
talk).
Where to get help
How to improve a draft
You can also browse Wikipedia:Featured articles and Wikipedia:Good articles to find examples of Wikipedia's best writing on topics similar to your proposed article. Improving your odds of a speedy review To improve your odds of a faster review, tag your draft with relevant WikiProject tags using the button below. This will let reviewers know a new draft has been submitted in their area of interest. For instance, if you wrote about a female astronomer, you would want to add the Biography, Astronomy, and Women scientists tags. Editor resources
| ![]() |
Submission declined on 21 December 2023 by
Theroadislong (
talk). This draft's references do not show that the subject
qualifies for a Wikipedia article. In summary, the draft needs multiple published sources that are: Declined by
Theroadislong 5 months ago.
| ![]() |
Submission declined on 18 December 2023 by
Rich Smith (
talk). This submission appears to be taken from
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10540961/.
Wikipedia cannot accept material copied from elsewhere, unless it explicitly and verifiably has been released to the world under a
suitably free and compatible copyright license or into the
public domain and is written in an
acceptable tone—this includes material that you own the copyright to. You should attribute the content of a draft to outside sources, using
citations, but
copying and pasting or
closely paraphrasing sources is not acceptable. The entire draft should be written using your own words and structure. Declined by
Rich Smith 5 months ago.
| ![]() |
UIBVFED [1] is the first database made up of synthetic avatars that categorizes up to 32 facial expressions. The dataset is composed of 660 facial images (1080 x 1920) from 20 virtual characters each creating 32 facial expressions. The avatars represent 10 men and 10 women, aged between 20 and 80, from different ethnicities. Expressions are classified based on the six universal emotions (Anger, Disgust, Fear, Joy, Sadness, and Surprise) according to Faigin’s classification [2] whose reference is considered the standard to follow by animators and 3D artists. In addition to Faigin’s expression classification, the database provides the equivalence of the Facial Action Coding System (FACS) [3] with information about the position of the 51 facial landmarks in the 3D space to facilitate expression recognition. This is because the images of the facial expressions in the dataset were generated following the FAC guidelines. Therefore, the deformations that were applied to the 3D models have a direct correspondence with the Action Units (AUs) that are associated with each expression. This procedure ensures an objective labelling of all the images. Information about the landmarks for all characters and expressions is included in the dataset.
Figure 1 shows the 32 expressions plus the neutral one of one of the 20 characters in the UIBVFED database and their associated emotion.
The dataset is provided together with an interactive application, the UIBVFED application GUI, that allows the users to activate and control the expression intensity of the characters from different points of view.
-
in-depth (not just passing mentions about the subject)
-
reliable
-
secondary
-
independent of the subject
Make sure you add references that meet these criteria before resubmitting. Learn about mistakes to avoid when addressing this issue. If no additional references exist, the subject is not suitable for Wikipedia.