ACL-OCL / Base_JSON /prefixS /json /signlang /2020.signlang-1.20.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "2020",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T15:22:19.528800Z"
},
"title": "Recognition of Static Features in Sign Language Using Key-Points",
"authors": [
{
"first": "Ioannis",
"middle": [],
"last": "Koulierakis",
"suffix": "",
"affiliation": {
"laboratory": "Intelligent Systems Laboratory",
"institution": "National Technical University of Athens",
"location": {}
},
"email": ""
},
{
"first": "Georgios",
"middle": [],
"last": "Siolas",
"suffix": "",
"affiliation": {
"laboratory": "Intelligent Systems Laboratory",
"institution": "National Technical University of Athens",
"location": {}
},
"email": "gsiolas@islab.ntua.gr"
},
{
"first": "Eleni",
"middle": [],
"last": "Efthimiou",
"suffix": "",
"affiliation": {
"laboratory": "Sign Language Technologies Team",
"institution": "ATHENA RC Zografou Campus",
"location": {
"addrLine": "9, Iroon Polytechniou str, Artemidos 6 & Epidavrou",
"postCode": "15780, 15125",
"settlement": "Zografou, Maroussi",
"country": "Greece, Greece"
}
},
"email": "eleni_e@athenarc.gr"
},
{
"first": "Stavroula-Evita",
"middle": [],
"last": "Fotinea",
"suffix": "",
"affiliation": {
"laboratory": "Sign Language Technologies Team",
"institution": "ATHENA RC Zografou Campus",
"location": {
"addrLine": "9, Iroon Polytechniou str, Artemidos 6 & Epidavrou",
"postCode": "15780, 15125",
"settlement": "Zografou, Maroussi",
"country": "Greece, Greece"
}
},
"email": ""
},
{
"first": "Andreas-Georgios",
"middle": [],
"last": "Stafylopatis",
"suffix": "",
"affiliation": {
"laboratory": "Intelligent Systems Laboratory",
"institution": "National Technical University of Athens",
"location": {}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In this paper we report on a research effort focusing on recognition of static features of sign formation in single sign videos. Three sequential models have been developed for handshape, palm orientation and location of sign formation respectively, which make use of key-points extracted via OpenPose software. The models have been applied to a Danish and a Greek Sign Language dataset, providing results around 96%. Moreover, during the reported research, a method has been developed for identifying the time-frame of real signing in the video, which allows to ignore transition frames during sign recognition processing.",
"pdf_parse": {
"paper_id": "2020",
"_pdf_hash": "",
"abstract": [
{
"text": "In this paper we report on a research effort focusing on recognition of static features of sign formation in single sign videos. Three sequential models have been developed for handshape, palm orientation and location of sign formation respectively, which make use of key-points extracted via OpenPose software. The models have been applied to a Danish and a Greek Sign Language dataset, providing results around 96%. Moreover, during the reported research, a method has been developed for identifying the time-frame of real signing in the video, which allows to ignore transition frames during sign recognition processing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "One of the problems relating to sign language recognition is the lack of appropriate datasets for algorithm training, since most datasets are recorded for academic purposes and as such, they concentrate in human learning rather than machine learning. Therefore, most data collections contain a very large number of different glosses with very few repetitions of each. This characteristic makes it very unlikely for these datasets to be used as training sets for classification algorithms in sign recognition level. Thus, we developed a system in the direction of \"phonological\" features recognition. This way we can extract a dataset with a lot of examples for every handshape, palm orientation and hand location out of the video collections.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "For the purposes of the project two collections of single gloss videos were used as datasets.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2."
},
{
"text": "The first one is \"Noema +\" which was developed by the Greek Institute for Language and Speech Processing (ILSP), Athena. It contains approximately 3000 lemmas of the Greek Language were signed by one native Greek signer and many of them are recorded two or three times. The total amount of videos is 3195 annotated with HamNoSys (Hanke, 2004) .",
"cite_spans": [
{
"start": 329,
"end": 342,
"text": "(Hanke, 2004)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2."
},
{
"text": "The second one is the \"Danish Sign Language Dictionary\". It was developed and edited at the Centre for Sign Language and Sign Supported Communication -KC in close cooperation with the Danish Deaf Association (DDL) Centre for Sign Language as a dictionary of the Danish Sign Language (DSL). The dictionary is consisted by single-sign videos as well as videos including short sentences in DSL. We used the single gloss videos which are 2714 in total, signed by several different signers. All these videos are annotated with a variation of HamNoSys that uses only one descriptor per instance (handshape, location, etc). For the description of the handshapes, 69 different names were used. Most of them are named after a letter of the Danish fingerspelling alphabet. Based on them all videos were annotated. This feature is making the whole process easier when trying to split the dictionary into handshape classes.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Datasets",
"sec_num": "2."
},
{
"text": "OpenPose is a software freely distributed by Carnegie Melon University, Perceptual Computing Lab (Cao et al., 2018) . It is used as a tool of human body keypoints extraction from a single image or video frame. It offers an estimation of 25 body/foot keypoints, 2x21 hand keypoints and 70 face keypoints. In the case of a 2D video input, for each keypoint it returns a vector containing 3 elements. The first 2 correspond to the (x,y) coordinates with reference to the upper left corner of the image. The third is a value in the range [0,1] which is quantification of the confidence given by the program that the specific keypoint is correctly located in the frame. The novelty behind OpenPose relies on the fact that it works for more than one person per image but more importantly the keypoint analysis is not affected when part of the individual's body is out of frame. This last feature is crucial for applications on sign language videos where the signer appears above the waist level (Figure 1 ).",
"cite_spans": [
{
"start": 97,
"end": 115,
"text": "(Cao et al., 2018)",
"ref_id": null
}
],
"ref_spans": [
{
"start": 989,
"end": 998,
"text": "(Figure 1",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "Openpose",
"sec_num": "3."
},
{
"text": "The first step in our method is transforming each video frame into keypoints using the OpenPose software. With this step we keep all the necessary information of the signer's posture and hand articulation, while we reduce the data dimensionality from 1280 x 720 pixels to 137 keypoints. Moreover, in our case the keypoints of the legs are redundant since none of the videos shows the bottom half of the signer. In addition, for each of the systems we will analyze below, we used a different number of keypoints related to the feature we are trying to classify in each case.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Method",
"sec_num": "4."
},
{
"text": "In general, the complete feature vector produced by OpenPose has the form:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Method",
"sec_num": "4."
},
{
"text": "= [ 0, 0, 0, 1, 1, 1, \u22ee 14, 14, 14, 0, 0, 0, 1, 1, 1, \u22ee 20, 20, 20, 0, \u210e 0, \u210e 0, \u210e 1, \u210e 1, \u210e 1, \u210e \u22ee 20, \u210e 20, \u210e 20, \u210e ]",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Method",
"sec_num": "4."
},
{
"text": "Where [ , , , , , ] is the i th keypoint of the \u03c4 th frame of the video. With the superscripts H, h we denote the keypoints of the dominant and non-dominant hand, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Our Method",
"sec_num": "4."
},
{
"text": "The problem of training a model on our data is a problem of semi-supervised learning. The reason is that in every video the annotation provides us with information on which are the static phonological features appearing in the video and the order in which they appear, but we lack a matching of the static features with individual frames. Moreover, we need a filtering of transitional frames that represent none of the annotated features. Those frames appear when a signer starts or stops signing moving his/her hands from or to resting pose, or during transition from sign to sign or from handshape to handshape into one sign. In all those cases the frames have no use in our training algorithm.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "In (Koller et al., 2016) this problem is solved by considering a \"junk\" state for those frames and using an Expectation Maximization algorithm for finding the most probable alignment between the frames and the annotation. On the other hand, we will use an alternative method to what was proposed by (Ko et al., 2018) . This method is relying on the work of (Choudhury et al., 2017) that categorizes the movement during signing in \"Movement Epenthesis\" and \"Signing\" based on the velocity of the centroid of the contour produced during the hand tracking stage. This method sets a velocity threshold and rejects every sequence of frames with greater velocity than the threshold.",
"cite_spans": [
{
"start": 3,
"end": 24,
"text": "(Koller et al., 2016)",
"ref_id": null
},
{
"start": 299,
"end": 316,
"text": "(Ko et al., 2018)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "In our method we are transforming each video frame into keypoints using the OpenPose software. This help us skip the hand recognition and tracking process while maintaining the maximal accuracy provided by OpenPose. In addition, we have the opportunity to calculate the velocity of the hand based on more than one point of interest. We calculate the total hand velocity as the sum of the velocities of each keypoint between two frames on the basis of the following equation:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "= \u2211 \u221a ( , \u2212 , + ) + ( , \u2212 , + ) =",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "We present our methodology based on the handshape recognition. Although, the method is outright extendable to the other two characteristics. We modify the method for rejecting redundant frames based on velocity threshold and extend it by adding one more rule. Every frame is removed from dataset unless it satisfies the following 3 rules:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "\u2022 Belong in a sequence of 5 frames with total velocity below a threshold Tv. \u2022 The logarithmic sum of certainty \u03c3i of all points is over a threshold T\u03c3. \u2022 Wrist is over the waist level (not a hand resting posture) Both video datasets were recorded at 25 frames/second and so the 5 frame sequence corresponds to 0.2 seconds. We remove the third column from the feature vector X\u03c4 and we use the provided information in the second rule in order to remove bad quality data from our dataset before training. The third rule was added to remove frames from the start and end of each video where the signer is crossing his/her hands on waist level. These frames do not involve signing but they pass the first two rules due to very low movement and clarity. Our final step is to match each frame remaining to the matching handshape. The advantage here is that the maximum number of different handshapes appearing in every video is 2 due to the fact that we have single gloss videos. If according to the HamNoSys annotation, only one handshape appears in the video, we know that the frames are representing that specific handshape. Otherwise, when two handshapes appear in the video, we are clustering the frames into two clusters using Gaussian Mixture Model (GMM) (Koller et al., 2016) (Theodorakis et al., 2014) (Pitsikalis et al. 2011). The first handshape is matched to the cluster the elements of which appear earlier by mean in the video and the second is matched to the other one.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "For example, in Figure 2 , there is an example of 8 frames extracted from a video and transformed into keypoints. Beforehand we know that there are two different handshapes. During the segmentation, the last frame is rejected because the signer has crossed his hands and so the right palm is below the elbow level. Otherwise this frame would be labeled as or as adding a false element into the dataset. The GMM algorithm splits the frames into clusters A,B. Cluster-A appears earlier in the sequence so we label the elements of cluster-A, cluster-B as , , respectively.",
"cite_spans": [],
"ref_spans": [
{
"start": 16,
"end": 24,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Segmentation",
"sec_num": "4.1."
},
{
"text": "At this point we have created a collection of frames representing each possible handshape. In our final dataset, 66 different handshapes, 12 different palm orientations, 33 different locations have found after the segmentations. Approximately 60000 frames ended in the final segmentation for each of the three static features. For the handshape training we used the keypoints extracted only from the dominant hand. We changed the reference system fixing the centroid of the 21 keypoints to (0,0) and the feature vector was normalised using the mean and standard deviation for each dimension, respectively. Moreover, the same process was used for creating the palm-orientation and location feature vectors. For the palm orientation we used all the upper body keypoints plus the dominant hand. Finally, for the location feature vector we included every keypoint including the non-dominant. The reason for this extension of the feature vectors was due to the fact that the orientation of the body of the signer is not the same for all videos and the hand orientation has to be recognized relatively to the body. Obviously, all keypoints are necessary for palm location due to the fact that every articulation is described relatively to a body part including the non-dominant hand and the face. For the two models we use Multi-Layer Perceptrons (MLPs) since we classify each frame independently. 5 hidden layers models with 128 neurons in each layer and softmax activation function are used both for handshape and orientation classifiers. In total we train our models for approximately 400 epochs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": "4.2."
},
{
"text": "For the training we isolated the 25% of all the videos as a test set. In this way the models are tested in classifying features from signs they have not come up with or even signers for DSL dataset that the systems are not trained on. In Table 1 we can see that the final accuracies of every model is over 95%. According to Figures 3,4 ,5 the models can almost perfectly classify the training set. We should, also, point out that many of the errors in our classification method could be related to errors during the segmentation. ",
"cite_spans": [],
"ref_spans": [
{
"start": 238,
"end": 245,
"text": "Table 1",
"ref_id": null
},
{
"start": 324,
"end": 335,
"text": "Figures 3,4",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Training",
"sec_num": "4.2."
}
],
"back_matter": [
{
"text": "Research efforts relating to recognition of static features of sign formation including the handshape, the palm orientation and the location of signs by means of sequential models, have provided encouraging results as shown in 4.3 above. Such results may prove especially helpful towards (semi-)automatic annotation of SL videos. Furthermore, embedding of the three models handling handshape, palm orientation and location of sign in recurrent neural networks is expected to pave the way towards continuous SL recognition.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "5."
},
{
"text": "The authors acknowledge support of this work by the project \"Computational Science and Technologies: Data, Content and Interaction\" (MIS 5002437), which is implemented under the Action \"Reinforcement of the Research and Innovation Infrastructure\", funded by the Operational Programme \"Competitiveness, Entrepreneurship and Innovation\" (NSRF 2014-2020) and co-financed by Greece and the European Union (European Regional Development Fund). Moreover, they wish to thank Center for Tegnsprog (2008) (2009) (2010) (2011) (2012) (2013) (2014) (2015) (2016) (2017) (2018) ",
"cite_spans": [
{
"start": 489,
"end": 495,
"text": "(2008)",
"ref_id": null
},
{
"start": 496,
"end": 502,
"text": "(2009)",
"ref_id": null
},
{
"start": 503,
"end": 509,
"text": "(2010)",
"ref_id": null
},
{
"start": 510,
"end": 516,
"text": "(2011)",
"ref_id": null
},
{
"start": 517,
"end": 523,
"text": "(2012)",
"ref_id": null
},
{
"start": 524,
"end": 530,
"text": "(2013)",
"ref_id": null
},
{
"start": 531,
"end": 537,
"text": "(2014)",
"ref_id": null
},
{
"start": 538,
"end": 544,
"text": "(2015)",
"ref_id": null
},
{
"start": 545,
"end": 551,
"text": "(2016)",
"ref_id": null
},
{
"start": 552,
"end": 558,
"text": "(2017)",
"ref_id": null
},
{
"start": 559,
"end": 565,
"text": "(2018)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgements",
"sec_num": "6."
}
],
"bib_entries": {},
"ref_entries": {
"FIGREF0": {
"type_str": "figure",
"uris": null,
"text": "Example of OpenPose",
"num": null
},
"FIGREF1": {
"type_str": "figure",
"uris": null,
"text": "Handshape recognition modelFigure 2: Example of Segmentation process",
"num": null
}
}
}
}