Yeah; I found that if there was a lot of other objects in view (laptop/phone/notebook/pens/etc) then it had a hard time (I'm not great at signing anyways, ha
It may have performed better if I had done some image segmentation processing before running inference, then it may have performed even better
You mean well, but it looks like David Lee isn't fluent either. Have y'all included Deaf folks in this? I'm not deaf but I learned from folks who are... and well-meaning hearing folks was a consistent theme in learning about Deaf culture.
The dataset is questionable, and I note that his classifier has problems recognizing the g and h handshapes because...
(quoth wikipedia) > In most drawings or illustrations of the American Manual Alphabet, some of the letters are depicted from the side to better illustrate the desired hand shape. For example, the letters G and H are frequently shown from the side to illustrate the position of the fingers. However, they are signed with the hand in an ergonomically neutral position, palm facing to the side and fingers pointing forward.
David's example shows him holding his wrist in a position that would quickly cause RSI (but his son gets it right). Consider: it takes me under a second to spell "gerontology", and I lost fluency years ago.
This chart depicts those handshapes more accurately:
Yeah; this was definitely just a very quick project to demonstrate what was possible. If I was trying to build an actual, production-level/commercial project I would work with a Deaf person with this to develop a more comprehensive dataset.
Fun (possibly local dialect) tidbit: the zz in pizza can be spelled in one go; similar to the V handshape but following the Z motion. If somebody is gonna take this to production level, it had better include more than one Deaf person (and, just like facial recognition... folks of diverse skin tone, gender expression and BMI)