Making Facial Gestures or, Using CIDetector with Streaming Video
Since iOS7 Apple introduced new feature to the CIDetector, specifically smile and blinks detection.
I’ve wondered if i could “feed” it streaming video instead of static image and get make something like a live smile facial gesture be registered as an action on the device.
Here is a demo video of it:
I’ve extracted all the logic into an easy to use facade so it can easily be added in any project. Have a look at the The main issues where:
CIDetector is a slow and takes up a lot of CPU
Smartley aggregating the feature detection results from the CIDetector until enough features are considered a registered gestured.
The code and example project: https://github.com/dannyshmueli/DSFacialGestureDetector