CIDetector With Live Video - UIFacialGestuures

- - posted in ios

Making Facial Gestures or, Using CIDetector with Streaming Video

Since iOS7 Apple introduced new feature to the CIDetector, specifically smile and blinks detection.

I’ve wondered if i could “feed” it streaming video instead of static image and get make something like a live smile facial gesture be registered as an action on the device.

Here is a demo video of it:

I’ve extracted all the logic into an easy to use facade so it can easily be added in any project. Have a look at the The main issues where:

  1. CIDetector is a slow and takes up a lot of CPU

  2. Smartley aggregating the feature detection results from the CIDetector until enough features are considered a registered gestured.

The code and example project: