• 0 Posts
  • 4 Comments
Joined 9 months ago
cake
Cake day: February 11th, 2024

help-circle

  • Thanks for the question, it actually made me look for the api. Looks like I misremembered it, and there aren’t actually any exposed APIs for developers regarding attention. Internally it’s used by iOS for checking when you’re looking at the screen for faceID and keeping the screen lit when you’re reading.

    There are APIs for developers that expose the position of the users head, but apparently it excludes eye information. Looks like it’s also pretty resource intensive, and mainly for AR applications.

    The faceID / touchID api essentially only returns “authenticated”, “authenticating”, and “unautheticated”. The prompts / UI are stock iOS and cannot be altered, save showing a reason.



  • I can really emphasise with Samir. Working in healthcare I’m basically limited to just the Office applications. However in the past few years I’ve been able to cook up solutions by reading / writing to file based databases, and using VBA to generate and bind to HTML contents on the fly for the built in IE11 instance. It’s as close to getting to some kind of web-stack within the confines of IT Sec in healthcare.