Inherently Privacy-Preserving Robotic Vision
Do you have a robot vacuum cleaner? Perhaps one of the new generation robots that uses a camera to navigate around your house? Where do those camera images go? Who can see them? Perhaps the images should never leave your house. Perhaps they should never leave the robot or the camera chip. Perhaps, to best protect your privacy, the images, as we know them, should never be formed in the first place.
- We present a robotic vision framework that never captures images or reveals information for reconstruction
- We propose guidelines for creating inherently privacy-preserving vision systems
- Our simulated localisation study shows comparable performance to conventional approaches
- We identify four practical approaches to implementing the system in hardware
This is a first step, and we hope to inspire future works that expand the range of applications open to sighted robotic systems.
Publications
• A. K. Taras, N. Suenderhauf, P. Corke, and D. G. Dansereau, “Inherently privacy-preserving vision for trustworthy autonomous systems: Needs and solutions,” Journal of Responsible Technology, vol. 17, p. 100079, 2024. Preprint here.
• Best Poster Award. Adam K. Taras, Niko Suenderhauf, Corke, and Donald G. Dansereau, “The Need for Inherently Privacy-Preserving Vision in Trustworthy Autonomous Systems,” International Conference on Robotics and Automation (ICRA) Workshop: Multidisciplinary Approaches to Co-Creating Trustworthy Autonomous Systems, 2023. Preprint, poster.
• Summary video, Honours thesis seminar
• Appearance on local news and associated press release.
Citing
@article{taras2024inherently, title = {Inherently Privacy-Preserving Vision for Trustworthy Autonomous Systems: Needs and Solutions}, author = {Adam K. Taras and Niko Suenderhauf and Peter Corke and Donald G. Dansereau}, journal = {Journal of Responsible Technology}, volume = {17}, pages = {100079}, year = {2024} }
Collaborators
Themes
Reconstruction challenge
See the github repo and discussion forum to get started on the challenge and share your progress and ideas.
Dataset
Inherently Private Imaging Architecture
Current robotic vision forms human-interpretable images that are rich with private data. Features are extracted and processed in the vulerable digital domain, where attackers can access all of this information.
Instead, we propose to shift processing into the optical-analogue domain. In this example a micromirror device filters light in a scene to multiplex data onto a single-pixel sensor. Analogue processing performs privacy-preserving summarisation and hashing before digitisation. Specialised task-specific algorithms then operate on the secure hashes in the digital domain to perform important robotics tasks. Privacy is ensured because reverting the optical-analogue hashing is intractable.
Gallery
(click to enlarge)
The signals from the sensor are processed by analogue electronics that form a summary, in this case by accumulating the maximum and minimum as a pair. This is repeated over many curves to build up a hash or "fingerprint" of the image.
The circle features find local edges at scales of tens of pixels, while lines find larger extremes more frequently.
For more results, including an analysis of which pixels the feature data comes from, please see the paper.
How-to: implementing inherently privacy-preserving vision
- Specialise the camera to the task; this sacrifices generality for privacy as the camera can only be used for the task(s) it's designed for,
- Shift as much processing as possible out of the digital domain, keeping it out of reach of remote attack,
- Maximise information-destroying operations prior to digitisation,
- Apply obscuration prior to digitisation such that brute-force attack becomes the only option for inverting the imaging process,
- Consider all information already available to the attacker, e.g. sequences of data and priors to improve domain performance and ensure privacy, and
- Maximise ambiguity, so that even a successful brute-force inversion of the imaging process is not likely to yield the correct image.
Call to Action
- Characterising and refining hardware implementations based on the approaches proposed here;
- Establishing meaningful metrics for privacy in the context of optical-analogue processing and hashing;
- End-to-end design of optical, analogue and algorithmic processing to tackle a broader range of vision tasks;
- Establishing trustworthiness through rigorous attack and refinement of the proposed concepts;
- Communicating and educating in an accessible manner to address the barriers to societal acceptance of sighted, privacy-preserving systems.