Section 4. Computer Vision#
In this lesson, users will be given insight into how Computer Vision (CV) methods can impact health equity. This lesson includes an overview of common methods and research objectives, health equity challenges, and a more detailed case study.
Health Equity and Computer Vision
|
What is Computer Vision#
Computer Vision (CV) is the scientific field that gains insight into using computers for processing visual information to determine the shape, appearance, or identity of objects. Computer vision is used for tasks such as monitoring safety compliance (i.e. tracking workers in an area), biomolecular research, and diagnostics. Computer vision tasks such as object recognition are inherently prone to bias because they are an inverse problem, where unknowns are estimated using partial information. Probabilistic models are used to help disambiguate between potential solutions, which introduces bias into the system.
The coded gaze, which is implanted into AI-driven computer vision platforms sees the world through the programmatic bias entered into it. Where computer vision platforms are built to detect different people, the implications of bias could result in incorrect identification of individuals. Historically, computer vision platforms have had issues detecting people of color and discriminating gender [Buolamwini, 2018],[Schwemmer et al., 2020],[Wang et al., 2022].
Computer vision problems often consist of one of the following research objectives:
Image Preprocessing: Involves formatting images prior to being used to train a model. This can include methods such as image warping (used for image scaling, translation, and rotation or may be used for correction of image distortion) and image de-noising (used for the retrieval of an unknown original image from a noisy one).
Image Clustering: Image clustering algorithms are used for classification of objects in an image when there is no classification training data, making is a type of unsupervised learning. These clustering methods find natural groups in feature space.
Image Classification: Image classification includes labeled training data and can produce stronger results than clustering. It is a type of supervised learning.
Boundary Detection: defines the boundary between two or more objects or locations in an image.
Object Detection: an object is identified and includes the use of feature analysis and neural networks.
Face Detection: detects the location of a face in an image and includes the use of feature analysis and neural networks.
How Computer Vision is Used#
Some examples of computer vision problems within a public health context are:
Medical Image Segmentation: Supervised medical image segmentation is used in order to perform edge detection and object detection to automate the identification of anatomical structure and regions of interest.
Health Monitoring: Computer vision classification algorithms may be applied on unlabeled facial scans for predicting early symptoms of infection and illness.
Early Diagnosis: A computer vision model may be built to detect cancerous cells in human-annotated tissue image samples.
Computer Vision Methods#
Bias resulting in misidentification of biometric attributes such as gender, age, and ethnicity carries potential for great consequences. Therefore, common methods in object detection are emphasized in more detail below:
Object Recognition is a type of object detection. In object detection, instances of an object are identified in a series of images. In object recognition, the object itself is identified. Control for bias in object recognition is of extreme pertinence because misidentifying objects carries potential for great consequences. [Waithe et al., 2020]
Haar Cascades is a method of object detection that uses a collection of positive and negative samples to identify the object. Positive samples include the object and negative samples do not.
Face Recognition Facial recognition is a type of object detection in which faces are recognized. Control for bias in facial recognition is of extreme pertinence because misidentifying faces carries potential for great consequences. [Libby and Ehrenfeld, 2021]
Feature Analysis is a method of facial recognition in which individual features are used to identify the individual. Feature examples include distance between the eyes, length of the chin, tone of the lips, etc.
Convolutional Neural Networks (CNNs) are artificial neural networks with convolution and pooling layers that identify features. They are multilayer perceptrons, meaning that each neuron in a layer is connected to all neurons in the adjacent layer and this makes them prone to overfitting. Regularization curbs this effect. We suggest taking special care when choosing a regularizer, testing multiple options before choosing the optimal regularizer.
Adaptive Neuro-Fuzzy Interference System (ANFIS) is a type of artificial neural network that takes advantage of fuzzy logic principles and is reported to show high accuracy in facial recognition.
Eigenfaces is a method of facial recognition that uses dimensionality reduction of images. The eigenfaces form the basis of images used to construct the covariance matrix for a vector space of images. This basis may be used to reconstruct any of the original images in the set.
Fischerfaces is an improvement upon the eigenfaces method that can interpolate and extrapolate over lighting and facial expressions.
Thermal Face Recognition models learn the unique temperature patterns of the face. The ability of these models to extract features does not depend on the presence of makeup or glasses.
The table below contains an overview of common methods in computer vision. More experienced readers may want to jump ahead to the next section on health equity considerations, which includes health equity challenges with computer vision and a more in depth case study.
If you are already familiar with computer vision methods, please continue to the next section. Otherwise click here.
Model | Common Usage | Suggested Usage | Suggested Scale | Interpretability | Common Concerns |
---|---|---|---|---|---|
Image De-noising |
|
Small to large datasets | High | ||
Warping Images |
|
Small to large datasets | Low | ||
Content-based Image Retrieval |
|
Small to large datasets | High | ||
K-Means Clustering |
|
Large datasets with multiple variables, but not too many variables | Moderate | ||
Hierarchical Clustering |
|
|
Moderate | ||
Spectral Clustering |
|
Large datasets with too many dimensions | Moderate | ||
K-Nearest Neighbors |
|
Medium to large datasets | Moderate | ||
Bayes Classifier |
|
P(y|X) = P(X|y)P(y)/P(X) , |
Medium to large datasets | High | |
Support Vector Machines |
|
Medium to large datasets | Moderate | ||
Edge-Based Segmentation |
|
Small to large datasets | Low | ||
Threshold-Based Segmentation |
|
Small to large datasets | Low | ||
Region-Based Segmentation |
|
Small to large datasets | Low | ||
Haar Cascades |
|
Large datasets | Low | ||
Feature Analysis |
|
Large datasets | Low | ||
Convolutional Neural Networks (CNNs) |
|
Large datasets | Low | ||
Adaptive Neuro-Fuzzy Interference System (ANFIS) |
|
Large datasets | Low | ||
Eigenfaces |
|
Large datasets | Low | ||
Fischerfaces |
|
Large datasets | Low | ||
Thermal Facial Recognition |
|
Large datasets | Low |
Health Equity Considerations#
Below are definitions of common sources of bias in computer vision and descriptions on how to mitigate these biases in the context of public health. For a broader review of bias in machine learning, please see the section on Machine Learning, which includes lessons on supervised, unsupervised, and reinforcement learning.
Challenge |
Challenge Description |
Health Equity Example |
Recommended Best Practice |
---|---|---|---|
Measurement Bias |
|
|
|
Selection Bias |
|
|
|
Recall Bias |
|
|
|
Confirmation Bias |
|
|
|
Demographic Bias |
|
|
|
Racial Bias |
|
|
|
Gender Bias |
|
|
|
Overfitting and Underfitting |
|
|
|
Outliers and Exclusion |
|
|
|
Case Study Example#
Case study is for illustrative purposes and does not represent a specific study from the literature.
Scenario: A researcher wants to explore the spread of influenza in the workplace with the help of a computer vision platform.
Specific Model Objective: Build a computer vision platform based on a labeled image dataset composed of masked facial recognition data to support downstream modeling in order to analyze the effect of demographic background on the spread of influenza.
Data Source: Masked facial recognition data coupled with thermal screenings was obtained for 5000 people of known vaccine status from 50 corporate office sites, one in each state. Racial distribution of participants in the training dataset were as follows: black (36%), white (44%), latino (18%), and asian (2%).
Analytic Method: A Convolutional Neural Network (CNN) was trained to perform facial recognition on masked participants.
Results: Leveraging the CNN model, a predictive model achieved an overall accuracy of 96% for racial classification and overall accuracy of 92% for health monitoring (sick vs healthy status).
Health Equity Considerations: Computer vision platforms used for facial recognition should be able to correctly identify demographic data, including people of all races equally. The failure to recognize some races may skew results and therefore public health recommendations or policies and lead to disparities. Below are additional considerations for this study:
There is an imbalance in the training data that leads to a large false positive rate for asian people. The computer vision platform for this study was written to recognize people wearing masks using a training dataset that consisted of only 2% asian participants. This is indicative of demographic bias as the algorithm has learned to classify black and white people with much higher accuracy.
Effects of this bias means that researchers may conclude that influenza spread is lower among the asian population than what it is in reality since fever measurement data is not properly tracked.
Facial recognition technology can produce racially biased results if the underlying data used is not representative of the community being served. One way to mitigate this bias is to ensure the training dataset has a large enough sample size with an acceptable minimum number of samples representing each possible race in the dataset.
In the analysis, outliers were defined as samples lying outside of three standard deviations from the mean. Post-analysis, it was observed that all of the outliers removed were of a specific minority population. By removing them they have introduced demographic bias into their system. To mitigate this bias, consider redefining the outlier boundary or develop a separate models for different populations.
When tested on unseen individuals, it was found that recall bias was present in predicting healthy vs infected individuals. This indicates that fever readings may have been inaccurate and the training data and improperly skewed toward a sick population with mislabeled data. Additional information such as a survey of how participants are feeling or PCR test could be useful in applying correct labels.
Finally, researcher may also want to carefully consider the metric used during model training and evaluation. For example, one might choose a specific test to reduce misidentification of participant race (reduce false positives). On the other hand, one might prioritize a sensitive test to detect fever/infection at the expense of misclassifying an individual as ill when they are in fact healthy (more false positives).
Considerations for Project Planning
|