There has been increasing concern within the machine learning community and beyond that Artificial Intelligence (AI) faces a bias and discrimination crisis, urgently requiring AI systems to incorporate fairness constraints. The US Congress has recognized this issue and has been trying to pass the Algorithmic Accountability Act. It demands systems be evaluated for “accuracy, fairness, bias, discrimination, privacy and security within automated systems and companies would be required to correct any issues they uncovered during the process.” Most existing work on evaluating fairness assumes the availability of records in which the source data is annotated with categories needed to apply the fairness definition and fairness algorithm at hand. This assumption, however, is impractical in a diversity of real-world, socially-sensitive applications, ranging from precision medicine to marketing analytics, actuarial analysis and recidivism prediction instruments. There is thus a critical need to study the problem that arises from the gap between the design of a “fair” model in the lab and its deployment in the real world. To this end, this project will revisit the foundational definitions of fairness and reveal idiosyncrasies in the existing fairness literature stemming from assuming information that is not available in practice. Next, this project will aim to bridge the gap between current AI fairness studies and their real-world deployment, leading to improved understanding of the societal impact of AI and significant reduction in its potential for social discrimination. <br/><br/>To achieve this goal, the project will formulate a new fairness-as-a-survival-analysis problem, where the availability of class labels is not always guaranteed, but there is still a requirement that similar individuals are treated similarly. The first research objective focuses on quantifying individual unfairness in the presence of missing labels from two different perspectives. Specifically, the first track will see fairness as the correlation of similarity in the input and output spaces, which enables defining a fairness measure usable on statistically censored data. The second definition will constitute another fairness issue arising from the perspective of robustness, evaluating whether similar individuals suffer dissimilar levels of prediction stability. The second research objective will make an initial investigation jointly addressing bias reduction and statistical censoring management in model building, so as to ensure utility maximization while minimizing bias across individuals. These criteria will be formulated as regularization terms for joint optimization and will not require all individuals to have a class label. The outcomes of this project are expected to include versatile artifacts that ensure fairness guarantees in various real-world socially-sensitive applications. Furthermore, the project will introduce a new task setting, paving the way for future research in the practical application of AI fairness.<br/><br/>This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.