A new greedy, brute-force solution improves fairness in ranking and encompasses realistic scenarios with multiple, unknown protected groups.
A new greedy, brute-force solution improves fairness in ranking and encompasses realistic scenarios with multiple, unknown protected groups.
Ranking systems are omnipresent in our digital daily lives, from search engines that return a list of relevant web pages, to e-commerce services that recommend items to buy, to social media platforms that suggest friends. The impact of such systems is often illustrated in the case of job applications where the candidates become the ‘data points’ to be ranked and where they’re subject to potential biases in the system. The biases found in ranking systems generally originate either from the data used by the (often, machine-learning-based) approach or from the ranking algorithm itself. The harm resulting from the ranking algorithm bias is generally related to the notion of ‘exposure’ inherent in ranking systems and its unfair distribution across ranked items. Exposure simply denotes the amount of attention one allocates to the different items in the ranking list. Naturally, exposure is not evenly distributed across items: items at the top of the ranking tend to receive far more attention than items at the bottom. For example, a search engine user usually settles for the first few results and rarely scrolls down to the bottom of the page.
Due to the potential unfairness stemming from algorithmic bias, researchers have recently been studying ways to balance the exposure allocated to ranking items while preserving the quality of the ranking—giving rise to the so-called ‘fair ranking’ problem. Most studies in the fair ranking literature (1, 2, 3, 4, 5) assume that the items to rank pertain to a ‘protected group’. In the case of job candidate ranking, one could for example consider two protected groups as the set of ‘male’ candidates and the set of ‘female’ candidates. The protected group membership is then utilized to enforce fairness, for example to fairly rank men and women for the job offer. In the widely adopted merit-based definition of fairness, a fair ranking system provides exposure to each protected group which is proportional to the overall merit of the group members. The merit of an item indicates its utility for the task underlying the ranking, such as the adequacy of a candidate to a job or the relevance of a web page to a user’s query.
Despite recent efforts aimed at achieving fair ranking, current approaches rely on two limiting assumptions:
Our work attempts to address these two limitations to get closer to providing a ranking system that ensures algorithmic fairness.
The problem we tackle is inspired by the set-up introduced in TREC Fair Ranking 2019 (5), which is a track in a long-standing series of information retrieval evaluation workshops. The goal of this track was to devise systems which, given a query about a scientific topic, would be able to rank a set of scientific articles so that higher-ranked documents would be more relevant to the query, and returned rankings would provide fair exposure to the different groups of authors who wrote those articles. The twist is that the challenge participants didn’t have access to the author-group assignments, and that several different, a priori unknown, grouping definitions would be used in the evaluation.
Our approach to fair ranking for multiple, unknown groupings (or ‘multi-grouping robust fair ranking’) relied on two components: a strategy to define some proxy grouping(s) to alleviate the lack of known groupings, and a fair ranking algorithm to embed those proxy groupings.
In the absence of known groupings, we explored different possible proxy groupings to act as substitutes. We focused on two types of strategies. The first one simply assumes that each author (or alternatively each article) made up their own group, known in the literature as ‘individual fairness’. The rationale for this choice is that individual fairness often subsumes group-level fairness (6). The second strategy we investigated relied on generating synthetic, random groupings. Although this may not appear to be very intuitive, the underlying idea is that enforcing fairness for a large enough number of random groupings leads to improving the coverage of possible groupings. It should therefore help in being more robust with respect to the set of ‘true’, unknown groupings considered for evaluation. Among the random groupings, we generated balanced groupings where the size of each group was approximately the same, and unbalanced groupings which reflected a more heterogeneous situation. To define the unbalanced random groupings, we relied on an existing stochastic process known as the Chinese restaurant process (CRP) (7). Figure 1 shows the different types of groupings considered.
Given one (or several) of the proxy groupings described above, we defined a fair ranking algorithm to issue rankings that are both useful with respect to the input queries and fair according to the chosen proxy grouping(s). Our approach is a simple, efficient heuristic which we named single-query greedy brute-force re-ranking (SGBR). It features the following characteristics:
Performing true brute force would consist of generating and evaluating all permutations of the documents which need to be ranked for a query. As this is not feasible for a large set of documents, we followed a two-step methodology which considerably reduced the number of rankings to consider. We briefly summarize these two steps below, which are described in full mathematical detail in our paper (8).
The first step, which we denote as pre-ordering, computes an initial ranking of the documents associated with a query. This ranking should account simultaneously for the documents’ relevance to the query and for the potential exposure deficit of the groups associated with the documents in previous occurrences of the query based on the merit of those groups. In other words, the documents ranked at the top of the pre-ordering are those which are relevant and whose groups (from the proxy groupings) received too little exposure in the rankings issued for past occurrences of the same query. This process is depicted in Figure 2(a).
The second step, labelled top-item brute-force re-ranking, generates all the permutations of the top items in the pre-ordering while keeping the tail items in the same position. The rationale for permuting only the top items is that they are the most likely to have a large impact on the overall relevance and fairness of the ranking. Each candidate ranking formed in this way is then evaluated with the target metric. This metric balances the utility of the ranking (i.e. relevance with respect to the query) and fairness according to the proxy grouping(s). Finally, the ranking with the best score is returned for the current occurrence of the query. This step is illustrated in Figure 2(b).
Our experiments based on the TREC Fair Ranking (5) benchmark simulate the scenario where fairness is evaluated with respect to a large number of different, unknown groupings. For that purpose, we randomly generated for the evaluation a new set of unbalanced groupings from a CRP. Unbalanced groupings were adopted as evaluation groupings to reflect that, in reality, protected groups are often unevenly distributed (e.g. the gender of applicants for computer engineering jobs). In our evaluation, we did not rely on the groupings proposed by the track’s organizers, as only two such groupings were provided—we judged this too limited for a reliable evaluation in a multi-grouping setting.
We compared the performance of the rankings generated by our SGBR-based approach on the different proxy groupings. We also considered as baselines two simple approaches: one that maximizes the utility of the ranking by ordering documents based on their relevance to the query, and another approach which returns a random permutation of the documents, ignoring their relevance. Our experimental results revealed that the variants based on SGBR and individual proxy groupings yielded both high utility and low unfairness in comparison to the other approaches. Indeed, the SGBR approach with individual groupings led to a reduction in the unfairness score of between 18 and 22% relative to the ‘maximum utility’ approach, while keeping utility within 0.0001% of the maximum possible utility. More details on the results are provided in our article (8). Our experiments empirically support the fact that adopting individual fairness seems to be the best choice when fairness is to be enforced with respect to a set of unknown groupings.
We’re currently working on characterizing when this individual fairness strategy is optimal from a theoretical standpoint, and on extending it to the multi-stakeholder setting, where fairness should be ensured simultaneously on the side of the producer (e.g. the document author) and on the side of the consumer (e.g. the reader). How we succeed will be the topic of a future post.
NAVER LABS Europe 6-8 chemin de Maupertuis 38240 Meylan France Contact
To make robots autonomous in real-world everyday spaces, they should be able to learn from their interactions within these spaces, how to best execute tasks specified by non-expert users in a safe and reliable way. To do so requires sequential decision-making skills that combine machine learning, adaptive planning and control in uncertain environments as well as solving hard combinatorial optimization problems. Our research combines expertise in reinforcement learning, computer vision, robotic control, sim2real transfer, large multimodal foundation models and neural combinatorial optimization to build AI-based architectures and algorithms to improve robot autonomy and robustness when completing everyday complex tasks in constantly changing environments. More details on our research can be found in the Explore section below.
For a robot to be useful it must be able to represent its knowledge of the world, share what it learns and interact with other agents, in particular humans. Our research combines expertise in human-robot interaction, natural language processing, speech, information retrieval, data management and low code/no code programming to build AI components that will help next-generation robots perform complex real-world tasks. These components will help robots interact safely with humans and their physical environment, other robots and systems, represent and update their world knowledge and share it with the rest of the fleet. More details on our research can be found in the Explore section below.
Visual perception is a necessary part of any intelligent system that is meant to interact with the world. Robots need to perceive the structure, the objects, and people in their environment to better understand the world and perform the tasks they are assigned. Our research combines expertise in visual representation learning, self-supervised learning and human behaviour understanding to build AI components that help robots understand and navigate in their 3D environment, detect and interact with surrounding objects and people and continuously adapt themselves when deployed in new environments. More details on our research can be found in the Explore section below.
Details on the gender equality index score 2024 (related to year 2023) for NAVER France of 87/100.
The NAVER France targets set in 2022 (Indicator n°1: +2 points in 2024 and Indicator n°4: +5 points in 2025) have been achieved.
—————
Index NAVER France de l’égalité professionnelle entre les femmes et les hommes pour l’année 2024 au titre des données 2023 : 87/100
Détail des indicateurs :
Les objectifs de progression de l’Index définis en 2022 (Indicateur n°1 : +2 points en 2024 et Indicateur n°4 : +5 points en 2025) ont été atteints.
Details on the gender equality index score 2024 (related to year 2023) for NAVER France of 87/100.
1. Difference in female/male salary: 34/40 points
2. Difference in salary increases female/male: 35/35 points
3. Salary increases upon return from maternity leave: Non calculable
4. Number of employees in under-represented gender in 10 highest salaries: 5/10 points
The NAVER France targets set in 2022 (Indicator n°1: +2 points in 2024 and Indicator n°4: +5 points in 2025) have been achieved.
——————-
Index NAVER France de l’égalité professionnelle entre les femmes et les hommes pour l’année 2024 au titre des données 2023 : 87/100
Détail des indicateurs :
1. Les écarts de salaire entre les femmes et les hommes: 34 sur 40 points
2. Les écarts des augmentations individuelles entre les femmes et les hommes : 35 sur 35 points
3. Toutes les salariées augmentées revenant de congé maternité : Incalculable
4. Le nombre de salarié du sexe sous-représenté parmi les 10 plus hautes rémunérations : 5 sur 10 points
Les objectifs de progression de l’Index définis en 2022 (Indicateur n°1 : +2 points en 2024 et Indicateur n°4 : +5 points en 2025) ont été atteints.
To make robots autonomous in real-world everyday spaces, they should be able to learn from their interactions within these spaces, how to best execute tasks specified by non-expert users in a safe and reliable way. To do so requires sequential decision-making skills that combine machine learning, adaptive planning and control in uncertain environments as well as solving hard combinatorial optimisation problems. Our research combines expertise in reinforcement learning, computer vision, robotic control, sim2real transfer, large multimodal foundation models and neural combinatorial optimisation to build AI-based architectures and algorithms to improve robot autonomy and robustness when completing everyday complex tasks in constantly changing environments.
The research we conduct on expressive visual representations is applicable to visual search, object detection, image classification and the automatic extraction of 3D human poses and shapes that can be used for human behavior understanding and prediction, human-robot interaction or even avatar animation. We also extract 3D information from images that can be used for intelligent robot navigation, augmented reality and the 3D reconstruction of objects, buildings or even entire cities.
Our work covers the spectrum from unsupervised to supervised approaches, and from very deep architectures to very compact ones. We’re excited about the promise of big data to bring big performance gains to our algorithms but also passionate about the challenge of working in data-scarce and low-power scenarios.
Furthermore, we believe that a modern computer vision system needs to be able to continuously adapt itself to its environment and to improve itself via lifelong learning. Our driving goal is to use our research to deliver embodied intelligence to our users in robotics, autonomous driving, via phone cameras and any other visual means to reach people wherever they may be.
This web site uses cookies for the site search, to display videos and for aggregate site analytics.
Learn more about these cookies in our privacy notice.
You may choose which kind of cookies you allow when visiting this website. Click on "Save cookie settings" to apply your choice.
FunctionalThis website uses functional cookies which are required for the search function to work and to apply for jobs and internships.
AnalyticalOur website uses analytical cookies to make it possible to analyse our website and optimize its usability.
Social mediaOur website places social media cookies to show YouTube and Vimeo videos. Cookies placed by these sites may track your personal data.
This content is currently blocked. To view the content please either 'Accept social media cookies' or 'Accept all cookies'.
For more information on cookies see our privacy notice.