AI for Trust, Safety and Fraud - February 2024

9 minute read

Published:


Welcome back readers to AI for Trust, Safety and Fraud, where we explore insights in real world applications of machine learning with a focus on cybersecurity and risk management. 

In this February 2024 edition, we present a discussion with Shikhar Gupta, an Applied Scientist at Amazon. The conversation revolves around a crucial topic of sourcing labeled data for minority classes, employing research from their team on target-aware active learning. The topics covered in this article are:
  1. Identifying shortcomings in traditional active learning methods,
  2. Exploring the ethical implications of missing labels for minority classes,
  3. Unveiling real world applications of active learning, including fraud detection,
  4. Establishing a feedback loop to integrate active learning at the foundational model stage for large language models (LLMs), subsequently using LLMs for annotations.  
  5. image.png 

    An overview of target-aware active learning


    Kunal Kotian, Indranil Bhattacharya, Shikhar Gupta, Kaushik Pavani, Naval Bhandari, & Sunny Dasgupta (2023). TAAL: Target-aware active learning. In KDD 2023 Workshop on e-Commerce and NLP (ECNLP 6).

    Traditional active learning techniques have been proven effective in generating accurate classifiers with minimal labeling requirements compared to random sampling. For instance,  fraud detection teams building classifiers can benefit from active learning to find suspicious events for investigation, which when labeled increase the overall  classifier performance the most.  However, it might have a problem: the standard active learning methods overlook class-specific business targets . Consider a fraud detection classifier that is doing well on catching card-not-present fraud and fake accounts, but not on account takeovers. This can happen if the training data collection is agnostic to different fraud types and is especially harmful if account takeovers cause more monetary harm than other fraud types. This problem arises because the  standard active learning methods do not care about which class is being labeled, as long as labeling an unknown event leads to a  better performance of the overall classifier. The result of this problem is low precision/recall on minority classes.

    (If you want to revisit the common types of fraud, read my  detailed article on 12 Types of Fraud and Abuse that Everyone Should Know About). 

    This is where the idea of target-aware active learning steps in. This framework transforms any active learning strategy into "target-aware", by considering the gap between each class's current estimated accuracy and its corresponding business target. By sampling data points from fraud categories where the model is underperforming (e.g. account takeovers) , target-aware active learning can ensure efficient allocation of annotation resources (investigators) while achieving performance metrics for all fraud types.

    Building Responsible AI


    Ethical AI development necessitates ensuring fair and inclusive representation across all data classes. This is especially crucial in tasks like loan approvals, where underrepresented data points can lead to biased decisions with negative societal impacts. 

    Imagine an AI system used for loan approvals  that classifies applicants into multiple categories based on loan type (e.g., mortgage, auto loan, small business loan). Let's  say the system performs very well for common loan types requested by individuals with high credit scores and stable financial backgrounds (represented by the majority class in the training data). However, the system underperforms for less frequent loan types, like small business loans requested by first-time entrepreneurs or minority-owned businesses (represented by the minority class).

    The lower performance of AI based loan approver on less frequent loan types can have negative societal implications, leading to limited access to capital for minority groups and distrust for aspiring entrepreneurs on the financial system. AI systems, often being opaque, make it difficult for loan applicants to understand why their application was rejected. Therefore, a multi-class loan approval classifier that misses good performance on a minority class can raise serious ethical concerns.    

    Target-Aware Active Learning: A Solution: Target-aware active learning can offer a powerful solution to this challenge. By mitigating bias in data selection, it actively prioritizes underrepresented classes for human labeling. This ensures a more balanced training set that avoids favoring common patterns over potentially risky but less frequent ones.  The result is a system that learns from a broader range of data points, promoting fairer and more inclusive outcomes in loan approvals and various other applications.
     


    Q/A Session With Shikhar

    Q: How does target-aware active learning improve traditional uncertainty sampling in imbalanced datasets?

    A: Traditional uncertainty sampling (a common active learning approach) focuses on identifying the most confusing samples overall, neglecting the specific classes that might benefit the most from labeling. Target-aware active learning addresses this by considering the target precision for each class and prioritizing labeling confusing samples that are predicted to benefit lagging classes with minimum labeled data. This helps improve the model's performance on under-represented classes in imbalanced data. The good part about this framework is that it is plug and play with any active learning algorithm. The framework's greatest benefit lies in settings with constrained annotation budgets. By prioritizing informative samples for under-represented classes, it aims to achieve good performance with fewer labels, reducing the overall annotation cost. 

    image.png
    Click here to listen to the detailed answer


    Q: Beyond classification, what other potential applications are there for this framework?

    A: The framework can be generally applicable to any scenario where data collection is expensive and there's an inherent imbalance across the classes of interest. This could be relevant in various domains beyond what we discussed in the paper. 
    1. Medical diagnosis: In medical imaging tasks like chest X-ray analysis, certain diseases are much rarer than others. Standard active learning can improve overall model accuracy of detecting a disease from an X-ray, but it might neglect rare diseases. Target-aware active learning can help prioritize these rare diseases for radiologist review.   
    2. Large language models (LLMs): During the "human-in-the-loop" phase of LLM training, where humans refine the model's outputs, data collection can be expensive. Target-aware active learning can be used to prioritize the selection of prompts for human evaluation, focusing on those that would benefit the model's performance across all categories. For instance,  in LAMM: Language aware active learning for multilingual models, the researchers show that a same active learning strategy can help achieve accuracy on low resource languages with limited training data. This can help in ensuring multilingual models are performant on all languages of operation.  
    3. Fraud detection: Fraudulent events are inherently rare compared to legitimate transactions. Target-aware active learning could be applied to select the most informative transactions for human review, improving fraud detection accuracy for each fraud type, while reducing the cost of manual investigation.
    4. Ethical considerations in training data: The framework could be adapted to incorporate ethical dimensions into data selection. By considering factors like potential biases in the training data, it could help ensure fairer and more responsible model development.
    5. image.png
      Click here to listen to the detailed answer

      Q: Is there a publicly available library for the target-aware active learning framework?

      A: Shikhar's team did not create a public library for their specific framework. However, they acknowledge the existence of other repositories like distil: https://github.com/decile-team/distil that offer implementations of base active learning algorithms. This publicly available resource potentially served as a foundation for their research and experimentation.

      Q. Does the potential of using large language models (LLMs) as annotators render active learning obsolete?

      A: There is an interesting dynamic between LLMs and active learning:

      Copyright © 2024 AI for Trust, Safety and Fraud, All rights reserved.

      Want to change how you receive these emails?
      You can update your preferences or unsubscribe from this list.
      </td> </tr> </tbody> </table> </td> </tr> </tbody> </table> </div> </div> </div>

      beacon