• ReLearn: Learning new things for Large Language Models

    Summary

    The paper “ReLearn: Unlearning via Learning for Large Language Models” presents a novel method for unlearning in LLMs while preserving fluency and relevance. It introduces a data augmentation and fine-tuning pipeline as an alternative to ‘gradient ascent (GA)’ and ’negative preference optimization (NPO]’, which degrade linguistic coherence.

    How to Implement This Paper

    To implement ReLearn, we will follow these key steps:

    1️⃣ Understanding the Core Approach

    • Data Augmentation: Generate diverse question-answer (QA) variations for forgetting while ensuring non-sensitive yet relevant responses.
    • Fine-Tuning: Replace the knowledge to be forgotten with relevant but non-sensitive responses.
    • Evaluation Metrics: Use the paper’s Knowledge Forgetting Rate (KFR), Knowledge Retention Rate (KRR), and Linguistic Score (LS) for performance assessment.

    2️⃣ Setting Up the Development Environment

    We need:

  • Rag: Retrieval-Augmented Generation

    Summary

    Retrieval-Augmented Generation (RAG) is a powerful technique that enhances large language models (LLMs) by allowing them to use external knowledge sources.

    An Artificial Intelligence (AI) system consists of components working together to apply knowledge learned from data. Some common components of those systems are:

    • Large Language Model (LLM): Typically the core component of the system, often there is more than one. These are large models that have been trained on massive amounts of data and can make intelligent predictions based on their training.