Models Post-Training - 25

Welcome,



1. Post-training


1.1 Model Editing

  • ACL 2024 - An Easy-to-use Knowledge Editing Framework for LLMs
  • Easy Edit




1.2 Model Merging


  • 2021 Task Arithmetic - Linear combination of model weights
  • LoRA Merging - Fusion of multiple LoRA adapters


Knowledge Map


Knowledge Map


1.3 Machine Un-Learning




1.4 Hard to modify – update knowledge to LMs




2. Multi-LLM Agent





3. Why Can Neural Networks Be Compressed


1. Redundancy

Most neural networks contain abundant redundant parameters Many weights contribute minimally to final predictions Network capacity typically exceeds actual requirements



2. Over-parameterization

Modern NNs often have millions to billions of parameters Actual tasks may only require a small subset Large capacity needed for training, but can be streamlined for inference




References








Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • Neuroscience Basis For Hearing & Vision - 25
  • AI Character Consistency x Memory - 25
  • Mathematical Fractals
  • Structures for Speech Processing - 25
  • Speech Processing x Transfer Learning - 25