20241212 修订:

新增 CTR Prediction 相关的 5 篇论文:EDCN, GDCN, DCN V3, FINAL, FinalMLP

20230920 修订:

新增 LLM 量化章节,新增 9 篇相关的论文

  • 《Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference》
  • 《Mixed Precision Training》
  • 《The case for 4-bit precision: k-bit Inference Scaling Laws》
  • 《SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models》
  • 《LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale》
  • 《ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers》
  • 《SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot》
  • 《GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers》
  • 《LUT-GEMM: Quantized Matrix Multiplication based on LUTs for Efficient Inference in Large-Scale Generative Language Models》
  • 20230828 修订:

    7.Transformer(9) 章节新增两篇论文:LIMA、LLAMA2

    新增 PEFT 章节,新增 10 篇关于 LORA 和 ADAPTER 相关的热门论文

  • 《Parameter-Efficient Transfer Learning for NLP》
  • 《BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models》
  • 《LoRA: Low-Rank Adaptation of Large Language Models》
  • 《Towards a Unified View of Parameter-Efficient Transfer Learning》
  • 《AdapterDrop: On the Efficiency of Adapters in Transformers》
  • 《AdapterFusion: Non-Destructive Task Composition for Transfer Learning》
  • 《QLoRA: Efficient Finetuning of Quantized LLMs》
  • 《AdapterHub: A Framework for Adapting Transformers》
  • 《Compacter: Efficient Low-Rank Hypercomplex Adapter Layers》
  • 《MAD-X: An Adapter-based Framework for Multi-task Cross-lingual Transfer》
  • 20230801 修订:

    新增 36 篇关于 Prompt Engineering 的热门论文:

  • 《Chain of Thought Prompting Elicits Reasoning in Large Language Models》
  • 《Least-to-Most Prompting Enables Complex Reasoning in Large Language Models》
  • 《Automatic Chain of Thought Prompting in Large Language Models》
  • 《Self-Consistency Improves Chain of Thought Reasoning in Language Models》
  • 《Large Language Models are Zero-Shot Reasoners》
  • 《Calibrate Before Use: Improving Few-Shot Performance of Language Models》
  • 《What Makes Good In-Context Examples for GPT-3?》
  • 《Making Pre-trained Language Models Better Few-shot Learners》
  • 《It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners》
  • 《Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference》
  • 《GPT Understands, Too》
  • 《P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks》
  • 《Prefix-Tuning: Optimizing Continuous Prompts for Generation》
  • 《The Power of Scale for Parameter-Efficient Prompt Tuning》
  • 《How Can We Know What Language Models Know?》
  • 《Eliciting Knowledge from Language Models Using Automatically Generated Prompts》
  • 《Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity》
  • 《Can language models learn from explanations in context?》
  • 《Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?》
  • 《Multitask Prompted Training Enables Zero-Shot Task Generalization》
  • 《Language Models as Knowledge Bases?》
  • 《Do Prompt-Based Models Really Understand the Meaning of Their Prompts?》
  • 《Finetuned Language Models Are Zero-Shot Learners》
  • 《Factual Probing Is [MASK]: Learning vs. Learning to Recall》
  • 《How many data points is a prompt worth?》
  • 《Learning How to Ask: Querying LMs with Mixtures of Soft Prompts》
  • 《Learning To Retrieve Prompts for In-Context Learning》
  • 《PPT: Pre-trained Prompt Tuning for Few-shot Learning》
  • 《Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm》
  • 《Show Your Work: Scratchpads for Intermediate Computation with Language Models》
  • 《True Few-Shot Learning with Language Models》
  • 《Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning》
  • 《Improving and Simplifying Pattern Exploiting Training》
  • 《MetaICL: Learning to Learn In Context》
  • 《SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer》
  • 《Noisy Channel Language Model Prompting for Few-Shot Text Classification》
  • 20230524 修订:

  • 新增 Transformer 7、8、9 三章,包括:《Scaling Laws for Neural Language Models》、 《Training Compute-Optimal Large Language Models》、LLaMA、GLM、GLM-130B、 GPT-NeoX-20B、Bloom、PaLM、PaLM2、Self-Instruct 等十篇论文。
  • 20230516 修订:

  • 新增 HuggingFace Transformer 应用、Gradio。 所有 HuggingFace Transformer 官方教程和 API , 包括 Tokenizer、Dataset、Trainer、Evaluator、Pipeline、Model、Accelerate、AutoClass、应用,等九章内容
  • 20230408 修订:

  • 整理了 Paragraph Vector、Skip-Thought Vector、FastSent、InferSent、Simple-But-Tough-To-Beat Baseline For Sentence Embedding 、QuickThoughts 等六篇传统 sentence embedding 论文
  • 整理了 T5、mT5、ExT5、Muppet、Self-Attention with Relative Position Representation、 DeCLUTR、CLEAR、ConSERT、Sentence-T5、ULMFiT、USE、Sentence-BERT、SimCSE、BERT-Flow、BERT-Whitening、 Compare the Geometry of BERT/ELMo/GPT-2 Embedding、CERT 等十七篇论文
  • 整理了 CodeGen 一篇自动生成代码的论文
  • 20230221 修订:

  • 整理了 CTR 预估模型(神经网络方法 3)、CTR 预估模型(神经网络方法 4)、CTR 预估模型(神经网络方法 5), 共包括 AutoInt、 Fi-GNN、FwFM、FM2、FiBiNet、AutoFIS、DCN-V2、AFN、FGCNN、AutoCross、InterHAt、xDeepInt、BarsCTR、AutoDis、 MDE、NIS、AutoEmb、AutoDim、PEP、DeepLight 等二十篇经典论文
  • 20230213 修订:

  • 整理了 HuggingFace Transformer 官方教程和 API ,包括 Tokenizer、Dataset、Trainer、Evaluator、Pipeline、Model、Accelerate、AutoClass 等八章内容
  • 20221219 修订:

    20221219 修订:

    20221127 修订:

    20221106 修订:

    20221022 修订:

    20220925 修订:

    20220917 修订:

    20220822 修订:

    20220731 修订:

    20220526 修订:

    20220509 修订:

    20220403 修订:

    20220306 修订:

    20211121 修订:

    20211114 修订:

    20211031 修订:

    20211019 修订:

    20210925 修订:

    20210919 修订:

    20210807 修订:

    20210726 修订:

    20210620 修订:

    20210516 修订:

    20210420 修订:

    20210330 新增 scala 笔记

    来自于 《Scala 编程第三版》的笔记。

    20210317 修订:

    20210227 修订:

    20210120 修订:

    20201212 修订:

    20201117 修订:

    20201018 修订:

    20200921 修订:

    20200816 修订:

    20200726 修订:

    20200601 修订:

    20200405 修订:

    20200112 修订:

    20191201 修订:

    20190928 修订:

    20190825 修订:

    20190802 修订: