论文¶
通用¶
Agents¶
- React
- Chat with the Environment
- Reflexion: Language Agents with Verbal Reinforcement Learning
- TaskMatrix.AI
- Generative Agents
- ChatDev: Communicative Agents for Software Development
- MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework
- AgentSims: An Open-Source Sandbox for Large Language Model Evaluation
- AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation
- AgentVerse: Facilitating Multi-Agent Collaboration and Exploring Emergent Behaviors
- Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models
- Data Interpreter: An LLM Agent For Data Science
- Internet of Agents: Weaving a Web of Heterogeneous Agents for Collaborative Intelligence
- ADAS: Automated Design of Agentic Systems
- SELA: Tree-Search Enhanced LLM Agents for Automated Machine Learning
- AFlow: Automating Agentic Workflow Generation
- FACT: Examining the Effectiveness of Iterative Context Rewriting for Multi-fact Retrieval
大模型调优¶
- Prefix-Tuning: Optimizing Continuous Prompts for Generation
- p-tuning: GPT Understands, Too
- Prompt Tuning: The Power of Scale for Parameter-Efficient Prompt Tuning
- LoRA: Low-Rank Adaptation of Large Language Models
- FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
- QLoRA: Efficient Finetuning of Quantized LLMs
- Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models
- DoRA: Weight-Decomposed Low-Rank Adaptation
- LoRA+: Efficient Low Rank Adaptation of Large Models
- GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
- LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models
NLP LLM¶
- GPT1: Improving Language Understanding by Generative Pre-Training
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
- GPT2: Language Models are Unsupervised Multitask Learners
- CPM: A Large-scale Generative Chinese Pre-trained Language Model
- LLaMA: Open and Efficient Foundation Language Models
- Llama 2: Open Foundation and Fine-Tuned Chat Models
- Qwen Technical Report
- DeepSeek-Coder: When the Large Language Model Meets Programming – The Rise of Code Intelligence
- MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
- DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
- ChatGLM: A Family of Large Language Models from GLM-130B to GLM-4 All Tools
Vision LLm¶
Framework¶
ML¶
- WebGPT: Browser-assisted question-answering with human feedback
- Teaching language models to support answers with verified quotes
- FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation
- Evaluating Verifiability in Generative Search Engines
- Citation: A Key to Building Responsible and Accountable Large Language Models
- HAGRID: A Human-LLM Collaborative Dataset for Generative Information-Seeking with Attribution
- Enabling Large Language Models to Generate Text with Citations
RAG¶
- Retrieval-Augmented Generation for Large Language Models: A Survey
- CRAG: Corrective Retrieval Augmented Generation
- Adaptive-RAG
- From Local to Global: A Graph RAG Approach to Query-Focused Summarization
- GRAG: Graph Retrieval-Augmented Generation
- GraphRAG 官方文档
- Multi-Meta-RAG: Improving RAG for Multi-Hop Queries using Database Filtering with LLM-Extracted Metadata
Tools¶
手机业务¶
AGI¶
others¶
Highlighting the top ML papers every week: https://github.com/dair-ai/ML-Papers-of-the-Week