Open2
LLM重要事項メモ

- GraphRAG Methods
- LLMs for Tabular Data
- Automated Agentic Systems
- Enhancing Robustness in LLMs
- Controllable Text Generation for LLMs
- LLM Pruning and Distillation in Practice
- rStar
- HybirdRAG
- LongWriter
- EfficientRAG
- RAGChecker
- The AI Scientist
- SAM 2
- RAGEval
- Survey of Mamba
- Transformer Explainer
- Self-Taught Evaluators
- Conversational Prompt Engineering
-
- MindSearch
- Refusal in LLMs
- Constrained-CoT
- Meta-Rewarding LLMs
- Evaluating Persona Agents
- Improved RAG with Self-Reasoning
- LazyLLM
- Llama 3.1
- OpenDevin
- AlphaProof
- Text-to-SQL Survey
- RAG vs. Long-Context LLMs
- SpreadsheetLLM
- Weak-to-Strong Reasoning
- Improving LLM Output Legibility
- Distilling System 2 into System 1
- A Survey of Prompt Engineering in LLMs
- Context Embeddings improves RAG Efficiency
- RankRAG
- RouteLLM
- FlashAttention-3
- Internet of Agents
- Learning at Test Time
- Mixture of A Million Experts
- APIGen
- CriticGPT
- Agentless
- LLM See, LLM Do
- Scaling Synthetic Data Creation
- Searching for Best Practices in RAG
- ESM3
- Gemma 2
- Adam-mini
- GraphReader
- LLM Compiler
- Enhancing RAG with Long-Context LLMs
- TextGrad
- PlanRAG
- Claude 3.5 Sonnet
- DeepSeek-Coder-V2
- Mitigating Memorization in LLMs
- Tree Search for Language Model Agents
- SelfGoal
- Mixture-of-Agents
- Nemotron-4 340B
- Self-Tuning with LLMs
- Mixture of Memory Experts
- Multimodal Table Understanding
- NLLB
- Mamba-2
- AgentGym
- MatMul-free LLMs
- Buffer of Thoughts
- SimPO
- GNN-RAG
- Attention as an RNN
- Abacus Embeddings
- Symbolic Chain-of-Thought
- Contextual Position Encoding
- Guide for Evaluating LLMs
- Efficient Multimodal LLMs
- Scientific Applications of LLMs
- Enhancing Answer Selection in LLMs
- Claude 3 Sonnet Interpretable Features
- Agent Planning with World Knowledge Model
- Veo
- GPT-4o
- Chameleon
- RLHF Workflow
- Gemini 1.5 Flash
- Fine-tuning and Hallucinations
- xLSTM
- DrEureka
- AlphaFold 3
- DeepSeek-V2
- Consistency LLMs
- AlphaMath Almost Zero
- Med-Gemini
- When to Retrieve?
- Kolmogorov-Arnold Networks
- Multimodal LLM Hallucinations
- Self-Play Preference Optimization
- In-Context Learning with Long-Context Models
- Phi-3
- OpenELM
- AutoCrawler
- Self-Evolution of LLMs
- AI-powered Gene Editors
- Make Your LLM Fully Utilize the Context
- Llama 3
- Mixtral 8x22B
- A Survey on RAG
- How Faithful are RAG Models?
- Emerging AI Agent Architectures
- Chinchilla Scaling: A replication attempt
- CodeGemma
- Infini-Transformer
- Overview of Multilingual LLMs
- LM-Guided Chain-of-Thought
- The Physics of Language Models
- Best Practices and Lessons on Synthetic Data
- SWE-Agent
- Mixture-of-Depths
- Many-shot Jailbreaking
- Visualization-of-Thought
- Advancing LLM Reasoning
- Representation Finetuning for LMs
- Extracting Concepts from GPT-4

- DBRX
- Grok-1.5
- LLM2LLM
- Mini-Gemini
- Agent Lumos
- Long-form factuality in LLMs
- Grok-1
- TacticAI
- Agent-FLAN
- LLM4Decompile
- Evolutionary Model Merge
- Retrieval-Augmented Fine-Tuning
- SIMA
- Multimodal LLM Pre-training
- Knowledge Conflicts for LLMs
- Retrieval Augmented Thoughts
- LLMs Predict Neuroscience Results
- LMs Can Teach Themselves to Think Before Speaking
- Claude 3
• KnowAgent
• LLM for Law
• Design2Code
• RAG Enhancements Overview
• Robust Evaluation of Reasoning - Genie
- LearnAct
- Mistral Large
- Dataset for LLMs
- The Era of 1-bit LLMs
- LLMs on Tabular Data
- LoRA+
- Gemma
- Stable Diffusion 3
- OpenCodeInterpreter
- Revisiting REINFORCE in RLHF
- CoT Reasoning without Prompting
- Sora
- Gemini 1.5
- OS-Copilot
- TestGen-LLM
- Large World Model
- LLM Agents can Hack
- ALOHA 2
- DeepSeekMath 7B
- Self-Reflective Agents
- LLMs for Table Processing
- Indirect Reasoning with LLMs
- Grandmaster-Level Chess Without Search
- OLMo
- SliceGPT
- MoE-LLaVA
- Corrective RAG
- Advances in Multimodal LLMs
- LLMs for Mathematical Reasoning
- WARM
- Medusa
- AgentBoard
- MambaByte
- Knowledge Fusion of LLMs
- Resource-efficient LLMs & Multimodal Models
- AlphaCodium
- AlphaGeometry
- RAG vs. Finetuning
- Self-Rewarding Models
- Overview of LLMs for Evaluation
- Tuning Language Models by Proxy
- InseRF
- Sleeper Agent
- Blending Is All You Need
- Jailbreaking Aligned LLMs
- From LLM to Conversational Agents
- Prompting LLMs for Table Understanding
- DocLLM
- Mobile ALOHA
- Self-Play Fine-tuning
- Fast Inference of MoE
- LLM Augmented LLMs
- Mitigating Hallucination in LLMs
2024/01/01