cs.CL(2023-12-14)

📊 共 15 篇论文 | 🔗 1 篇有代码

🎯 兴趣领域导航

支柱九:具身大模型 (Embodied Foundation Models) (13 🔗1) 支柱二:RL算法与架构 (RL & Architecture) (2)

🔬 支柱九:具身大模型 (Embodied Foundation Models) (13 篇)

#题目一句话要点标签🔗
1 Self-Evaluation Improves Selective Generation in Large Language Models 自我评估提升大语言模型的选择性生成能力 large language model
2 TAP4LLM: Table Provider on Sampling, Augmenting, and Packing Semi-structured Data for Large Language Model Reasoning TAP4LLM:通过采样、增强和打包半结构化数据,提升大语言模型在表格推理任务中的性能。 large language model
3 Evaluating Large Language Models for Health-related Queries with Presuppositions UPHILL数据集揭示大型语言模型在处理带预设的健康查询时的事实性缺陷 large language model
4 Language Modeling on a SpiNNaker 2 Neuromorphic Chip 在SpiNNaker 2神经形态芯片上实现语言建模,性能媲美LSTM large language model
5 Arabic Mini-ClimateGPT : A Climate Change and Sustainability Tailored Arabic LLM 提出Arabic Mini-ClimateGPT,一个面向气候变化和可持续性的阿拉伯语LLM。 large language model
6 The Earth is Flat because...: Investigating LLMs' Belief towards Misinformation via Persuasive Conversation 通过说服式对话研究大语言模型对错误信息的信念操纵 large language model
7 TigerBot: An Open Multilingual Multitask LLM TigerBot:开源多语言多任务大语言模型 large language model
8 Metacognition-Enhanced Few-Shot Prompting With Positive Reinforcement 提出元认知增强的少样本提示方法,结合正向激励提升大语言模型性能 large language model
9 Measurement in the Age of LLMs: An Application to Ideological Scaling 利用大型语言模型进行意识形态倾向性测量,解决社会科学中概念模糊问题。 large language model
10 Towards Verifiable Text Generation with Evolving Memory and Self-Reflection 提出VTG框架,通过演进记忆和自反思实现可验证的文本生成。 large language model
11 Zebra: Extending Context Window with Layerwise Grouped Local-Global Attention Zebra模型:通过分层分组局部-全局注意力机制扩展LLM上下文窗口 large language model
12 ZeroQuant(4+2): Redefining LLMs Quantization with a New FP6-Centric Strategy for Diverse Generative Tasks 提出FP6中心策略以优化LLMs的量化方法 large language model
13 Identifying Planetary Names in Astronomy Papers: A Multi-Step Approach 提出一种多步骤流程,用于识别天文论文中的行星地名,并实现高精度消歧。 large language model

🔬 支柱二:RL算法与架构 (RL & Architecture) (2 篇)

#题目一句话要点标签🔗
14 Fewer is More: Boosting LLM Reasoning with Reinforced Context Pruning CoT-Influx:强化上下文剪枝提升大语言模型数学推理能力 reinforcement learning large language model chain-of-thought
15 Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision 弱监督到强泛化:利用弱监督激发强大模型能力 reinforcement learning RLHF

⬅️ 返回 cs.CL 首页 · 🏠 返回主页