Category: AI News

  • AI Research Highlights | Week 47, 2023

    AI Research Highlights | Week 47, 2023

    1. The Impact of Large Language Models on Scientific Discovery: a Preliminary Study using GPT-4 Source: https://arxiv.org/abs/2311.07361 In this report, Microsoft researchers delved into the performance of LLMs within the context of scientific discovery, focusing on GPT-4, the state-of-the-art language model. The investigation spans a diverse range of scientific areas…

  • AI Research Highlights | Week 43, 2023

    AI Research Highlights | Week 43, 2023

    1. Theory of Mind for Multi-Agent Collaboration via Large Language Models Source: https://arxiv.org/abs/2310.10701 Researchers evaluated LLM-based agents in a multi-agent cooperative text game with Theory of Mind (ToM) inference tasks, comparing their performance with Multi-Agent Reinforcement Learning (MARL) and planning-based baselines. The results demonstrate that LLM-based agents can handle complex…

  • AI Research Highlights | Week 42, 2023

    AI Research Highlights | Week 42, 2023

    1. Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models Source: https://arxiv.org/abs/2310.06117 In this paper, Deepmind researchers suggested Step-back Prompting as a way to enhance LLMs’ capacity for reasoning. This prompting approach consists of two steps: abstraction and reasoning. It was inspired by how humans use abstraction…

  • AI Research Highlights | Week 41, 2023

    AI Research Highlights | Week 41, 2023

    1. Language Models Represent Space and Time Source: https://arxiv.org/abs/2310.02207 Linguists once pointed out that LLMs were just stochastic parrots, however, this paper proved that an LLM is not just a collection of superficial statistics, but a world model that truly learns structured knowledge of fundamental dimensions. A group of MIT…

  • AI Research Highlights | Week 44, 2023

    1. Open-Ended Instructable Embodied Agents with Memory-Augmented Large Language Models Source: https://arxiv.org/abs/2310.15127 Researchers from Carnegie Mellon University introduced HELPER (Human-instructable Embodied Language Parsing via Evolving Routines), an embodied agent equipped with external memory of language-program pairs that parses free-form human-robot dialogue into action programs through retrieval-augmented LLM prompting. The project…

  • AI Research Highlights | Week 45, 2023

    1. ChatCoder: Chat-based Refine Requirement Improves LLMs’ Code Generation Source: https://arxiv.org/abs/2311.00272 PKU researchers proposed ChatCoder: a method to refine the requirements via chatting with LLMs. They designed a chat scheme (shown below) in which the LLMs will guide the human users to refine their expression of requirements to be more…

  • Top 10 Most Influential Projects on AI in 2022

    The year 2022 saw significant advancements in the realm of Artificial Intelligence(AI). Natural Language Processing, Image Generation, Machine Learning, and other related fields have all made remarkable strides, making AI tools widely used across various domains. In the first half of 2022, a large number of generative AI models emerged,…

  • AI Research Highlights | Week 46, 2023

    1. MFTCoder: Boosting Code LLMs with Multitask Fine-Tuning Source: https://arxiv.org/abs/2311.02303 Researchers from the Ant group proposed MFTCoder, an open-source project of CodeFuse for multitasking Code-LLMs, which includes models, datasets, training codebases, and inference guides. The focus is on addressing the issues of data balance and convergence speed that commonly arise…

  • AI Research Highlights | Week 40, 2023

    1. Large Language Model Alignment: A Survey Source: https://arxiv.org/abs/2309.15025 A research team from Tianjin University published a review paper on large language model alignment, covering more than 300 references, and providing a macro perspective of this topic. They divided LLM alignment into three categories: outer alignment, inner alignment, and mechanical…

  • AI Research Highlights | Week 39, 2023

    1. Textbooks Are All You Need II: phi-1.5 technical report Source: https://arxiv.org/abs/2309.05463 Microsoft phi-1.5 has been released, and it outperforms most state-of-the-art open-source LLMs on multi-step reasoning tasks like GSM8k, HumanEval, and MBPP when training a small 1.3B model with high-quality data. This paper calls into question the widely held…