top of page

AI Research Highlights | Week 39, 2023

Updated: Feb 20


AI research highlights weekly

1. Textbooks Are All You Need II: phi-1.5 technical report

Microsoft phi-1.5 has been released, and it outperforms most state-of-the-art open-source LLMs on multi-step reasoning tasks like GSM8k, HumanEval, and MBPP when training a small 1.3B model with high-quality data. This paper calls into question the widely held belief that the capabilities of LLMs are solely determined by their scale. The model can be found at Huggingface.

phi-1.5 outperform most non-frontier LLMs on multi-step reasoning tasks
Benchmark Results

2. Explaining Agent Behavior with Large Language Models

In this paper, researchers proposed a three-stage model-agnostic method to generate natural language explanations for an agent’s behavior based only on observations of states and actions. This approach enables users to prompt an LLM to reason about agent behavior in a way that yields plausible and useful explanations and causes the least amount of hallucinations.

Explaining Agent Behavior with Large Language Models


3. Chain-of-Verification Reduces Hallucination in Large Language Models

Meta AI researchers developed Chain-of-Verification (CoVe), a method using short-form questions to reduce long-form hallucination in LLMs. CoVe first (1) drafts an initial response; then (2) plans verification questions; (3) answers those questions independently and (4) generates its final verified response. This method successfully decreases hallucinations in a range of tasks.

Chain-of-Verification (CoVe)


4. PoSE: Efficient Context Window Extension of LLMs via Positional Skip-wise Training

This paper proposed Positional Skip-wisE (PoSE) training that revolutionizes how we extend context windows of LLMs, potentially supporting infinite lengths. It successfully extends LLaMA to 128k tokens and has been verified on other LLMs (GPT-J, BaiChuan) and PI strategies (Linear, NTK, YaRN). PoSE significantly decreases memory and time overhead when compared to fine-tuning on the full length, with little to no performance impact.

 PoSE: LLMs context window extension


5. LLMR: Real-time Prompting of Interactive Worlds using Large Language Models

Researchers from MIT and Microsoft presented a Large Language Model for Mixed Reality (LLMR), a framework for the real-time creation and modification of interactive MR experiences using LLMs. LLMR can be deployed across platforms with high accuracy, satisfying its users.

real-time creation and modification of interactive MR experiences using LLMs


6. LongLoRA: Efficient Fine-tuning of Long-Context Large Language Models

Researchers from CUHK and MIT presented LongLoRA, an efficient fine-tuning approach that extends the context sizes of LLMs with limited computation cost and comparable performance to full fine-tuning.

LongLoRA extends the context sizes of LLMs


7. BTLM-3B-8K: 7B Parameter Performance in a 3B Parameter Model

BTLM-3B-8K, a 3B open-source language model, was introduced in this paper. It uses just 3GB of memory with 4-bit precision, yet it requires 2.5x less inference computation than 7B models with negligible performance impact. This contributes to the availability of a sophisticated language model on mobile and edge devices. The model can be found at: https://huggingface.co/cerebras/btlm-3b-8k-base



8. LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset

LMSYS-Chat-1M is a large-scale dataset containing 1M real-world conversations with 25 SotA LLMs, collected from 210K unique IP addresses. This dataset helps understand human-LLM interactions in real-world scenarios. It can be found at: https://huggingface.co/datasets/lmsys/lmsys-chat-1m



9. CoT-BERT: Enhancing Unsupervised Sentence Representation through Chain-of-Thought

CoT-BERT is a pioneering strategy introduced by researchers from Tsinghua University. It combines the Chain-of-Thought (CoT) process with text representation tasks, ending up in a two-stage approach for sentence representation: comprehension and summarization. CoT-BERT was shown to outperform a range of baselines and achieve state-of-the-art performance without relying on other text representation models or external databases.



10. Investigating the Catastrophic Forgetting in Multimodal Large Language Models

Catastrophic forgetting is a notorious problem for fine-tuned models, especially in multimodal large language models (MLLMs). The authors proposed EMT, the first framework to evaluate the catastrophic forgetting problem in MLLMs through classification. EMT reveals poor classification performance retention for most tested models. Additional fine-tuning experiments on LLaVA indicate that while moderate fine-tuning is beneficial, excessive fine-tuning ultimately leads to catastrophic forgetting.

EMT pipeline
EMT pipeline

11. Stabilizing RLHF through Advantage Model and Selective Rehearsal

Advantage Model and Selective Rehearsal are two novel methods developed by researchers from Tencent AI Lab in order to stabilize the RLHF training process, alleviating the issues brought on by reward hacking and catastrophic forgetting. Advantage Model aims to maintain balanced reward score distributions across diverse categories and examples. Selective Rehearsal identifies optimal examples for PPO training, encouraging the retention of crucial knowledge from the SFT stage.



12. The Reversal Curse: LLMs trained on "A is B" fail to learn "B is A"

If a model is trained on a sentence of the form "A is B", it will not automatically generalize to the reverse direction "B is A". That is the reversal curse of LLMs, pointing to a lack of meta-learning ability, regardless of their size or the questions posed.




*The researchers behind the publications deserve full credit for their work.

 





bottom of page