Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

Inside COSP and USP: Google Research New Methods to Advance Reasoning in LLMs

Jan 25, 2024 - pub.towardsai.net
The article discusses two new methods introduced by Google Research, Consistency-Based Self-Adaptive Prompting (COSP) and Universal Self-Adaptive Prompting (USP), that aim to improve common sense reasoning capabilities in large language models (LLMs). COSP leverages unlabeled samples and the model’s own predictions to generate suitable prompts, bridging the performance gap between zero-shot and few-shot while preserving the advantages of zero-shot prompting. USP extends this concept to a wide array of natural language understanding and generation tasks, showcasing its effectiveness across various domains.

Both methods utilize the model’s zero-shot outputs as demonstrations for prompting itself, selecting reliable self-generated demonstrations based on confident and consistent model predictions. The methods were evaluated across different benchmarks, with results showing that both COSP and USP outperform standard zero-shot baselines and remain competitive when compared to prompting with golden examples. The research reinforces the efficacy of these methods in enhancing the performance of language models across diverse natural language understanding and generation tasks.

Key takeaways:

  • Google Research has introduced two new methods, Consistency-Based Self-Adaptive Prompting (COSP) and Universal Self-Adaptive Prompting (USP), to enhance common sense reasoning capabilities in large language models (LLMs).
  • COSP and USP utilize the model’s zero-shot outputs as demonstrations for prompting itself, selecting reliable self-generated demonstrations based on confident and consistent model predictions.
  • USP extends the approach to a wider array of natural language processing tasks, including classification, short-form generation, and long-form generation, adapting the confidence measurement techniques accordingly.
  • Both COSP and USP have shown to outperform baseline methods in various tasks and benchmarks, demonstrating their effectiveness in enhancing the performance of language models across diverse natural language understanding and generation tasks.
View Full Article

Comments (0)

Be the first to comment!