AI/etc 48

언어 모델을 사용한 인증된 추론

https://arxiv.org/abs/2306.04031 Certified Reasoning with Language Models Language models often achieve higher accuracy when reasoning step-by-step in complex tasks. However, their reasoning can be unsound, inconsistent, or rely on undesirable prior assumptions. To tackle these issues, we introduce a class of tools for language arxiv.org 1. 언어 모델은 복잡한 작업에서 단계별로 추론을 할 때 종종 더 높은 정확도를 보입니다. 그러나 그들의..

AI/etc 2023.06.08

생각 복제: 인간의 생각을 모방하여 행동하면서 생각하는 법 배우기

https://arxiv.org/abs/2306.00323 Thought Cloning: Learning to Think while Acting by Imitating Human Thinking Language is often considered a key aspect of human thinking, providing us with exceptional abilities to generalize, explore, plan, replan, and adapt to new situations. However, Reinforcement Learning (RL) agents are far from human-level performance in any arxiv.org https://twitter.com/jef..

AI/etc 2023.06.03

인간인가 아닌가? 튜링 테스트에 대한 게임화된 접근 방식

https://arxiv.org/abs/2305.20010 Human or Not? A Gamified Approach to the Turing Test We present "Human or Not?", an online game inspired by the Turing test, that measures the capability of AI chatbots to mimic humans in dialog, and of humans to tell bots from other humans. Over the course of a month, the game was played by over 1.5 million arxiv.org 1. 우리는 튜링 테스트에 영감을 받은 온라인 게임인 "Human or Not?"..

AI/etc 2023.06.01

데이터 제약이 있는 언어모델 확장

Chinchilla 스케일링 법칙 확장 설명: https://twitter.com/Muennighoff/status/1661895337248686081 트위터에서 즐기는 Niklas Muennighoff “How to keep scaling Large Language Models when data runs out? 🎢 We train 400 models with up to 9B params & 900B tokens to create an extension of Chinchilla scaling laws for repeated data. Results are interesting… 🧐 📜: https://t.co/586bWwvpba twitter.com 1. 이 연구에서는 데이터 제한 조건에서 언어 모델을..

AI/etc 2023.06.01

긴 컨텍스트 대형 모델을 위한 블록별 병렬 트랜스포머

abs: https://arxiv.org/abs/2305.19370 Blockwise Parallel Transformer for Long Context Large Models Transformers have emerged as the cornerstone of state-of-the-art natural language processing models, showcasing exceptional performance across a wide range of AI applications. However, the memory demands posed by the self-attention mechanism and the large arxiv.org 1. 트랜스포머 모델은 다양한 AI 응용 분야에서 최첨단 자..

AI/etc 2023.06.01

LIV: 로봇 제어를 위한 언어-이미지 표현 및 보상

https://penn-pal-lab.github.io/LIV/ LIV LIV as Representation for Language-Conditioned BC We use LIV's frozen multi-modal representation as backbone for LCBC and achieve impressive performance (46% success rate, absolute ~30% better than the second best baseline) on a challenging real-world muli penn-pal-lab.github.io 설명: https://twitter.com/JasonMa2020/status/1663618652778942464 트위터에서 즐기는 Jason..

AI/etc 2023.05.31