AI 89

Learning to Modulate pre-trained Models in RL

https://arxiv.org/abs/2306.14884 Learning to Modulate pre-trained Models in RL Reinforcement Learning (RL) has been successful in various domains like robotics, game playing, and simulation. While RL agents have shown impressive capabilities in their specific tasks, they insufficiently adapt to new tasks. In supervised learning, this arxiv.org 1. 강화 학습(RL)은 로봇 공학, 게임 플레이, 시뮬레이션 등 다양한 분야에서 성공적으로 ..

AI/Google&DeepMind 2023.06.27

Aidan Gomez 인터뷰

https://www.ft.com/content/732fc372-67ea-4684-9ab7-6b6f3cdfd736 Aidan Gomez: AI threat to human existence is ‘absurd’ distraction from real risks Co-founder of Cohere says we should be more worried about use of artificial intelligence in social media and medicine www.ft.com 코헤어의 공동 창업자이자 CEO인 에이든 고메즈는 인공지능(AI)에 대한 과도한 두려움이 실제 기술 문제로부터 우리를 방해하고 있다고 주장합니다. 그는 슈퍼인공지능(AGI)에 의한 인류 멸망에 대한 논의는 우리 시간과 공..

AI/etc 2023.06.23

ALP: 인식을 위한 행동 인식 구현 학습

https://arxiv.org/abs/2306.10190 ALP: Action-Aware Embodied Learning for Perception Current methods in training and benchmarking vision models exhibit an over-reliance on passive, curated datasets. Although models trained on these datasets have shown strong performance in a wide variety of tasks such as classification, detection, and segm arxiv.org 1. 현재의 시각 모델 학습 및 벤치마킹 방법은 수동적이고 선별된 데이터셋에 과도하게..

AI/etc 2023.06.21

대화형 AI 안전의 교차성: 베이지안 다단계 모델이 안전에 대한 다양한 인식을 이해하는 데 도움이 되는 방법

https://arxiv.org/abs/2306.11530 Intersectionality in Conversational AI Safety: How Bayesian Multilevel Models Help Understand Diverse Perceptions of Safety Conversational AI systems exhibit a level of human-like behavior that promises to have profound impacts on many aspects of daily life -- how people access information, create content, and seek social support. Yet these models have also shown..

AI/Google&DeepMind 2023.06.21

π2vec: Policy Representations with Successor Features

https://arxiv.org/abs/2306.09800 $\pi2\text{vec}$: Policy Representations with Successor Features This paper describes $\pi2\text{vec}$, a method for representing behaviors of black box policies as feature vectors. The policy representations capture how the statistics of foundation model features change in response to the policy behavior in a task agno arxiv.org 1. 이 논문은 행동의 특징 벡터로 블랙박스 정책을 표현하는..

AI/Google&DeepMind 2023.06.19