AI 89

온라인 비확률적 모델 없는 강화학습

abs: https://arxiv.org/abs/2305.17552 Online Nonstochastic Model-Free Reinforcement Learning In this work, we explore robust model-free reinforcement learning algorithms for environments that may be dynamic or even adversarial. Conventional state-based policies fail to accommodate the challenge imposed by the presence of unmodeled disturbances in arxiv.org 1. 이 연구에서는 동적이거나 적대적일 수 있는 환경에 대한 강건한 모..

AI/Google&DeepMind 2023.05.30

네이버가 보는 AGI 는?

https://themiilk.com/articles/a6946e73d?u=15ed0645&t=abef9db15&from&fbclid=IwAR1I6NThOH6Dnb3hrL-ZG_3qbFfC8QEs2dgzaHmwGhNSBkkXIEetHED7_lQ AI 휩쓰는 미국 빅테크에 '디지털 영토' 또 내줄 것인가? - 더밀크 [네이버 클라우드 AI이노베이션 하정우 센터장 인터뷰] ●일본, 동남아, 중동 등 글로벌 시장, 소버린(sovereign) AI서비스 및 인프라 구축 계획 하는 네이버. ●챗GPT등 기존 빅테크 서비스에 종속 막 themiilk.com 하 센터장은 AI가 발전한다면 AGI가 이론적으로는 가능은 하지만, ‘지능(intelligence)를 어떻게 정의하느냐?’에 따라서 달라질 수도 있다. 또한 언..

AI/etc 2023.05.30

마음의 눈 재구성: fMRI-to-Image with Contrastive Learning 및 Diffusion Priors

https://arxiv.org/abs/2305.18274 github: https://medarc-ai.github.io/mindeye/ Reconstructing the Mind's Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors We present MindEye, a novel fMRI-to-image approach to retrieve and reconstruct viewed images from brain activity. Our model comprises two parallel submodules that are specialized for retrieval (using contrastive learning) and re..

AI/etc 2023.05.30

Q-러닝에 대한 보다 효율적인 대안인 VA-러닝

abs: https://arxiv.org/abs/2305.18161 VA-learning as a more efficient alternative to Q-learning In reinforcement learning, the advantage function is critical for policy improvement, but is often extracted from a learned Q-function. A natural question is: Why not learn the advantage function directly? In this work, we introduce VA-learning, which dire arxiv.org 1. 강화 학습에서 이점 함수는 정책 개선에 있어 중요하지만, ..

AI/Google&DeepMind 2023.05.30

Ghost in the Minecraft: 텍스트 기반 지식 및 메모리를 갖춘 대규모 언어 모델을 통해 일반적으로 오픈 월드 환경에 사용할 수 있는 에이전트

abs: https://arxiv.org/abs/2305.17144 github: https://github.com/OpenGVLab/GITM Ghost in the Minecraft: Generally Capable Agents for Open-World Enviroments via Large Language Models with Text-based Knowledge The captivating realm of Minecraft has attracted substantial research interest in recent years, serving as a rich platform for developing intelligent agents capable of functioning in open-wo..

AI/etc 2023.05.30

대규모 언어 모델은 게으른 학습자가 될 수 있습니다: 상황 내 학습의 지름길 분석

abs: https://arxiv.org/abs/2305.17256 Large Language Models Can be Lazy Learners: Analyze Shortcuts in In-Context Learning Large language models (LLMs) have recently shown great potential for in-context learning, where LLMs learn a new task simply by conditioning on a few input-label pairs (prompts). Despite their potential, our understanding of the factors influencing end-tas arxiv.org 1. 대규모 언..

AI/etc 2023.05.30