AI/Google&DeepMind 29

Block-State Transformer

https://arxiv.org/abs/2306.09539 Block-State Transformer State space models (SSMs) have shown impressive results on tasks that require modeling long-range dependencies and efficiently scale to long sequences owing to their subquadratic runtime complexity. Originally designed for continuous signals, SSMs have sho arxiv.org 1. 상태 공간 모델(SSM)은 장기적인 의존성을 모델링하는 데 필요한 작업에서 뛰어난 결과를 보여주며, 서브쿼드럴릭 실행 시간 복잡..

AI/Google&DeepMind 2023.06.19

이미지 캡셔너는 확장 가능한 비전 학습자이기도 합니다.

https://arxiv.org/abs/2306.07915 Image Captioners Are Scalable Vision Learners Too Contrastive pretraining on image-text pairs from the web is one of the most popular large-scale pretraining strategies for vision backbones, especially in the context of large multimodal models. At the same time, image captioning on this type of data is co arxiv.org 1. 웹에서 이미지-텍스트 쌍에 대한 대조적 사전학습은, 특히 대형 다중모달 모델의 맥..

AI/Google&DeepMind 2023.06.14

딥마인드 AlphaDev: 새로운 접근법으로 더 빠른 정렬 알고리즘 발견

https://www.deepmind.com/blog/alphadev-discovers-faster-sorting-algorithms AlphaDev discovers faster sorting algorithms In our paper published today in Nature, we introduce AlphaDev, an artificial intelligence (AI) system that uses reinforcement learning to discover enhanced computer science algorithms – surpassing those honed by scientists and engineers over decades. www.deepmind.com DeepMind의 ..

AI/Google&DeepMind 2023.06.08

SQL-PaLM: Text-to-SQL을 위한 개선된 대규모 언어 모델 적응

abs: https://arxiv.org/abs/2306.00739 SQL-PaLM: Improved Large Language ModelAdaptation for Text-to-SQL One impressive emergent capability of large language models (LLMs) is generation of code, including Structured Query Language (SQL) for databases. For the task of converting natural language text to SQL queries, Text-to-SQL, adaptation of LLMs is of paramo arxiv.org 1. 대형 언어 모델(LLM)의 놀라운 능력 중 ..

AI/Google&DeepMind 2023.06.02

브레인포머: 효율성을 위한 거래 단순성

abs: https://arxiv.org/abs/2306.00008 Brainformers: Trading Simplicity for Efficiency Transformers are central to recent successes in natural language processing and computer vision. Transformers have a mostly uniform backbone where layers alternate between feed-forward and self-attention in order to build a deep network. Here we investigat arxiv.org 1. 트랜스포머는 자연어 처리와 컴퓨터 비전의 최근 성공에 중심 역할을 하는데, ..

AI/Google&DeepMind 2023.06.02

더 크게, 더 좋게, 더 빠르게: 인간 수준의 효율성을 갖춘 인간 수준의 Atari

https://arxiv.org/abs/2305.19452 Bigger, Better, Faster: Human-level Atari with human-level efficiency We introduce a value-based RL agent, which we call BBF, that achieves super-human performance in the Atari 100K benchmark. BBF relies on scaling the neural networks used for value estimation, as well as a number of other design choices that enable this sca arxiv.org https://github.com/google-re..

AI/Google&DeepMind 2023.06.01

PaLI-X: 다국어 비전 및 언어 모델 확장

https://arxiv.org/abs/2305.18565 PaLI-X: On Scaling up a Multilingual Vision and Language Model We present the training recipe and results of scaling up PaLI-X, a multilingual vision and language model, both in terms of size of the components and the breadth of its training task mixture. Our model achieves new levels of performance on a wide-range of arxiv.org PaLI-X라는 다국어 비전 및 언어 모델을 확대 규모로 훈련하..

AI/Google&DeepMind 2023.05.31

온라인 비확률적 모델 없는 강화학습

abs: https://arxiv.org/abs/2305.17552 Online Nonstochastic Model-Free Reinforcement Learning In this work, we explore robust model-free reinforcement learning algorithms for environments that may be dynamic or even adversarial. Conventional state-based policies fail to accommodate the challenge imposed by the presence of unmodeled disturbances in arxiv.org 1. 이 연구에서는 동적이거나 적대적일 수 있는 환경에 대한 강건한 모..

AI/Google&DeepMind 2023.05.30

Q-러닝에 대한 보다 효율적인 대안인 VA-러닝

abs: https://arxiv.org/abs/2305.18161 VA-learning as a more efficient alternative to Q-learning In reinforcement learning, the advantage function is critical for policy improvement, but is often extracted from a learned Q-function. A natural question is: Why not learn the advantage function directly? In this work, we introduce VA-learning, which dire arxiv.org 1. 강화 학습에서 이점 함수는 정책 개선에 있어 중요하지만, ..

AI/Google&DeepMind 2023.05.30