Wed Apr 05 2023
Tue Apr 04 2023

Summary of ChatGPT/GPT-4 Research and Perspective Towards the Future of Large Language Models

Data analytics
Language models
Natural language processing
Direct natural language processing applications
Education, history, mathematics, medicine, and physics

Presents a comprehensive survey of ChatGPT and GPT-4 and their prospective applications across diverse domains.

Offers insights into ChatGPT's capabilities, potential implications, ethical concerns, and direction for future advancements.

Rethinking the Role of Token Retrieval in Multi-Vector Retrieval

Information retrieval
Machine learning
Information retrieval
Document ranking

Presents XTR, ConteXtualized Token Retriever, which introduces a simple, yet novel, objective function that encourages the model to retrieve the most important document tokens first.

Introduces a simplified multi-vector retrieval model that advances the state-of-the-art.

TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings

Hardware systems
Machine learning
Computing systems
Machine learning workloads
Large language models

TPU v4 outperforms TPU v3 by 2.1x, and The TPU v4 pod is 4x larger at 4096 chips and thus ~10x faster overall.

Offers improved performance and energy efficiency in machine learning workloads.

LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models

Natural language processing
Artificial intelligence for text and language processing
Parameter-efficient fine-tuning techniques
Language translation
Text summarization
Sentiment analysis

This paper presents LLM-Adapters, an easy-to-use framework that integrates various adapters into LLMs for different tasks. The framework includes state-of-the-art open-access LLMs and widely used adapters, allowing the integration of new adapters and evaluation with new and larger-scale LLMs. Experiments conducted on six math reasoning datasets show that using adapter-based PEFT in smaller-scale LLMs yields comparable, and sometimes superior, performance to that of powerful LLMs in zero-shot inference on simple math reasoning datasets.

Implementing adapter-based PEFT in smaller-scale LLMs can yield comparable or even better performance than powerful LLMs in downstream tasks. This can result in cost-effective and accessible alternatives for businesses looking to fine-tune open-access LLMs with task-specific data or instruction data.

MonoAvatar: Learning Personalized High Quality Volumetric Head Avatars from Monocular RGB Videos

Virtual reality
Computer vision
3D modelling
Gaming
Virtual events
Advertising

The paper proposes a method to learn a high-quality implicit 3D head avatar from a monocular RGB video to achieve user-controlled facial expressions and head poses. The method combines the geometry prior and dynamic tracking of a 3DMM with a neural radiance field to achieve fine-grained control and photorealism. The paper also proposes to predict local features anchored on the 3DMM geometry to improve out-of-model expressions synthesis. The proposed method is compared to other state-of-the-art approaches and shows good generalization to out-of-training expressions and quantitatively superior renderings.

MonoAvatar can be used to create high-quality personalized volumetric head avatars from monocular RGB videos, which can be useful for businesses in fields that require virtual representation such as gaming, advertising, and virtual events.

Mon Apr 03 2023
Sun Apr 02 2023
Fri Mar 31 2023
Thu Mar 30 2023