LLaMA: Open and Efficient Foundation Language Models
Dropout Reduces Underfitting
Cross-domain Compositing with Pretrained Diffusion Models
REaLTabFormer: Generating Realistic Relational and Tabular Data using Transformers
Large-scale Multi-Modal Pre-trained Models: A Comprehensive Survey
Fine-Tuning Language Models from Human Preferences
AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities
Demonstrate-Search-Predict: Composing retrieval and language models for knowledge-intensive NLP
Mastering Diverse Domains through World Models
Adding Conditional Control to Text-to-Image Diffusion Models
A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech
Token Merging: Your ViT But Faster
BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining
Dual PatchNorm
Reversible Vision Transformers
Offsite-Tuning: Transfer Learning without Full Model
A Length-Extrapolatable Transformer
NaturalSpeech: End-to-End Text to Speech Synthesis with Human-Level Quality
Multimodal Chain-of-Thought Reasoning in Language Models
BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models
Join Podbean Ads Marketplace and connect with engaged listeners.
Advertise Today
Create your
podcast in
minutes
It is Free
Cyber Security Headlines
The WAN Show
The 404 Media Podcast
Babbage from The Economist
Big Technology Podcast