Editing Models with Task Arithmetic
What do Vision Transformers Learn? A Visual Exploration
Yuan 1.0: Large-Scale Pre-trained Language Model in Zero-Shot and Few-Shot Learning
Programming Is Hard - Or at Least It Used to Be: Educational Opportunities And Challenges of AI Code Generation
MeMViT: Memory-Augmented Multiscale Vision Transformer for Efficient Long-Term Video Recognition
DAMO-YOLO : A Report on Real-Time Object Detection Design
TorchScale: Transformers at Scale
InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions
OneFormer: One Transformer to Rule Universal Image Segmentation
Large Language Models Are Human-Level Prompt Engineers
Efficient Spatially Sparse Inference for Conditional GANs and Diffusion Models
On the Versatile Uses of Partial Distance Correlation in Deep Learning
SeaPearl: A Constraint Programming Solver guided by Reinforcement Learning
What Makes Convolutional Models Great on Long Sequence Modeling?
Amos: An Adam-style Optimizer with Adaptive Weight Decay towards Model-Oriented Scale
TabPFN: A Transformer That Solves Small Tabular Classification Problems in a Second
Long Range Graph Benchmark
Taming Transformers for High-Resolution Image Synthesis
Time Will Tell: New Outlooks and A Baseline for Temporal Multi-View 3D Object Detection
GLM-130B: An Open Bilingual Pre-trained Model
Join Podbean Ads Marketplace and connect with engaged listeners.
Advertise Today
Create your
podcast in
minutes
It is Free
Cyber Security Headlines
The WAN Show
Cybersecurity Today
Babbage from The Economist
The 404 Media Podcast