Observation of Gravitational Waves from a Binary Black Hole Merger

This paper reports the first direct detection of gravitational waves on September 14, 2015, originating from the merger of a binary stellar-mass black hole system, a discovery that confirms the existence of such binaries and validates key predictions of general relativity.

# The Universe's First "Scream": A Simple Explanation of GW150914 Imagine the universe as a giant, invisible trampoline made of space and time.…

The LIGO Scientific Collaboration, the Virgo Collaboration2016-02-11⚛️ gr-qc

First M87 Event Horizon Telescope Results. I. The Shadow of the Supermassive Black Hole

The Event Horizon Telescope collaboration successfully imaged the supermassive black hole at the center of galaxy M87, revealing a bright asymmetric ring surrounding a dark shadow consistent with general relativity predictions and enabling a precise mass measurement of approximately 6.5 billion solar masses.

Imagine trying to take a photograph of a ghost that is so heavy it bends the very fabric of space around it, but the ghost itself is invisible because it swallows all light.…

The Event Horizon Telescope Collaboration2019-06-26🔭 astro-ph.GA

Deep Residual Learning for Image Recognition

This paper introduces a residual learning framework that reformulates network layers to learn residual functions, enabling the successful training of extremely deep neural networks (up to 152 layers) that significantly outperform previous models and achieved first place in multiple 2015 computer vision competitions.

Imagine you are trying to teach a student how to solve a very complex math problem.…

Kaiming He, Xiangyu Zhang, Shaoqing Ren + 1 more2015-12-10👁️ cs.CV

Generative Adversarial Networks

This paper proposes a new generative modeling framework based on a minimax two-player game between a generative model that captures data distribution and a discriminative model that distinguishes real from generated samples, which can be trained efficiently using backpropagation without Markov chains.

Imagine you are trying to teach a computer how to create art, music, or realistic photos.…

Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza + 5 more2014-06-10📊 stat.ML

Adam: A Method for Stochastic Optimization

This paper introduces Adam, an efficient and memory-light stochastic optimization algorithm that uses adaptive moment estimates to effectively handle large-scale, noisy, and sparse gradient problems while offering strong theoretical convergence guarantees and competitive empirical performance.

Imagine you are trying to find the lowest point in a vast, foggy mountain range (the "optimal solution" for a machine learning problem).…

Diederik P. Kingma, Jimmy Ba2014-12-22🤖 cs.LG

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

This paper introduces Batch Normalization, a technique that accelerates deep neural network training and improves performance by normalizing layer inputs to reduce internal covariate shift, thereby allowing for higher learning rates, better initialization robustness, and reduced reliance on regularization methods like Dropout.

Imagine you are trying to teach a team of 20 chefs (a **Deep Neural Network**) to cook a perfect steak.…

Sergey Ioffe, Christian Szegedy2015-02-11🤖 cs.LG

Auto-Encoding Variational Bayes

This paper introduces the Auto-Encoding Variational Bayes (AEVB) framework, which enables efficient stochastic variational inference and learning in directed probabilistic models with continuous latent variables by utilizing a reparameterized lower bound estimator and an approximate inference model to handle intractable posteriors and scale to large datasets.

Imagine you are trying to teach a robot to understand a massive library of books.…

Diederik P Kingma, Max Welling2013-12-20📊 stat.ML

Improving neural networks by preventing co-adaptation of feature detectors

This paper introduces the "dropout" technique, which randomly omits feature detectors during training to prevent complex co-adaptations and overfitting, thereby significantly improving neural network performance on tasks like speech and object recognition.

Imagine you are trying to teach a class of students how to recognize different animals.…

Geoffrey E. Hinton, Nitish Srivastava, Alex Krizhevsky + 2 more2012-07-03💻 cs.NE

Efficient Estimation of Word Representations in Vector Space

This paper introduces two novel, computationally efficient model architectures for learning high-quality continuous word vector representations from massive datasets, which achieve state-of-the-art performance in measuring syntactic and semantic word similarities at a fraction of the previous computational cost.

Imagine you are trying to teach a computer how to understand human language.…

Tomas Mikolov, Kai Chen, Greg Corrado + 1 more2013-01-16💬 cs.CL