Courses from 1000+ universities
Two years after its first major layoff round, Coursera announces another, impacting 10% of its workforce.
600 Free Google Certifications
Web Development
Software Development
Graphic Design
Functional Programming Principles in Scala
Mountains 101
Industrial Pharmacy-I
Organize and share your learning with Class Central Lists.
View our Lists Showcase
Explore HyperTransformer, a novel approach to few-shot learning using transformers to generate CNN weights. Learn about its architecture, advantages, and potential applications in machine learning.
Exploring how large language models can translate high-level tasks into actionable steps for virtual environments, with techniques to improve executability and potential real-world applications.
Explore connections between deep learning and neuroscience, focusing on unsupervised brain models and their potential to explain visual processing streams and representation learning in the brain.
Explore deep symbolic regression for predicting rules behind number sequences using transformers. Learn about data encoding, training processes, and applications beyond mathematics in this in-depth interview.
Explore Neural Interpreters: a modular deep learning architecture for systematic generalization. Learn about its structure, function routing, and applications in image classification and abstract reasoning.
Explore Noether Networks: a novel approach to meta-learning conserved quantities in sequential prediction problems, inspired by Noether's theorem and aimed at discovering useful symmetries and inductive biases.
Explore LaMa, an advanced image inpainting system using Fourier convolutions for large mask removal. Learn about its architecture, loss function, and impressive results in reconstructing complex structures.
Explore NÃœWA, a unified multimodal model for visual synthesis tasks. Learn about its 3D transformer framework, 3D Nearby Attention mechanism, and applications in text-to-image/video generation and manipulation.
Detailed explanation of Scaling Transformers and Terraformer architecture, focusing on leveraging sparsity to improve efficiency and speed in large language models while maintaining accuracy.
Explores ExT5, an advanced NLP model pre-trained on 107 diverse tasks. Analyzes multi-task scaling benefits, task co-training effects, and demonstrates ExT5's superior performance across various NLP benchmarks.
Explore parameter prediction for neural networks using graph hypernetworks. Learn about DeepNets-1M dataset, training techniques, and experimental results for efficient network training paradigms.
Explore grafting technique for transferring learning rate schedules between optimizers, improving deep learning model performance and reducing computational costs in hyperparameter tuning.
Explores limitations of differentiable programming in machine learning, focusing on chaos-based failures in various systems. Discusses alternatives to backpropagation for gradient estimation in complex, stochastic environments.
Explore Autoregressive Diffusion Models, a novel approach combining autoregressive and diffusion models for efficient, order-agnostic generation and compression of text and image data.
Explore EfficientZero, a groundbreaking reinforcement learning algorithm achieving human-level performance on Atari games with minimal data, outperforming previous methods in sample efficiency and performance.
Get personalized course recommendations, track subjects and courses with reminders, and more.