InfoQ Homepage Neural Networks Content on InfoQ
-
Microsoft Satin Audio Codec Uses AI to Outperform Skype Silk
Microsoft announced Satin, a new audio codec that leverages AI techniques to outperform Skype's Silk codec over ultra-low bandwidth and highly constrained network conditions.
-
Google Open-Sources Trillion-Parameter AI Language Model Switch Transformer
Researchers at Google Brain have open-sourced the Switch Transformer, a natural-language processing (NLP) AI model. The model scales up to 1.6T parameters and improves training time up to 7x compared to the T5 NLP model, with comparable accuracy.
-
OpenAI Announces GPT-3 Model for Image Generation
OpenAI has trained a 12B-parameter AI model based on GPT-3 that can generate images from textual description. The description can specify many independent attributes, including the position of objects as well as image perspective, and can also synthesize combinations of objects that do not exist in the real world.
-
Facebook Open-Sources Multilingual Speech Recognition Deep-Learning Model
Facebook AI Research (FAIR) open-sourced Cross-Lingual Speech Recognition (XSLR), a multilingual speech recognition AI model. XSLR is trained on 53 languages and outperforms existing systems when evaluated on common benchmarks.
-
TensorFlow 2.4 Release Includes CUDA 11 Support and API Updates
The TensorFlow project announced the release of version 2.4.0 of the deep-learning framework, featuring support for CUDA 11, cuDNN 8, and NVIDIA's Ampere GPU architecture, as well as new strategies and profiling tools for distributed training. Other API updates include mixed-precision in Keras and a NumPy frontend.
-
AI Models from Google and Microsoft Exceed Human Performance on Language Understanding Benchmark
Research teams from Google and Microsoft have recently developed natural language processing (NLP) AI models which have scored higher than the human baseline score on the SuperGLUE benchmark. SuperGLUE measures a model's score on several natural language understanding (NLU) tasks, including question answering and reading comprehension.
-
DeepMind's AlphaFold2 AI Solves 50-Year-Old Biology Challenge
The Protein Structure Prediction Center announced that AlphaFold2, an AI system developed by DeepMind, has solved its Protein Structure Prediction challenge. AlphaFold2 achieved a median score of 92.4 on the Global Distance Test (GDT) metric, above the threshold considered competitive with traditional methods.
-
Caltech Open-Sources AI for Solving Partial Differential Equations
Researchers from Caltech's DOLCIT group have open-sourced Fourier Neural Operator (FNO), a deep-learning method for solving partial differential equations (PDEs). FNO outperforms other existing deep-learning techniques for solving PDEs and is three orders of magnitude faster than traditional solvers.
-
Google Open-Sources Fast Attention Module Performer
Google has open-sourced Performer, a Transformer deep-learning architecture that scales linearly with input sequence length. This allows Performer to be used for tasks that require long sequences, including pixel-prediction and protein sequence modeling.
-
NVIDIA's AI Reduces Video Streaming Bandwidth Consumption by 10x
GPU-manufacturer NVIDIA announced their Maxine platform for AI-enhanced video-conferencing services, which includes a technology that can reduce bandwidth requirements by an order of magnitude. By moving much of the data processing to the cloud, end-users can take advantage of the compression without needing specialized hardware.
-
NVIDIA Releases a $59 Jetson Nano 2GB Kit to Make AI More Accessible to Developers
With the Jetson series of devices and software SDKs, NVIDIA creates a coherent development environment to learn and develop GPU-based AI applications.
-
AI Training Method Exceeds GPT-3 Performance with 99.9% Fewer Parameters
A team of scientists at LMU Munich have developed Pattern-Exploiting Training (PET), a deep-learning training technique for natural language processing (NLP) models. Using PET, the team trained a Transformer NLP model with 223M parameters that out-performed the 175B-parameter GPT-3 by over 3 percentage points on the SuperGLUE benchmark.
-
Facebook Releases AI Model for Protein Sequence Processing
A team of scientists at Facebook AI Research have released a deep-learning model for processing protein data from DNA sequences. The model contains approximately 700M parameters, was trained on 250 million protein sequences, and learned representations of biological properties that can be used to improve current state-of-the-art in several genomics prediction tasks.
-
Salesforce Releases Photon Natural Language Interface for Databases
A team of scientists from Salesforce Research and Chinese University of Hong Kong have released Photon, a natural language interface to databases (NLIDB). The team used deep-learning to construct a parser that achieves 63% accuracy on a common benchmark and an error-detecting module that prompts users to clarify ambiguous questions.
-
Artificial Intelligence Can Create Sound Tracks for Silent Videos
Researchers Ghose and Prevost created a deep learning algorithm which, given a silent video, can generate a realistic sounding synchronised soundtrack. They trained the neural network to classify the class of the sound to generate, and they also trained a sequential network to generate the sound. They thus could go from temporally aligned images to the generation of sound: a different modality!