Follow Datanami:

Tag: GPT-3

Conversational AI Poised to Be Major Disrupter

Chatbots and conversational AI systems got an extended tryout during COVID as companies scrambled for ways to keep their operations running amid lockdowns. The technology fared better than expected, and now is on the cus Read more…

IBM Collaboration Looks to Bring Massive AI Models to Any Cloud

Training machine learning foundation models with sometimes billions of parameters demands serious computing power. For example, the largest version of GPT-3, the famous large language model behind OpenAI’s DALL-E 2, ha Read more…

Data Intelligence Platform Seek AI Launches to Automate Repetitive Tasks

With all the hype around generative AI and its text-to-image capabilities throwing the art world for a loop, it may be easy to forget that you can write code with it, too. Seek AI, a new data intelligence platform tha Read more…

OpenAI’s DALL·E 2 Is Surreal

About 15 months ago, OpenAI—famed for its eerily effective GPT-3 large language model—introduced a child system to that language model: the cleverly named “DALL·E,” a 12-billion parameter neural network that gen Read more…

Google’s Massive New Language Model Can Explain Jokes

Nearly two years ago, OpenAI’s 175 billion-parameter GPT-3 language model opened the world’s eyes to what large language models (LLMs) could accomplish with relatively little input, sensibly answering questions, tran Read more…

d-Matrix Gets Funding to Build SRAM ‘Chiplets’ for AI Inference

Hardware startup d-Matrix says the $44 million it raised in a Series A round today will help it continue development of a novel “chiplet” architecture that uses 6 nanometer chip embedded in SRAM memory modules for ac Read more…

Mantium Lowers the Barrier to Using Large Language Models

Large language models like GPT-3 are bringing powerful AI capabilities to organizations around the world, but putting them into production in a secure and responsible manner can be difficult. Now a company called Mantium Read more…

Cerebras Hits the Accelerator for Deep Learning Workloads

When it comes to large neural networks like BERT or GPT-3, organizations often must wait weeks or even months for a training task to complete if they’re using traditional CPU and GPU clusters. But with its massive Wafe Read more…

An All-Volunteer Deep Learning Army

A team of researchers says their novel framework, called Distributed Deep Learning in Open Collaborations, or DeDLOC, brings the potential to train large deep learning models from scratch in a distributed, grid-like mann Read more…

Inside eBay’s Optimization Techniques for Scaling AI

Getting the software right is important when developing machine learning models, such as recommendation or classification systems. But at eBay, optimizing the software to run on a particular piece of hardware using disti Read more…

Unlocking the True Potential of ML: How Self-Supervised Learning in Language Can Beat Human Performance

A core goal for many organizations using artificial intelligence (AI) systems is to have them mirror human language and intelligence. However, mimicking human language and mastering its unique complexities continues to b Read more…

Deci Shows a NAC for Automated Neural Net Construction

Deep learning researchers have been dancing around a looming performance wall in recent months, as huge neural networks push the limits in terms of computation and power consumption. But now a company called Deci says it Read more…

The Perfect Storm: How the Chip Shortage Will Impact AI Development

This chip shortage has brought to light our dependency on hardware to run high-tech economies and the everyday lives of consumers. Today, chips can be found in everything from gaming consoles like Xbox Series, PlayStatio Read more…

Google’s ‘Breakthrough’ LaMDA Promises to Elevate the Common Chatbot

Many of Google’s language processing efforts – like BERT and, more recently, MUM – are focused on returning search queries. But as Google moves more toward Assistant – and search queries in general become more de Read more…

Experts Disagree on the Utility of Large Language Models

Large language models like OpenAI’s GPT-3 and Google Brain’s Switch Transformer have caught the eye of AI experts, who have expressed surprise at the rapid pace of improvement. However, not everybody is jumping onto the bandwagon, and others see significant limitations in the new technology, as well as ethical implications. Read more…

AI Experts Discuss Implications of GPT-3

Last July, GPT-3 took the internet by storm. The massive 175 billion-parameter autoregressive language model, developed by OpenAI, showed a startling ability to translate languages, answer questions, and – perhaps most Read more…

One Model to Rule Them All: Transformer Networks Usher in AI 2.0, Forrester Says

The recent advent of massive transformer networks is ushering in a new age of AI that will give customers advanced natural language capabilities with just a fraction of the skills and data previously required, according Read more…

Google’s New Switch Transformer Model Achieves 1.6 Trillion Parameters, Efficiency Gains

Last year, OpenAI wowed the world with its eerily human language generator, GPT-3. The autoregressive model stood at a then-staggering 175 billion parameters, ten times higher than its predecessors. Now, Google is upping Read more…

Low-Code Can Lower the Barrier to Entry for AI

Organizations that want to get started quickly with machine learning may be interested in investigating emerging low-code options for AI. While low-code techniques will never completely replace hand-coded systems, they c Read more…

OpenAI’s GPT-3 Language Generator Is Impressive, but Don’t Hold Your Breath for Skynet

Just a couple of months after releasing a paper describing GPT-3, AI development and deployment company OpenAI has begun testing its novel, AI-powered language generator with a select group of users through a private bet Read more…

Datanami