NLP – PyTorch https://pytorch.org Mon, 10 Feb 2025 06:49:04 +0000 en-US hourly 1 https://wordpress.org/?v=6.8.2 https://pytorch.org/wp-content/uploads/2024/10/cropped-favicon-32x32.webp NLP – PyTorch https://pytorch.org 32 32 The Why and How of Scaling Large Language Models https://www.youtube.com/watch?v=qscouq3lo0s Tue, 04 Jan 2022 05:43:00 +0000 https://pytorch.org/?p=2999 Anthropic is an AI safety and research company that’s working to build reliable, interpretable, and steerable AI systems. Over the past decade, the amount of compute used for the largest training runs has increased at an exponential pace. We’ve also seen in many domains that larger models are able to attain better performance following precise scaling laws. The compute needed to train these models can only be attained using many coordinated machines that are communicating data between them.

]]>
Using a Grapheme to Phoneme Model in Cisco’s Webex Assistant https://blogs.cisco.com/developer/graphemephoneme01 Tue, 07 Sep 2021 05:52:00 +0000 https://pytorch.org/?p=3007 Grapheme to Phoneme (G2P) is a function that generates pronunciations (phonemes) for words based on their written form (graphemes). It has an important role in automatic speech recognition systems, natural language processing, and text-to-speech engines. In Cisco’s Webex Assistant, we use G2P modelling to assist in resolving person names from voice. See here for further details of various techniques we use to build robust voice assistants.

]]>
University of Pécs enables text and speech processing in Hungarian, builds the BERT-large model with just 1,000 euro with Azure https://www.microsoft.com/en/customers/story/1402696956382669362-university-of-pecs-higher-education-azure-en-hungary Tue, 10 Aug 2021 05:54:00 +0000 https://pytorch.org/?p=3011 Everyone prefers to use their mother tongue when communicating with chat agents and other automated services. However, for languages like Hungarian—spoken by only 15 million people—the market size will often be viewed as too small for large companies to create software, tools or applications that can process Hungarian text as input. Recognizing this need, the Applied Data Science and Artificial Intelligence team from University of Pécs decided to step up. 

]]>
AI21 Labs Trains 178-Billion-Parameter Language Model Using Amazon EC2 P4d Instances, PyTorch https://aws.amazon.com/solutions/case-studies/AI21-case-study-p4d/ Mon, 07 Jun 2021 05:57:00 +0000 https://pytorch.org/?p=3015 AI21 Labs uses machine learning to develop language models focused on understanding meaning, and in 2021 it set a goal to train the recently released Jurassic-1 Jumbo, an autoregressive language model with 178 billion parameters. Developers who register for beta testing will get access to Jurassic-1 Jumbo and can immediately start to customize the model for their use case. The software startup wanted to train the model efficiently, so it looked to Amazon Web Services (AWS) 

]]>
Deepset achieves a 3.9x speedup and 12.8x cost reduction for training NLP models by working with AWS and NVIDIA https://aws.amazon.com/blogs/machine-learning/deepset-achieves-a-3-9x-speedup-and-12-8x-cost-reduction-for-training-nlp-models-by-working-with-aws-and-nvidia/ Wed, 27 Jan 2021 06:33:00 +0000 https://pytorch.org/?p=3029 At deepset, we’re building the next-level search engine for business documents. Our core product, Haystack, is an open-source framework that enables developers to utilize the latest NLP models for semantic search and question answering at scale. Our software as a service (SaaS) platform, Haystack Hub, is used by developers from various industries, including finance, legal, and automotive, to find answers in all kinds of text documents. 

]]>
PyTorch at Dolby Labs https://www.youtube.com/watch?v=K5hD0et_wUc&list=PL_lsbAsL_o2BY-RrqVDKDcywKnuUTp-f3&index=20 Wed, 06 Nov 2019 06:48:00 +0000 https://pytorch.org/?p=3059 Hear how Dolby Labs is using PyTorch to develop deep learning for audio, and learn about the challenges that audio AI presents and the breakthroughs and applications they’ve built at Dolby to push the field forward.

]]>