compute-intensive Archives - 成人VR视频 Institute https://blogs.thomsonreuters.com/en-us/innovation-topics/compute-intensive/ 成人VR视频 Institute is a blog from 成人VR视频, the intelligence, technology and human expertise you need to find trusted answers. Tue, 17 Sep 2024 14:10:57 +0000 en-US hourly 1 https://wordpress.org/?v=6.8.3 成人VR视频 Labs: Training large language models using Amazon SageMaker HyperPod /en-us/posts/innovation/thomson-reuters-labs-training-large-language-models-using-amazon-sagemaker-hyperpod/ Tue, 17 Sep 2024 09:12:44 +0000 https://blogs.thomsonreuters.com/en-us/?post_type=innovation_post&p=63089 2023聽proved to be聽an inflection point for AI, prompting 成人VR视频 to consider how our聽high-value, curated, data could improve general language models on customer-specific tasks. Training and finetuning a large language model (LLM) is compute-intensive and requires specialized hardware.

quickly discovered that it was extremely difficult to acquire these resources on-demand and at scale in our cloud environments. Further, looking to other third parties presented its own set of risks and challenges.

We turned to聽(AWS), which has long been a trusted partner in secure and scalable solutions, to get early access to . With our computing platform acquired, we were ready to roll up our sleeves and do the hard work of exploring how to optimally train and finetune models to our domain. In our first phase of experimentation, we peaked at 16聽 compute instances, 128 A100 GPUs, with the longest job taking 36 days to complete training a 70 billion parameter model.

Initial results of our custom models look promising and our research continues, supported by the release of . Our explores the journey that 成人VR视频 took to enable cutting-edge research in training domain-adapted LLMs using聽Amazon SageMaker HyperPod.

This is a guest post from John Duprey, distinguished engineer, 成人VR视频.

]]>