Skip to content Skip to sidebar Skip to footer

Announcements

GraphStorm 0.3: User-friendly APIs offering scalable, multitasking learning on graphs.

GraphStorm, a low-code enterprise graph machine learning (GML) framework designed for building, training, and deploying GML solutions swiftly on complex, large-scale graphs, announces the launch of GraphStorm 0.3. The new version includes native support for multi-task learning on graphs, enabling users to define multiple training targets on different nodes and edges within a single training…

Read More

Amazon SageMaker Inference introduces enhanced auto scaling for AI generative models, improving its speed.

Amazon SageMaker has introduced a new capability that can help reduce the time it takes for the generative artificial intelligence (AI) models it supports to automatically scale. With this enhancement, the responsiveness of AI applications can be improved when demand becomes volatile. The emergence of foundation models (FMs) and large language models (LLMs) has brought…

Read More

AI chips provided by AWS ensure efficient performance and affordability for the Llama 3.1 models hosted on AWS.

Today AWS announced Trainium and Inferentia support for the Llama 3.1 models' fine-tuning and inference. The Llama 3.1 is a collection of large language models (LLMs) available in three sizes: 8B, 70B, and 405B and supports a range of capabilities such as search, image generation, code execution, and mathematical reasoning. Notably, the Llama 3.1 405B…

Read More

Amazon SageMaker introduces the fine-tuning Cohere Command R model.

Amazon Web Services (AWS) has announced the availability of the Cohere Command R fine-tuning model on Amazon SageMaker, making it the newest addition to the SageMaker suite of machine learning capabilities. The Cohere Command R model is a scalable, large language model (LLM) designed to handle enterprise-grade workloads and is optimized for conversational interaction and…

Read More

Simplify the development of generative AI in Amazon Bedrock using Prompt Management and Prompt Flows (preliminary view).

Amazon Bedrock is introducing two new features; Prompt Management and Prompt Flows. These tools are set to expedite the design, testing, and execution of generative Artificial Intelligence (AI) applications. Developers will be able to create more effective solutions, which will also be easier to maintain. As generative AI gains traction, many organizations are facing challenges in…

Read More

Improve throughput by almost 2 times and decrease expenses by about 50% for generative AI inference on Amazon SageMaker using the innovative inference optimization toolkit – Segment 1.

Amazon SageMaker has released a new inference optimization toolkit, which significantly shortens the time it takes to enhance generative artificial intelligence (AI) models. The toolkit offers several optimization techniques that can be applied to AI models and validated in just a few simple steps, ultimately reducing costs and boosting performance. It uses methods such as…

Read More

With the new inference optimization toolkit – part 2, increase your generative AI inference’s efficiency by up to 2 times on Amazon SageMaker while cutting down the expenses by almost 50%.

Amazon has launched an inference optimization toolkit as a feature of Amazon SageMaker to help speed up generative artificial intelligence (AI) operations. This tool allows businesses to balance cost-effectiveness and productivity with the use of various optimization techniques. It offers up to twice the throughput and reduces costs by up to 50% for AI models…

Read More