GraphStorm, a low-code enterprise graph machine learning (GML) framework designed for building, training, and deploying GML solutions swiftly on complex, large-scale graphs, announces the launch of GraphStorm 0.3. The new version includes native support for multi-task learning on graphs, enabling users to define multiple training targets on different nodes and edges within a single training…
Amazon SageMaker has introduced a new capability that can help reduce the time it takes for the generative artificial intelligence (AI) models it supports to automatically scale. With this enhancement, the responsiveness of AI applications can be improved when demand becomes volatile. The emergence of foundation models (FMs) and large language models (LLMs) has brought…
Today AWS announced Trainium and Inferentia support for the Llama 3.1 models' fine-tuning and inference. The Llama 3.1 is a collection of large language models (LLMs) available in three sizes: 8B, 70B, and 405B and supports a range of capabilities such as search, image generation, code execution, and mathematical reasoning. Notably, the Llama 3.1 405B…
Amazon Web Services (AWS) has announced the availability of the Cohere Command R fine-tuning model on Amazon SageMaker, making it the newest addition to the SageMaker suite of machine learning capabilities. The Cohere Command R model is a scalable, large language model (LLM) designed to handle enterprise-grade workloads and is optimized for conversational interaction and…
Amazon Bedrock is introducing two new features; Prompt Management and Prompt Flows. These tools are set to expedite the design, testing, and execution of generative Artificial Intelligence (AI) applications. Developers will be able to create more effective solutions, which will also be easier to maintain.
As generative AI gains traction, many organizations are facing challenges in…
Amazon SageMaker has released a new inference optimization toolkit, which significantly shortens the time it takes to enhance generative artificial intelligence (AI) models. The toolkit offers several optimization techniques that can be applied to AI models and validated in just a few simple steps, ultimately reducing costs and boosting performance. It uses methods such as…
Amazon has launched an inference optimization toolkit as a feature of Amazon SageMaker to help speed up generative artificial intelligence (AI) operations. This tool allows businesses to balance cost-effectiveness and productivity with the use of various optimization techniques. It offers up to twice the throughput and reduces costs by up to 50% for AI models…