|
Clarifai Launches First Vendor-Agnostic Compute Orchestration Capabilities to Optimize AI Performance and Spend
Wednesday, December 11, 2024
Customers Can Orchestrate AI Workloads Better, Avoid Vendor Lock-In, and Use Compute Spend Efficiently
WASHINGTON, Dec. 3, 2024 /PRNewswire/ -- Clarifai, a global leader in AI and pioneer of the full-stack AI platform, today unveiled the public preview of the first compute orchestration capabilities for AI workloads across any AI model, on any compute, at any scale.
With Clarifai's new compute orchestration capabilities, customers can maximize their compute investments, better leverage their cloud commitments and hardware for AI, and seamlessly use Clarifai's SaaS compute -- all while centrally managing and monitoring costs, performance, and governance. Clarifai is the only platform capable of building and orchestrating AI workloads across any hardware provider, cloud provider, on-premises, or air-gapped environment, which helps enterprises eliminate vendor lock-in.
"Clarifai has always been ahead of the curve, with over a decade of experience supporting large enterprise and mission-critical government needs with the full stack of AI tools to create custom AI workloads. Now, we're opening up capabilities we built internally to optimize our compute costs as we scaled to serve millions of models simultaneously. Our customers can now have the same world-leading tools to scale their AI workloads on their compute, wherever it may be," said Matt Zeiler, Ph.D., Founder and CEO of Clarifai. "As generative AI grows, our platform enables customers to reduce complexity and seamlessly build and deploy AI in minutes, at a lower cost, with room to scale and flex to meet future business needs easily."
As one of the fastest, production-grade deep learning platforms for developers, data scientists, and MLOps/ML engineers to build on, Clarifai's compute orchestration layer provides unique benefits, including the ability to:
-- Cost-efficiently use compute and abstract away complexity. With an
easy-to-use control plane, customers can effectively govern access to AI
resources, monitor performance, and manage costs while Clarifai's
expert-built platform handles dependencies and optimizations. The
platform can automatically optimize resource usage through model
packing, simple dependency management, and customizable autoscaling
options, including scale-to-zero for both model replicas and compute
nodes. These optimizations have been shown to reduce compute usage by
3.7x through model packing and support 1.6 million+ inputs per second
with 99.9997% reliability. Depending on the configuration, this can lead
to reduced costs of at least 60% and up to 90%.
-- Deploy on any hardware or environment. Customers can deploy any model
using any hardware vendor in any cloud, on-premises, air-gapped, or SaaS
environment for inference. Most competitors focus on inference in
managed cloud or customer Virtual Private Cloud (VPC) deployments. Few
offer on-premises offerings, and none offer air-gapped environment
support with proven success in meeting standards for the most demanding
military deployments.
-- Integrate with a full stack of world-leading AI tools. Clarifai offers a
full-stack platform for the whole AI lifecycle: data labeling, training,
evaluation, workflows, and feedback, which is easy enough for any team
to collaborate on. This allows teams to customize AI workloads for their
business needs in minutes and seamlessly integrate with compute
orchestration to route those workloads to the optimal compute
dynamically.
-- Maintain enterprise security and flexibility. Clarifai can deploy the
compute plane into a customer's cloud VPC or on-premise Kubernetes
cluster without opening up ports into the customer environment, setting
up VPC peering, or creating custom Identity and Access Management (IAM)
roles in their cloud or data centers. Admins are also able to allocate
access to AI resources across teams and projects. This allows users to
manage, monitor, and efficiently utilize multiple compute planes from a
single control plane without sacrificing security.
Clarifai's team of researchers engaged deeply with customers to understand their challenges in optimizing AI performance and spend. Customers highlighted the need for tools that enable cost-effective scaling, seamless provisioning, and flexibility.
"It's the only way you can scale in the end because you make the choices easier for the practitioners rather than getting them to go back to the AI engineering every time, or even worse, trying to upskill them on cloud computing," one customer noted.
Others emphasized the importance of optionality across environments, "If we had a way to think about it holistically and look at our on-prem costs compared to our cloud costs, and then be able to orchestrate across environments with a cost basis, that would be incredibly valuable," another said.
Clarifai's compute orchestration innovation extends Clarifai's focus and leadership position beyond Computer Vision. With over 2 billion operations powering AI for vision, language, and audio, Clarifai has consistently maintained 99.99+% uptime and 24/7 availability, delivering the reliability and flexibility needed for critical applications.
Learn more here or check out the documentation here.
About Clarifai
Clarifai is a global leader in AI and the pioneer of the full-stack AI platform that helps organizations, teams, and developers build, deploy, and operationalize AI at scale. Clarifai's cutting-edge AI platform supports today's modern AI technologies like Large Language Models (LLMs), Large Vision Models (LVMs), Retrieval Augmented Generation (RAG), automated data labeling, high-volume production inference, and more. Founded in 2013, Clarifai is available in cloud, on-premises, or hybrid environments and has been used to build more than 1.5 million AI models with more than 400,000 users in 170 countries. Learn more at www.clarifai.com.
For more information or media requests, contact: pr@clarifai.com.
View original content to download multimedia:https://www.prnewswire.com/news-releases/clarifai-launches-first-vendor-agnostic-compute-orchestration-capabilities-to-optimize-ai-performance-and-spend-302320419.html
SOURCE Clarifai
|
|
|
|
This Week in Tech News: 10 Stories You Need to See | Dec 26, 2024
|
|
GRAMMY-WINNING ARTIST ALESSIA CARA PARTNERS WITH LENOVO AND INTEL FOR THE "MADE BY" CAMPAIGN, FEATURING A BEHIND-THE-SCENES MINISERIES AND EXCLUSIVE LIVE EXPERIENCE | Dec 26, 2024
|
|
Autonomous Driving Technology Leader Plus Adds H.E.L.P.® Alerts in Preparation for Launch of Commercial Driverless Trucking Operations | Dec 26, 2024
|
|
ezW2 2024 Is Now Available For 2025 Tax Season To Businesses At No Cost Increase From Previous Year | Dec 26, 2024
|
|
Omdia: Demand for Google's TPU chips accelerates challenging NVIDIA's dominance | Dec 26, 2024
|
|
Hyperscale Capex Surges 82 Percent in 3Q 2024, Fueled by AI Infrastructure Spending, According to Dell'Oro Group | Dec 26, 2024
|
|
Illinois Public Health Institute Taps 4medica to Lead Chicago Regionwide Community Information Exchange | Dec 26, 2024
|
|
RE/MAX HOLDINGS LAUNCHES THE RE/MAX MEDIA NETWORK, THE FIRST IN REAL ESTATE | Dec 26, 2024
|
|
NAMUGA Launches Next-Generation 3D Sensing Solutions, Powered by Lumotive | Dec 26, 2024
|
|
Smartkem Receives £900,000 Grant from Innovate UK for Advanced MicroLED Displays | Dec 26, 2024
|
|
|
|