SUPERCOMPUTING NEWS SUPERCOMPUTING NEWS
    • EMAIL NEWSLETTER SUBSCRIPTION

    • ACADEMIA
    • AEROSPACE
    • APPLICATIONS
    • ASTRONOMY
    • AUTOMOTIVE
    • BIG DATA
    • BIOLOGY
    • CHEMISTRY
    • CLIENTS
    • CLOUD
    • DEFENSE
    • DEVELOPER TOOLS
    • ECONOMICS
    • EARTH SCIENCES
    • ENGINEERING
    • ENTERTAINMENT
    • GAMING
    • GOVERNMENT
    • HEALTH
    • INDUSTRY
    • INTERCONNECTS
    • MANUFACTURING
    • MIDDLEWARE
    • MOVIES
    • NETWORKS
    • OIL & GAS
    • PHYSICS
    • PROCESSORS
    • RETAIL
    • SCIENCE
    • STORAGE
    • SYSTEMS
    • VISUALIZATION
    • POPULAR ARTICLES
    • RSS FEED
    • ADD YOUR VIDEOS
    • MANAGE VIDEOS
    • EVENTS
      • CALENDAR
      • POST YOUR EVENT
      • GENERAL EVENTS CATEGORY
      • MEETING EVENTS CATEGORY
    • CONVERSATION INBOX
    • SOCIAL ADVERTISER
    • SOCIAL NETWORK VIDEOS
    • SOCIAL ADVERTISEMENTS
    • SURVEYS
    • GROUPS
    • PAGES
    • MARKETPLACE LISTINGS
    • APPLICATIONS BROWSER
    • PRIVACY CONFIRM REQUEST
    • PRIVACY CREATE REQUEST
    • LEADERBOARD
    • POINTS LISTING
      • BADGES
    • MEDIA KIT
    • ADD BANNERS
    • ADD CAMPAIGN
    • CAMPAIGNS PAGE
    • MANAGE ADS
    • MY ORDERS
    • LOGIN/REGISTER
Sign In
How to resolve AdBlock issue?
Refresh this page

FEATURED

Featured

CoreWeave, Perplexity forge a strategic HPC-driven AI partnership

O'NEAL LATEST March 4, 2026, 8:00 am
CoreWeave, Inc. has entered a multi-year partnership with Perplexity AI to provide the infrastructure for Perplexity’s next-generation inference workloads via its specialized AI cloud platform. This strategic collaboration demonstrates how advanced HPC-grade architectures, especially GPU clusters optimized for AI inference, are enabling production-scale AI systems with stringent performance, scalability, and reliability demands.
 
The partnership centers on deploying Perplexity’s inference workloads on CoreWeave’s cloud infrastructure, leveraging dedicated NVIDIA GB200 NVL72-powered clusters to support the high throughput and low latency needed by Perplexity’s Sonar and Search API ecosystem as usage scales.

Inference at Scale: Technical Imperatives

AI inference, serving predictions from pre-trained models in real time, poses unique computational challenges compared with training. While training benefits from large batch sizes and long-duration GPU utilization, inference workloads demand ultra-low latency responses, predictable performance under bursty query patterns, and efficient resource utilization across multi-tenant clusters. For a company like Perplexity, which handles billions of user queries per month, infrastructure that can orchestrate inference workloads at scale with minimal jitter is critical.
 
CoreWeave’s platform is built on a Kubernetes-orchestrated service layer that abstracts and automates resource allocation across GPU clusters. By pairing container orchestration with dedicated hardware, specifically GB200 NVL72 accelerators, CoreWeave ensures that inference models can be deployed without rigid re-architecture while maintaining consistent latency profiles, even at peak demand. This pattern is particularly important as AI models grow in size and complexity, often requiring substantial GPU memory and bandwidth to serve real-time applications effectively.
 
From an engineering perspective, this deployment highlights several critical infrastructure considerations:
  • Workload specialization: Automated tiering of resources for inference vs. training, recognizing that inference tasks often require different memory and throughput characteristics than model training.
  • Latency control: Optimization of GPU-to-network pathways to reduce end-to-end inference time, a key metric for conversational AI and search APIs.
  • Scalability: Dynamic scaling mechanisms that transparently add or remove GPU nodes as load fluctuates, coupled with robust orchestration to prevent resource fragmentation.
  • Cost predictability: Infrastructure designed to avoid over-provisioning while meeting performance SLAs, aided by load-aware scheduling and GPU utilization monitoring. 
Perplexity has already begun running inference workloads on CoreWeave’s platform through its Kubernetes Service and is leveraging tools such as W&B Models to manage models from experimentation to production. This reflects a broader multi-cloud strategy that allows Perplexity to balance resilience, capacity, and vendor flexibility as its AI footprint expands.

Implications for the HPC Community

For supercomputing engineers and architects, this collaboration is emblematic of a broader trend: HPC technologies are transitioning from niche scientific workloads to mainstream AI infrastructure stacks. Traditionally, HPC clusters were associated with physics simulations, climate modeling, and other numerically intensive domains. Increasingly, similar architectures, especially GPU-centric clusters, are now critical for production AI services, requiring operational excellence not just in computational throughput but also in orchestration, fault tolerance, and real-time responsiveness.
 
Platforms like CoreWeave demonstrate that HPC principles, such as parallelism, memory hierarchy optimization, and workload specialization, are foundational to delivering commercial AI services at a global scale. For inference workloads in particular, engineers must consider not just peak compute, but sustained, predictable performance across thousands of queries per second.
 
This shift also presents opportunities for HPC professionals to influence how AI infrastructure evolves: from advising on cluster design and interconnect topologies to developing efficiency-aware scheduling policies that reduce energy consumption without sacrificing performance, an increasingly important consideration as production AI systems grow in scale and footprint.
 
In summary, the CoreWeave, Perplexity alliance exemplifies how cloud platforms purpose-built with HPC knowledge and advanced GPUs are forming the foundation of modern AI services. As inference workloads expand and diversify, platforms that consistently deliver high performance at scale will set themselves apart from general-purpose clouds, reshaping the architecture and deployment of AI applications across industries.
Marina Sirota
Marina Sirota
Featured

AI agents open new frontiers in predicting preterm birth

Tyler O'Neal, Staff Editor LATEST March 3, 2026, 5:36 pm
In a compelling example of artificial intelligence (AI) and high-performance computing (HPC) revolutionizing medical research, scientists at the University of California, San Francisco (UCSF) have created advanced AI tools capable of precisely analyzing vast healthcare datasets to predict preterm birth, a major contributor to infant mortality and long-term health issues globally. Their findings, recently published in Cell Reports Medicine, offer fresh optimism for early intervention and underscore the transformative potential of supercomputing-powered data science in addressing complex biological challenges.
 
Preterm birth, defined as delivery before 37 weeks of gestation, impacts about one in ten pregnancies worldwide and carries a heightened risk of complications, including respiratory distress, neurodevelopmental disorders, and chronic long-term illnesses. Despite years of research, accurately pinpointing which pregnancies are most at risk has proven difficult, primarily because of the complex mix of genetic, environmental, clinical, and lifestyle factors influencing gestational outcomes.
 
The UCSF team, led by Marina Sirota, PhD, professor of Pediatrics and interim director of the Bakar Computational Health Sciences Institute, approached the problem not by narrowing the dataset, but by embracing its scale.
 
The UCSF team addressed this complexity by harnessing machine learning algorithms trained on a vast multi-institutional dataset encompassing millions of electronic health records (EHRs), biomarker measurements, and demographic information. To manage and extract meaningful patterns from such a voluminous and heterogeneous dataset, the researchers relied on a supercomputing infrastructure that could efficiently process and analyze large-scale data in parallel, an essential capability when training and validating predictive AI models.
 
Their model integrates clinical features such as maternal age, blood test results, previous obstetric outcomes, and lifestyle information. Through iterative learning and exposure to diverse cases, AI developed the ability to distinguish subtle signals predictive of preterm birth, achieving significantly higher accuracy than traditional risk scoring systems. The findings reported in Cell Reports Medicine affirm that AI models trained on robust, high-dimensional data can discern patterns that may elude even experienced clinicians.
 
Crucially, the supercomputing element of this research was not merely about speed, but scale and integration. Handling millions of records, each with potentially hundreds of variables, demands computational resources capable of orchestrating complex matrix operations, optimization routines, and cross-validation loops that ensure model generalizability. Standard computing environments struggle with datasets of this magnitude, but HPC systems equipped with parallel processing and optimized data pipelines enabled researchers to train, test, and refine models within feasible time frames.
 
According to the study, this approach represents a paradigm shift in obstetric research. By applying AI to large-scale datasets, we can identify risk profiles long before symptoms manifest. This opens the door to earlier, more personalized interventions that could improve outcomes for mothers and infants alike.
 
The implications are profound. Early prediction of preterm birth could allow clinicians to tailor monitoring schedules, recommend targeted therapies, and provide proactive support to high-risk patients, ultimately reducing the incidence of complications and associated healthcare costs. In regions with limited access to specialized care, AI-driven models could empower frontline providers with actionable insights based on data patterns derived from large cohorts.
 
For the supercomputing community, the model illustrates the expanding role of HPC beyond traditional domains like physics, climate modeling, and astrophysics. In the era of digital medicine, vast datasets generated by electronic health records, genomic sequencing, and wearable sensors present both a challenge and an opportunity: how to turn data into life-saving knowledge. Supercomputers, with their ability to orchestrate trillions of calculations across distributed architectures, are becoming essential partners in this transformation.
 
Moreover, the success of the AI underscores the importance of ethical, transparent, and clinically grounded AI development. The UCSF researchers emphasize that predictive models must be rigorously validated across diverse populations to ensure fairness and avoid perpetuating healthcare disparities. Supercomputing resources make such comprehensive validation feasible, enabling researchers to test model performance across subgroups defined by race, socioeconomic status, and geographic region.
 
As AI continues to mature alongside advances in supercomputing, the pace of medical discovery is poised to accelerate. From predicting preterm birth to personalized cancer therapies and beyond, computational models trained on big data are charting new frontiers in health science, turning complexity into clarity and raw data into actionable insight. 
 
As Sirota and her colleagues demonstrate, when scientific AI meets scalable computing, the result is more than faster analysis. It is the possibility of foresight, the ability to identify risk before crisis emerges.

In maternal health, that foresight could mean healthier pregnancies, stronger newborns, and lives changed by the power of computation.

The low-surface-brightness galaxy CDG-2, within the dashed red circle at right, is dominated by dark matter and contains only a sparse scattering of stars. The full image from NASA’s Hubble Space Telescope is at left. NASA, ESA, Dayi Li (UToronto); Image Processing: Joseph DePasquale (STScI)
Featured

Peering into cosmic darkness: Supercomputers illuminate one of the faintest galaxies ever seen

Tyler O'Neal, Staff Editor LATEST February 24, 2026, 10:00 am
Astronomers have made a discovery that redefines how we think of galaxies, which are often pictured as shining collections of stars. By leveraging cutting-edge space telescopes and state-of-the-art data analysis, scientists have pinpointed one of the dimmest galaxies ever observed, a nearly invisible structure where just a few scattered stars hint at a vast, hidden mass. This achievement, made possible by advanced computational methods, demonstrates how supercomputing and data science are pushing the boundaries of our ability to detect the universe’s faintest and most mysterious objects.
 
Named Candidate Dark Galaxy-2 (CDG-2), this galaxy lies about 300 million light-years from Earth in the Perseus cluster. Instead of shining with billions of stars like most galaxies, CDG-2 emits barely any light, with a visible brightness equivalent to just six million suns. Even more astonishing, over 99 percent of its mass seems to be made of dark matter, the enigmatic, unseen substance that dominates the universe’s mass but does not emit or absorb light.
 
What makes this discovery especially groundbreaking is how the galaxy was found. Rather than detecting the galaxy directly by its stars, researchers used globular clusters, densely packed, gravitationally bound spheres of old stars, as cosmic beacons. These compact clusters were identified as unusually tight groupings in survey data, hinting that they might be orbiting an unseen underlying galaxy. Follow-up imaging with NASA’s Hubble Space Telescope confirmed the clusters’ presence, while data from ESA’s Euclid mission and the Subaru Telescope in Hawaii revealed an ultra-faint diffuse glow around them, the first direct evidence of the dark galaxy itself. 

This detection would have been impossible without high-performance computing and sophisticated statistical models, which are capable of sifting through vast datasets and isolating subtle signals. Modern astrophysical research increasingly relies on supercomputer-assisted analyses to combine multi-telescope observations, model faint features buried in noise, and test competing interpretations of the observed data. In essence, HPC enables astronomers to digitally construct cosmic systems too faint or distant to examine through direct observation alone.
 
CDG-2 stands apart from most known systems not just for its dimness, but for what it may reveal about the role of dark matter in galaxy formation and evolution. The prevailing view in cosmology holds that dark matter provides the gravitational scaffolding around which normal matter, gas, and stars accumulate to form galaxies. Yet the extreme case of CDG-2 suggests scenarios in which star formation was suppressed or stripped away, leaving behind a halo rich in dark matter but poor in visible stars. Such galaxies are thought to be exceedingly rare, making this one a crucial testbed for theories of cosmic structure formation.
 
The supercomputing community should take particular pride in this discovery, as the identification and analysis of CDG-2 depended on algorithms and models developed to handle petabyte-scale datasets from ongoing and upcoming sky surveys. As observatories like the Vera C. Rubin Observatory and the Nancy Grace Roman Space Telescope begin mapping the sky with unprecedented depth and breadth, the role of HPC will only grow, not just in storing and processing data, but in helping astronomers ask new questions about the dark universe and find answers hidden within noise.
 
Moreover, the methods used to detect CDG-2, effectively letting computational exploration precede direct detection, open a new frontier in observational astronomy. In future surveys, machine learning and other supercomputer-powered techniques may routinely uncover objects too faint or too exotic to be seen by the naked eye of a telescope, blurring the line between observation and inference.
 
While CDG-2 may be one of the darkest galaxies yet discovered, its detection casts an inspiring light on the future of astrophysics. It reminds us that the universe still holds countless hidden wonders and that with the synergy of powerful telescopes and supercomputing, we are just beginning to uncover them.
  1. Supercomputers tackle a stellar puzzle, but have we really solved it?
  2. Can Scientific AI truly solve quantum chemistry’s hardest problems?
  • 1
  • 2
Page 1 of 2
POPULAR RIGHT NOW
  • Supercomputing accelerates breakthroughs in diabetes drug discovery
  • Supercomputers power evolutionary insight: From house sparrows to conservation strategies
  • Supercomputing advances the quest to resolve the Hubble tension in cosmology
  • Supercomputing’s next frontier: NVIDIA, CoreWeave unite to build the AI factories of tomorrow
  • Supercomputing drives materials breakthrough for green computing: 3D graphene-like electronic behavior unlocks new low-energy electronics
  • Supercomputing reveals hidden galactic architecture around the Milky Way
  • ML, supercomputing unite to revolutionize high-power laser optics
  • Supercomputers unravel the mystery of missing Tatooine-like planets
  • Supercomputers illuminate deep Earth: How giant 'blobs' shape our magnetic shield
  • Glasgow sets its sights on 'cognitive' cities, where urban systems learn, predict, adapt
How to resolve AdBlock issue?
Refresh this page
THIS YEAR'S MOST READ
  • UVA unveils the power of AI in accelerating new treatment discoveries
  • New study tracks pollution worldwide
  • AI meets DNA: Scientists create custom gene editors with machine learning
  • Darkening oceans: New study reveals alarming decline in marine light zones
  • Supercomputers unlock the chemistry of gecko binding: Vienna team breaks new ground in modeling large molecules
  • At SC25, Phison pushes AI storage to Gen5 speeds, brings AI agents to everyday laptops
  • Harnessing the fury of plasma turbulence: Supercomputer simulations illuminate fusion’s next frontier
  • WSU study pinpoints molecular weak spot in virus entry; supercomputing helps reveal the hidden dance
  • Big numbers, big bets: Dell scales up HPC for the AI era
  • Climate whiplash approaches: A warning from supercomputer simulations
  • FRONTPAGE
  • LATEST
  • POPULAR
  • SOCIAL
  • EVENTS
  • VIDEO
  • SUBSCRIPTION
  • RSS
  • GUIDELINES
  • PRIVACY
  • TOS
  • ABOUT
  • +1 (816) 799-4488
  • editorial@supercomputingonline.com
© 2001 - 2026 SuperComputingOnline.com, LLC.
Sign In
  • FRONT PAGE
  • LATEST
    • POPULAR ARTICLES
    • RSS FEED
    • ACADEMIA
    • AEROSPACE
    • APPLICATIONS
    • ASTRONOMY
    • AUTOMOTIVE
    • BIG DATA
    • BIOLOGY
    • CHEMISTRY
    • CLIENTS
    • CLOUD
    • DEFENSE
    • DEVELOPER TOOLS
    • EARTH SCIENCES
    • ECONOMICS
    • ENGINEERING
    • ENTERTAINMENT
    • HEALTH
    • INDUSTRY
    • INTERCONNECTS
    • GAMING
    • GOVERNMENT
    • MANUFACTURING
    • MIDDLEWARE
    • MOVIES
    • NETWORKS
    • OIL & GAS
    • PHYSICS
    • PROCESSORS
    • RETAIL
    • SCIENCE
    • STORAGE
    • SYSTEMS
    • VISUALIZATION
    • REGISTER
  • VIDEOS
    • ADD YOUR VIDEOS
    • MANAGE VIDEOS
  • COMMUNITY
    • LEADERBOARD
    • APPLICATIONS BROWSER
    • CONVERSATION INBOX
    • GROUPS
    • MARKETPLACE LISTINGS
    • PAGES
    • POINTS LISTING
      • BADGES
    • PRIVACY CONFIRM REQUEST
    • PRIVACY CREATE REQUEST
    • SOCIAL ADVERTISER
    • SOCIAL ADVERTISEMENTS
    • SOCIAL NETWORK VIDEOS
    • SURVEYS
    • EVENTS
      • CALENDAR
      • POST YOUR EVENT
      • GENERAL EVENTS CATEGORY
      • MEETING EVENTS CATEGORY
  • ADVERTISE
    • ADD CAMPAIGN
    • ADD BANNERS
    • CAMPAIGNS PAGE
    • MANAGE ADS
    • MY ORDERS
    • MEDIA KIT
    • LOGIN/REGISTER
  • +1 (816) 799-4488
  • editorial@supercomputingonline.com

Hey there! We noticed you’re using an ad blocker.